Test Report: Docker_Linux_containerd_arm64 17114

                    
                      51f3d9893db86a392fa9064ae9bce74bae887273:2023-08-31:30790
                    
                

Test fail (9/304)

x
+
TestAddons/parallel/Ingress (37.98s)

                                                
                                                
=== RUN   TestAddons/parallel/Ingress
=== PAUSE TestAddons/parallel/Ingress

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/Ingress
addons_test.go:183: (dbg) Run:  kubectl --context addons-015166 wait --for=condition=ready --namespace=ingress-nginx pod --selector=app.kubernetes.io/component=controller --timeout=90s
addons_test.go:208: (dbg) Run:  kubectl --context addons-015166 replace --force -f testdata/nginx-ingress-v1.yaml
addons_test.go:221: (dbg) Run:  kubectl --context addons-015166 replace --force -f testdata/nginx-pod-svc.yaml
addons_test.go:226: (dbg) TestAddons/parallel/Ingress: waiting 8m0s for pods matching "run=nginx" in namespace "default" ...
helpers_test.go:344: "nginx" [a12b11df-507c-4a87-817d-433666d09d86] Pending / Ready:ContainersNotReady (containers with unready status: [nginx]) / ContainersReady:ContainersNotReady (containers with unready status: [nginx])
helpers_test.go:344: "nginx" [a12b11df-507c-4a87-817d-433666d09d86] Running
addons_test.go:226: (dbg) TestAddons/parallel/Ingress: run=nginx healthy within 9.014292872s
addons_test.go:238: (dbg) Run:  out/minikube-linux-arm64 -p addons-015166 ssh "curl -s http://127.0.0.1/ -H 'Host: nginx.example.com'"
addons_test.go:262: (dbg) Run:  kubectl --context addons-015166 replace --force -f testdata/ingress-dns-example-v1.yaml
addons_test.go:267: (dbg) Run:  out/minikube-linux-arm64 -p addons-015166 ip
addons_test.go:273: (dbg) Run:  nslookup hello-john.test 192.168.49.2
addons_test.go:273: (dbg) Non-zero exit: nslookup hello-john.test 192.168.49.2: exit status 1 (15.054338515s)

                                                
                                                
-- stdout --
	;; connection timed out; no servers could be reached
	
	

                                                
                                                
-- /stdout --
addons_test.go:275: failed to nslookup hello-john.test host. args "nslookup hello-john.test 192.168.49.2" : exit status 1
addons_test.go:279: unexpected output from nslookup. stdout: ;; connection timed out; no servers could be reached

                                                
                                                

                                                
                                                

                                                
                                                
stderr: 
addons_test.go:282: (dbg) Run:  out/minikube-linux-arm64 -p addons-015166 addons disable ingress-dns --alsologtostderr -v=1
addons_test.go:282: (dbg) Done: out/minikube-linux-arm64 -p addons-015166 addons disable ingress-dns --alsologtostderr -v=1: (1.098150782s)
addons_test.go:287: (dbg) Run:  out/minikube-linux-arm64 -p addons-015166 addons disable ingress --alsologtostderr -v=1
addons_test.go:287: (dbg) Done: out/minikube-linux-arm64 -p addons-015166 addons disable ingress --alsologtostderr -v=1: (8.076059064s)
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======>  post-mortem[TestAddons/parallel/Ingress]: docker inspect <======
helpers_test.go:231: (dbg) Run:  docker inspect addons-015166
helpers_test.go:235: (dbg) docker inspect addons-015166:

                                                
                                                
-- stdout --
	[
	    {
	        "Id": "a8fb941f1f0f953c7350d6b1c276cf4d522d308c1599eeae95f89ff74b160ef0",
	        "Created": "2023-08-30T22:55:07.556201355Z",
	        "Path": "/usr/local/bin/entrypoint",
	        "Args": [
	            "/sbin/init"
	        ],
	        "State": {
	            "Status": "running",
	            "Running": true,
	            "Paused": false,
	            "Restarting": false,
	            "OOMKilled": false,
	            "Dead": false,
	            "Pid": 1226334,
	            "ExitCode": 0,
	            "Error": "",
	            "StartedAt": "2023-08-30T22:55:07.889181596Z",
	            "FinishedAt": "0001-01-01T00:00:00Z"
	        },
	        "Image": "sha256:879c6efc994c345ac84dd4ebb4fc5b49dd2a4b340e335879382e51233f79b51a",
	        "ResolvConfPath": "/var/lib/docker/containers/a8fb941f1f0f953c7350d6b1c276cf4d522d308c1599eeae95f89ff74b160ef0/resolv.conf",
	        "HostnamePath": "/var/lib/docker/containers/a8fb941f1f0f953c7350d6b1c276cf4d522d308c1599eeae95f89ff74b160ef0/hostname",
	        "HostsPath": "/var/lib/docker/containers/a8fb941f1f0f953c7350d6b1c276cf4d522d308c1599eeae95f89ff74b160ef0/hosts",
	        "LogPath": "/var/lib/docker/containers/a8fb941f1f0f953c7350d6b1c276cf4d522d308c1599eeae95f89ff74b160ef0/a8fb941f1f0f953c7350d6b1c276cf4d522d308c1599eeae95f89ff74b160ef0-json.log",
	        "Name": "/addons-015166",
	        "RestartCount": 0,
	        "Driver": "overlay2",
	        "Platform": "linux",
	        "MountLabel": "",
	        "ProcessLabel": "",
	        "AppArmorProfile": "unconfined",
	        "ExecIDs": null,
	        "HostConfig": {
	            "Binds": [
	                "addons-015166:/var",
	                "/lib/modules:/lib/modules:ro"
	            ],
	            "ContainerIDFile": "",
	            "LogConfig": {
	                "Type": "json-file",
	                "Config": {}
	            },
	            "NetworkMode": "addons-015166",
	            "PortBindings": {
	                "22/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "2376/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "32443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "5000/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "8443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ]
	            },
	            "RestartPolicy": {
	                "Name": "no",
	                "MaximumRetryCount": 0
	            },
	            "AutoRemove": false,
	            "VolumeDriver": "",
	            "VolumesFrom": null,
	            "ConsoleSize": [
	                0,
	                0
	            ],
	            "CapAdd": null,
	            "CapDrop": null,
	            "CgroupnsMode": "host",
	            "Dns": [],
	            "DnsOptions": [],
	            "DnsSearch": [],
	            "ExtraHosts": null,
	            "GroupAdd": null,
	            "IpcMode": "private",
	            "Cgroup": "",
	            "Links": null,
	            "OomScoreAdj": 0,
	            "PidMode": "",
	            "Privileged": true,
	            "PublishAllPorts": false,
	            "ReadonlyRootfs": false,
	            "SecurityOpt": [
	                "seccomp=unconfined",
	                "apparmor=unconfined",
	                "label=disable"
	            ],
	            "Tmpfs": {
	                "/run": "",
	                "/tmp": ""
	            },
	            "UTSMode": "",
	            "UsernsMode": "",
	            "ShmSize": 67108864,
	            "Runtime": "runc",
	            "Isolation": "",
	            "CpuShares": 0,
	            "Memory": 4194304000,
	            "NanoCpus": 2000000000,
	            "CgroupParent": "",
	            "BlkioWeight": 0,
	            "BlkioWeightDevice": [],
	            "BlkioDeviceReadBps": [],
	            "BlkioDeviceWriteBps": [],
	            "BlkioDeviceReadIOps": [],
	            "BlkioDeviceWriteIOps": [],
	            "CpuPeriod": 0,
	            "CpuQuota": 0,
	            "CpuRealtimePeriod": 0,
	            "CpuRealtimeRuntime": 0,
	            "CpusetCpus": "",
	            "CpusetMems": "",
	            "Devices": [],
	            "DeviceCgroupRules": null,
	            "DeviceRequests": null,
	            "MemoryReservation": 0,
	            "MemorySwap": 8388608000,
	            "MemorySwappiness": null,
	            "OomKillDisable": false,
	            "PidsLimit": null,
	            "Ulimits": null,
	            "CpuCount": 0,
	            "CpuPercent": 0,
	            "IOMaximumIOps": 0,
	            "IOMaximumBandwidth": 0,
	            "MaskedPaths": null,
	            "ReadonlyPaths": null
	        },
	        "GraphDriver": {
	            "Data": {
	                "LowerDir": "/var/lib/docker/overlay2/a319c15f302f306a6e594fea92e83e1a7e485ec90885dcec209d164efbefcd4f-init/diff:/var/lib/docker/overlay2/ecca48d9e9e3ae5c70dec79bae341547fcaf05f85377796ae2cd1184b3d7168f/diff",
	                "MergedDir": "/var/lib/docker/overlay2/a319c15f302f306a6e594fea92e83e1a7e485ec90885dcec209d164efbefcd4f/merged",
	                "UpperDir": "/var/lib/docker/overlay2/a319c15f302f306a6e594fea92e83e1a7e485ec90885dcec209d164efbefcd4f/diff",
	                "WorkDir": "/var/lib/docker/overlay2/a319c15f302f306a6e594fea92e83e1a7e485ec90885dcec209d164efbefcd4f/work"
	            },
	            "Name": "overlay2"
	        },
	        "Mounts": [
	            {
	                "Type": "volume",
	                "Name": "addons-015166",
	                "Source": "/var/lib/docker/volumes/addons-015166/_data",
	                "Destination": "/var",
	                "Driver": "local",
	                "Mode": "z",
	                "RW": true,
	                "Propagation": ""
	            },
	            {
	                "Type": "bind",
	                "Source": "/lib/modules",
	                "Destination": "/lib/modules",
	                "Mode": "ro",
	                "RW": false,
	                "Propagation": "rprivate"
	            }
	        ],
	        "Config": {
	            "Hostname": "addons-015166",
	            "Domainname": "",
	            "User": "",
	            "AttachStdin": false,
	            "AttachStdout": false,
	            "AttachStderr": false,
	            "ExposedPorts": {
	                "22/tcp": {},
	                "2376/tcp": {},
	                "32443/tcp": {},
	                "5000/tcp": {},
	                "8443/tcp": {}
	            },
	            "Tty": true,
	            "OpenStdin": false,
	            "StdinOnce": false,
	            "Env": [
	                "container=docker",
	                "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
	            ],
	            "Cmd": null,
	            "Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec",
	            "Volumes": null,
	            "WorkingDir": "/",
	            "Entrypoint": [
	                "/usr/local/bin/entrypoint",
	                "/sbin/init"
	            ],
	            "OnBuild": null,
	            "Labels": {
	                "created_by.minikube.sigs.k8s.io": "true",
	                "mode.minikube.sigs.k8s.io": "addons-015166",
	                "name.minikube.sigs.k8s.io": "addons-015166",
	                "role.minikube.sigs.k8s.io": ""
	            },
	            "StopSignal": "SIGRTMIN+3"
	        },
	        "NetworkSettings": {
	            "Bridge": "",
	            "SandboxID": "fc88c85ba1ec60a1c5079c503f3fe903dd9ebe6885d4e7935ac796e10deea0f2",
	            "HairpinMode": false,
	            "LinkLocalIPv6Address": "",
	            "LinkLocalIPv6PrefixLen": 0,
	            "Ports": {
	                "22/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34314"
	                    }
	                ],
	                "2376/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34313"
	                    }
	                ],
	                "32443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34310"
	                    }
	                ],
	                "5000/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34312"
	                    }
	                ],
	                "8443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34311"
	                    }
	                ]
	            },
	            "SandboxKey": "/var/run/docker/netns/fc88c85ba1ec",
	            "SecondaryIPAddresses": null,
	            "SecondaryIPv6Addresses": null,
	            "EndpointID": "",
	            "Gateway": "",
	            "GlobalIPv6Address": "",
	            "GlobalIPv6PrefixLen": 0,
	            "IPAddress": "",
	            "IPPrefixLen": 0,
	            "IPv6Gateway": "",
	            "MacAddress": "",
	            "Networks": {
	                "addons-015166": {
	                    "IPAMConfig": {
	                        "IPv4Address": "192.168.49.2"
	                    },
	                    "Links": null,
	                    "Aliases": [
	                        "a8fb941f1f0f",
	                        "addons-015166"
	                    ],
	                    "NetworkID": "e4b094cfd79395f0e8d6b44a20c7acdc584d8c0b0e8f44f096b4b8ac417f3d8a",
	                    "EndpointID": "48a430ee68952ef9f1764d82d16ddd42052970c53df7fb70b2c0a91fe225a23d",
	                    "Gateway": "192.168.49.1",
	                    "IPAddress": "192.168.49.2",
	                    "IPPrefixLen": 24,
	                    "IPv6Gateway": "",
	                    "GlobalIPv6Address": "",
	                    "GlobalIPv6PrefixLen": 0,
	                    "MacAddress": "02:42:c0:a8:31:02",
	                    "DriverOpts": null
	                }
	            }
	        }
	    }
	]

                                                
                                                
-- /stdout --
helpers_test.go:239: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p addons-015166 -n addons-015166
helpers_test.go:244: <<< TestAddons/parallel/Ingress FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======>  post-mortem[TestAddons/parallel/Ingress]: minikube logs <======
helpers_test.go:247: (dbg) Run:  out/minikube-linux-arm64 -p addons-015166 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-arm64 -p addons-015166 logs -n 25: (1.660481453s)
helpers_test.go:252: TestAddons/parallel/Ingress logs: 
-- stdout --
	* 
	* ==> Audit <==
	* |---------|--------------------------------|------------------------|---------|---------|---------------------|---------------------|
	| Command |              Args              |        Profile         |  User   | Version |     Start Time      |      End Time       |
	|---------|--------------------------------|------------------------|---------|---------|---------------------|---------------------|
	| start   | -o=json --download-only        | download-only-635688   | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC |                     |
	|         | -p download-only-635688        |                        |         |         |                     |                     |
	|         | --force --alsologtostderr      |                        |         |         |                     |                     |
	|         | --kubernetes-version=v1.16.0   |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	|         | --driver=docker                |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	| start   | -o=json --download-only        | download-only-635688   | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC |                     |
	|         | -p download-only-635688        |                        |         |         |                     |                     |
	|         | --force --alsologtostderr      |                        |         |         |                     |                     |
	|         | --kubernetes-version=v1.28.1   |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	|         | --driver=docker                |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	| delete  | --all                          | minikube               | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC | 30 Aug 23 22:54 UTC |
	| delete  | -p download-only-635688        | download-only-635688   | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC | 30 Aug 23 22:54 UTC |
	| delete  | -p download-only-635688        | download-only-635688   | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC | 30 Aug 23 22:54 UTC |
	| start   | --download-only -p             | download-docker-266424 | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC |                     |
	|         | download-docker-266424         |                        |         |         |                     |                     |
	|         | --alsologtostderr              |                        |         |         |                     |                     |
	|         | --driver=docker                |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	| delete  | -p download-docker-266424      | download-docker-266424 | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC | 30 Aug 23 22:54 UTC |
	| start   | --download-only -p             | binary-mirror-621146   | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC |                     |
	|         | binary-mirror-621146           |                        |         |         |                     |                     |
	|         | --alsologtostderr              |                        |         |         |                     |                     |
	|         | --binary-mirror                |                        |         |         |                     |                     |
	|         | http://127.0.0.1:44437         |                        |         |         |                     |                     |
	|         | --driver=docker                |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	| delete  | -p binary-mirror-621146        | binary-mirror-621146   | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC | 30 Aug 23 22:54 UTC |
	| start   | -p addons-015166               | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC | 30 Aug 23 22:57 UTC |
	|         | --wait=true --memory=4000      |                        |         |         |                     |                     |
	|         | --alsologtostderr              |                        |         |         |                     |                     |
	|         | --addons=registry              |                        |         |         |                     |                     |
	|         | --addons=metrics-server        |                        |         |         |                     |                     |
	|         | --addons=volumesnapshots       |                        |         |         |                     |                     |
	|         | --addons=csi-hostpath-driver   |                        |         |         |                     |                     |
	|         | --addons=gcp-auth              |                        |         |         |                     |                     |
	|         | --addons=cloud-spanner         |                        |         |         |                     |                     |
	|         | --addons=inspektor-gadget      |                        |         |         |                     |                     |
	|         | --driver=docker                |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	|         | --addons=ingress               |                        |         |         |                     |                     |
	|         | --addons=ingress-dns           |                        |         |         |                     |                     |
	| addons  | disable cloud-spanner -p       | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	|         | addons-015166                  |                        |         |         |                     |                     |
	| addons  | addons-015166 addons           | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	|         | disable metrics-server         |                        |         |         |                     |                     |
	|         | --alsologtostderr -v=1         |                        |         |         |                     |                     |
	| addons  | enable headlamp                | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	|         | -p addons-015166               |                        |         |         |                     |                     |
	|         | --alsologtostderr -v=1         |                        |         |         |                     |                     |
	| ip      | addons-015166 ip               | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	| addons  | addons-015166 addons disable   | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	|         | registry --alsologtostderr     |                        |         |         |                     |                     |
	|         | -v=1                           |                        |         |         |                     |                     |
	| ssh     | addons-015166 ssh curl -s      | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	|         | http://127.0.0.1/ -H 'Host:    |                        |         |         |                     |                     |
	|         | nginx.example.com'             |                        |         |         |                     |                     |
	| ip      | addons-015166 ip               | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	| addons  | addons-015166 addons disable   | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	|         | ingress-dns --alsologtostderr  |                        |         |         |                     |                     |
	|         | -v=1                           |                        |         |         |                     |                     |
	| addons  | addons-015166 addons disable   | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:58 UTC |
	|         | ingress --alsologtostderr -v=1 |                        |         |         |                     |                     |
	|---------|--------------------------------|------------------------|---------|---------|---------------------|---------------------|
	
	* 
	* ==> Last Start <==
	* Log file created at: 2023/08/30 22:54:44
	Running on machine: ip-172-31-31-251
	Binary: Built with gc go1.20.7 for linux/arm64
	Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
	I0830 22:54:44.288127 1225867 out.go:296] Setting OutFile to fd 1 ...
	I0830 22:54:44.288286 1225867 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 22:54:44.288296 1225867 out.go:309] Setting ErrFile to fd 2...
	I0830 22:54:44.288301 1225867 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 22:54:44.288579 1225867 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
	I0830 22:54:44.289010 1225867 out.go:303] Setting JSON to false
	I0830 22:54:44.289885 1225867 start.go:128] hostinfo: {"hostname":"ip-172-31-31-251","uptime":27419,"bootTime":1693408666,"procs":183,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1043-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"982e3628-3742-4b3e-bb63-ac1b07660ec7"}
	I0830 22:54:44.289953 1225867 start.go:138] virtualization:  
	I0830 22:54:44.293534 1225867 out.go:177] * [addons-015166] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	I0830 22:54:44.295256 1225867 out.go:177]   - MINIKUBE_LOCATION=17114
	I0830 22:54:44.297092 1225867 out.go:177]   - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0830 22:54:44.295415 1225867 notify.go:220] Checking for updates...
	I0830 22:54:44.300354 1225867 out.go:177]   - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	I0830 22:54:44.302092 1225867 out.go:177]   - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	I0830 22:54:44.304093 1225867 out.go:177]   - MINIKUBE_BIN=out/minikube-linux-arm64
	I0830 22:54:44.305841 1225867 out.go:177]   - MINIKUBE_FORCE_SYSTEMD=
	I0830 22:54:44.307896 1225867 driver.go:373] Setting default libvirt URI to qemu:///system
	I0830 22:54:44.333491 1225867 docker.go:121] docker version: linux-24.0.5:Docker Engine - Community
	I0830 22:54:44.333600 1225867 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 22:54:44.425503 1225867 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:28 OomKillDisable:true NGoroutines:38 SystemTime:2023-08-30 22:54:44.414894496 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 22:54:44.425636 1225867 docker.go:294] overlay module found
	I0830 22:54:44.429303 1225867 out.go:177] * Using the docker driver based on user configuration
	I0830 22:54:44.430880 1225867 start.go:298] selected driver: docker
	I0830 22:54:44.430905 1225867 start.go:902] validating driver "docker" against <nil>
	I0830 22:54:44.430935 1225867 start.go:913] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
	I0830 22:54:44.431574 1225867 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 22:54:44.505406 1225867 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:28 OomKillDisable:true NGoroutines:38 SystemTime:2023-08-30 22:54:44.495875074 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 22:54:44.505563 1225867 start_flags.go:305] no existing cluster config was found, will generate one from the flags 
	I0830 22:54:44.505779 1225867 start_flags.go:919] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
	I0830 22:54:44.507592 1225867 out.go:177] * Using Docker driver with root privileges
	I0830 22:54:44.509430 1225867 cni.go:84] Creating CNI manager for ""
	I0830 22:54:44.509462 1225867 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
	I0830 22:54:44.509473 1225867 start_flags.go:314] Found "CNI" CNI - setting NetworkPlugin=cni
	I0830 22:54:44.509488 1225867 start_flags.go:319] config:
	{Name:addons-015166 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:4000 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.28.1 ClusterName:addons-015166 Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containe
rd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0}
	I0830 22:54:44.511549 1225867 out.go:177] * Starting control plane node addons-015166 in cluster addons-015166
	I0830 22:54:44.513090 1225867 cache.go:122] Beginning downloading kic base image for docker with containerd
	I0830 22:54:44.514803 1225867 out.go:177] * Pulling base image ...
	I0830 22:54:44.516427 1225867 preload.go:132] Checking if preload exists for k8s version v1.28.1 and runtime containerd
	I0830 22:54:44.516481 1225867 preload.go:148] Found local preload: /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.28.1-containerd-overlay2-arm64.tar.lz4
	I0830 22:54:44.516503 1225867 cache.go:57] Caching tarball of preloaded images
	I0830 22:54:44.516511 1225867 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec in local docker daemon
	I0830 22:54:44.516585 1225867 preload.go:174] Found /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.28.1-containerd-overlay2-arm64.tar.lz4 in cache, skipping download
	I0830 22:54:44.516595 1225867 cache.go:60] Finished verifying existence of preloaded tar for  v1.28.1 on containerd
	I0830 22:54:44.517014 1225867 profile.go:148] Saving config to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/config.json ...
	I0830 22:54:44.517034 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/config.json: {Name:mkc799c1e3aadbb7e65ebbcba8e932055b83a00b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:54:44.536561 1225867 cache.go:150] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec to local cache
	I0830 22:54:44.536714 1225867 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec in local cache directory
	I0830 22:54:44.536745 1225867 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec in local cache directory, skipping pull
	I0830 22:54:44.536750 1225867 image.go:105] gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec exists in cache, skipping pull
	I0830 22:54:44.536758 1225867 cache.go:153] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec as a tarball
	I0830 22:54:44.536776 1225867 cache.go:163] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec from local cache
	I0830 22:55:00.495233 1225867 cache.go:165] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec from cached tarball
	I0830 22:55:00.495270 1225867 cache.go:195] Successfully downloaded all kic artifacts
	I0830 22:55:00.495324 1225867 start.go:365] acquiring machines lock for addons-015166: {Name:mkd1caf7e2d7c8b873baf6be8f0355d0884d5b9c Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
	I0830 22:55:00.495457 1225867 start.go:369] acquired machines lock for "addons-015166" in 114.363µs
	I0830 22:55:00.495484 1225867 start.go:93] Provisioning new machine with config: &{Name:addons-015166 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:4000 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.28.1 ClusterName:addons-015166 Namespace:default APIServerName:minikubeCA A
PIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.28.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:
false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0} &{Name: IP: Port:8443 KubernetesVersion:v1.28.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
	I0830 22:55:00.495586 1225867 start.go:125] createHost starting for "" (driver="docker")
	I0830 22:55:00.497857 1225867 out.go:204] * Creating docker container (CPUs=2, Memory=4000MB) ...
	I0830 22:55:00.498120 1225867 start.go:159] libmachine.API.Create for "addons-015166" (driver="docker")
	I0830 22:55:00.498152 1225867 client.go:168] LocalClient.Create starting
	I0830 22:55:00.498277 1225867 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem
	I0830 22:55:00.918608 1225867 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem
	I0830 22:55:01.212104 1225867 cli_runner.go:164] Run: docker network inspect addons-015166 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	W0830 22:55:01.231323 1225867 cli_runner.go:211] docker network inspect addons-015166 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
	I0830 22:55:01.231415 1225867 network_create.go:281] running [docker network inspect addons-015166] to gather additional debugging logs...
	I0830 22:55:01.231435 1225867 cli_runner.go:164] Run: docker network inspect addons-015166
	W0830 22:55:01.254522 1225867 cli_runner.go:211] docker network inspect addons-015166 returned with exit code 1
	I0830 22:55:01.254568 1225867 network_create.go:284] error running [docker network inspect addons-015166]: docker network inspect addons-015166: exit status 1
	stdout:
	[]
	
	stderr:
	Error response from daemon: network addons-015166 not found
	I0830 22:55:01.254581 1225867 network_create.go:286] output of [docker network inspect addons-015166]: -- stdout --
	[]
	
	-- /stdout --
	** stderr ** 
	Error response from daemon: network addons-015166 not found
	
	** /stderr **
	I0830 22:55:01.254649 1225867 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	I0830 22:55:01.274472 1225867 network.go:209] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x40028627d0}
	I0830 22:55:01.274511 1225867 network_create.go:123] attempt to create docker network addons-015166 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
	I0830 22:55:01.274576 1225867 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-015166 addons-015166
	I0830 22:55:01.348031 1225867 network_create.go:107] docker network addons-015166 192.168.49.0/24 created
	I0830 22:55:01.348090 1225867 kic.go:117] calculated static IP "192.168.49.2" for the "addons-015166" container
	I0830 22:55:01.348166 1225867 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
	I0830 22:55:01.365657 1225867 cli_runner.go:164] Run: docker volume create addons-015166 --label name.minikube.sigs.k8s.io=addons-015166 --label created_by.minikube.sigs.k8s.io=true
	I0830 22:55:01.384685 1225867 oci.go:103] Successfully created a docker volume addons-015166
	I0830 22:55:01.384816 1225867 cli_runner.go:164] Run: docker run --rm --name addons-015166-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-015166 --entrypoint /usr/bin/test -v addons-015166:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec -d /var/lib
	I0830 22:55:03.279386 1225867 cli_runner.go:217] Completed: docker run --rm --name addons-015166-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-015166 --entrypoint /usr/bin/test -v addons-015166:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec -d /var/lib: (1.89451146s)
	I0830 22:55:03.279422 1225867 oci.go:107] Successfully prepared a docker volume addons-015166
	I0830 22:55:03.279449 1225867 preload.go:132] Checking if preload exists for k8s version v1.28.1 and runtime containerd
	I0830 22:55:03.279469 1225867 kic.go:190] Starting extracting preloaded images to volume ...
	I0830 22:55:03.279564 1225867 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.28.1-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v addons-015166:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec -I lz4 -xf /preloaded.tar -C /extractDir
	I0830 22:55:07.469719 1225867 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.28.1-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v addons-015166:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec -I lz4 -xf /preloaded.tar -C /extractDir: (4.190097173s)
	I0830 22:55:07.469750 1225867 kic.go:199] duration metric: took 4.190276 seconds to extract preloaded images to volume
	W0830 22:55:07.469900 1225867 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
	I0830 22:55:07.470010 1225867 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
	I0830 22:55:07.539204 1225867 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-015166 --name addons-015166 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-015166 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-015166 --network addons-015166 --ip 192.168.49.2 --volume addons-015166:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec
	I0830 22:55:07.900476 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Running}}
	I0830 22:55:07.925054 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:07.960094 1225867 cli_runner.go:164] Run: docker exec addons-015166 stat /var/lib/dpkg/alternatives/iptables
	I0830 22:55:08.055683 1225867 oci.go:144] the created container "addons-015166" has a running status.
	I0830 22:55:08.055709 1225867 kic.go:221] Creating ssh key for kic: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa...
	I0830 22:55:08.520104 1225867 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
	I0830 22:55:08.558856 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:08.584595 1225867 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
	I0830 22:55:08.584614 1225867 kic_runner.go:114] Args: [docker exec --privileged addons-015166 chown docker:docker /home/docker/.ssh/authorized_keys]
	I0830 22:55:08.686413 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:08.721361 1225867 machine.go:88] provisioning docker machine ...
	I0830 22:55:08.721393 1225867 ubuntu.go:169] provisioning hostname "addons-015166"
	I0830 22:55:08.721459 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:08.750886 1225867 main.go:141] libmachine: Using SSH client type: native
	I0830 22:55:08.751366 1225867 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34314 <nil> <nil>}
	I0830 22:55:08.751384 1225867 main.go:141] libmachine: About to run SSH command:
	sudo hostname addons-015166 && echo "addons-015166" | sudo tee /etc/hostname
	I0830 22:55:08.967147 1225867 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-015166
	
	I0830 22:55:08.967231 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:08.987891 1225867 main.go:141] libmachine: Using SSH client type: native
	I0830 22:55:08.988364 1225867 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34314 <nil> <nil>}
	I0830 22:55:08.988390 1225867 main.go:141] libmachine: About to run SSH command:
	
			if ! grep -xq '.*\saddons-015166' /etc/hosts; then
				if grep -xq '127.0.1.1\s.*' /etc/hosts; then
					sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-015166/g' /etc/hosts;
				else 
					echo '127.0.1.1 addons-015166' | sudo tee -a /etc/hosts; 
				fi
			fi
	I0830 22:55:09.147200 1225867 main.go:141] libmachine: SSH cmd err, output: <nil>: 
	I0830 22:55:09.147226 1225867 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/17114-1219981/.minikube CaCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/17114-1219981/.minikube}
	I0830 22:55:09.147249 1225867 ubuntu.go:177] setting up certificates
	I0830 22:55:09.147258 1225867 provision.go:83] configureAuth start
	I0830 22:55:09.147321 1225867 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-015166
	I0830 22:55:09.173237 1225867 provision.go:138] copyHostCerts
	I0830 22:55:09.173312 1225867 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem (1082 bytes)
	I0830 22:55:09.173433 1225867 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem (1123 bytes)
	I0830 22:55:09.173499 1225867 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem (1679 bytes)
	I0830 22:55:09.173556 1225867 provision.go:112] generating server cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem org=jenkins.addons-015166 san=[192.168.49.2 127.0.0.1 localhost 127.0.0.1 minikube addons-015166]
	I0830 22:55:11.969349 1225867 provision.go:172] copyRemoteCerts
	I0830 22:55:11.969424 1225867 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
	I0830 22:55:11.969473 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:11.987831 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:12.095177 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
	I0830 22:55:12.126863 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
	I0830 22:55:12.157479 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem --> /etc/docker/server.pem (1216 bytes)
	I0830 22:55:12.187894 1225867 provision.go:86] duration metric: configureAuth took 3.040615392s
	I0830 22:55:12.187926 1225867 ubuntu.go:193] setting minikube options for container-runtime
	I0830 22:55:12.188168 1225867 config.go:182] Loaded profile config "addons-015166": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
	I0830 22:55:12.188181 1225867 machine.go:91] provisioned docker machine in 3.466799986s
	I0830 22:55:12.188187 1225867 client.go:171] LocalClient.Create took 11.690030577s
	I0830 22:55:12.188216 1225867 start.go:167] duration metric: libmachine.API.Create for "addons-015166" took 11.690097513s
	I0830 22:55:12.188241 1225867 start.go:300] post-start starting for "addons-015166" (driver="docker")
	I0830 22:55:12.188250 1225867 start.go:329] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
	I0830 22:55:12.188336 1225867 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
	I0830 22:55:12.188381 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:12.206826 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:12.313851 1225867 ssh_runner.go:195] Run: cat /etc/os-release
	I0830 22:55:12.318474 1225867 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
	I0830 22:55:12.318515 1225867 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
	I0830 22:55:12.318527 1225867 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
	I0830 22:55:12.318534 1225867 info.go:137] Remote host: Ubuntu 22.04.3 LTS
	I0830 22:55:12.318544 1225867 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/addons for local assets ...
	I0830 22:55:12.318613 1225867 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/files for local assets ...
	I0830 22:55:12.318641 1225867 start.go:303] post-start completed in 130.394084ms
	I0830 22:55:12.318976 1225867 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-015166
	I0830 22:55:12.337680 1225867 profile.go:148] Saving config to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/config.json ...
	I0830 22:55:12.337979 1225867 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0830 22:55:12.338030 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:12.356290 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:12.459405 1225867 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
	I0830 22:55:12.465439 1225867 start.go:128] duration metric: createHost completed in 11.969838317s
	I0830 22:55:12.465462 1225867 start.go:83] releasing machines lock for "addons-015166", held for 11.969996611s
	I0830 22:55:12.465534 1225867 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-015166
	I0830 22:55:12.483132 1225867 ssh_runner.go:195] Run: cat /version.json
	I0830 22:55:12.483184 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:12.483206 1225867 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
	I0830 22:55:12.483306 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:12.507749 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:12.509782 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:12.610051 1225867 ssh_runner.go:195] Run: systemctl --version
	I0830 22:55:12.743718 1225867 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
	I0830 22:55:12.749595 1225867 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
	I0830 22:55:12.780799 1225867 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
	I0830 22:55:12.780879 1225867 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%!p(MISSING), " -exec sh -c "sudo mv {} {}.mk_disabled" ;
	I0830 22:55:12.815904 1225867 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
	I0830 22:55:12.815930 1225867 start.go:466] detecting cgroup driver to use...
	I0830 22:55:12.815961 1225867 detect.go:196] detected "cgroupfs" cgroup driver on host os
	I0830 22:55:12.816028 1225867 ssh_runner.go:195] Run: sudo systemctl stop -f crio
	I0830 22:55:12.830893 1225867 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
	I0830 22:55:12.844398 1225867 docker.go:196] disabling cri-docker service (if available) ...
	I0830 22:55:12.844462 1225867 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
	I0830 22:55:12.860658 1225867 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
	I0830 22:55:12.877580 1225867 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
	I0830 22:55:12.974862 1225867 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
	I0830 22:55:13.076455 1225867 docker.go:212] disabling docker service ...
	I0830 22:55:13.076559 1225867 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
	I0830 22:55:13.100824 1225867 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
	I0830 22:55:13.115486 1225867 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
	I0830 22:55:13.208014 1225867 ssh_runner.go:195] Run: sudo systemctl mask docker.service
	I0830 22:55:13.312758 1225867 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
	I0830 22:55:13.326519 1225867 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///run/containerd/containerd.sock
	" | sudo tee /etc/crictl.yaml"
	I0830 22:55:13.346444 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.9"|' /etc/containerd/config.toml"
	I0830 22:55:13.358491 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
	I0830 22:55:13.370743 1225867 containerd.go:145] configuring containerd to use "cgroupfs" as cgroup driver...
	I0830 22:55:13.370850 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
	I0830 22:55:13.383098 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
	I0830 22:55:13.395209 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
	I0830 22:55:13.407233 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
	I0830 22:55:13.419163 1225867 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
	I0830 22:55:13.430722 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
	I0830 22:55:13.442720 1225867 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
	I0830 22:55:13.453316 1225867 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
	I0830 22:55:13.464475 1225867 ssh_runner.go:195] Run: sudo systemctl daemon-reload
	I0830 22:55:13.564609 1225867 ssh_runner.go:195] Run: sudo systemctl restart containerd
	I0830 22:55:13.702714 1225867 start.go:513] Will wait 60s for socket path /run/containerd/containerd.sock
	I0830 22:55:13.702833 1225867 ssh_runner.go:195] Run: stat /run/containerd/containerd.sock
	I0830 22:55:13.707912 1225867 start.go:534] Will wait 60s for crictl version
	I0830 22:55:13.708009 1225867 ssh_runner.go:195] Run: which crictl
	I0830 22:55:13.712710 1225867 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
	I0830 22:55:13.762208 1225867 start.go:550] Version:  0.1.0
	RuntimeName:  containerd
	RuntimeVersion:  1.6.22
	RuntimeApiVersion:  v1
	I0830 22:55:13.762349 1225867 ssh_runner.go:195] Run: containerd --version
	I0830 22:55:13.793307 1225867 ssh_runner.go:195] Run: containerd --version
	I0830 22:55:13.830513 1225867 out.go:177] * Preparing Kubernetes v1.28.1 on containerd 1.6.22 ...
	I0830 22:55:13.832626 1225867 cli_runner.go:164] Run: docker network inspect addons-015166 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	I0830 22:55:13.854629 1225867 ssh_runner.go:195] Run: grep 192.168.49.1	host.minikube.internal$ /etc/hosts
	I0830 22:55:13.859439 1225867 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1	host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
	I0830 22:55:13.874109 1225867 preload.go:132] Checking if preload exists for k8s version v1.28.1 and runtime containerd
	I0830 22:55:13.874184 1225867 ssh_runner.go:195] Run: sudo crictl images --output json
	I0830 22:55:13.920916 1225867 containerd.go:604] all images are preloaded for containerd runtime.
	I0830 22:55:13.920939 1225867 containerd.go:518] Images already preloaded, skipping extraction
	I0830 22:55:13.921000 1225867 ssh_runner.go:195] Run: sudo crictl images --output json
	I0830 22:55:13.966563 1225867 containerd.go:604] all images are preloaded for containerd runtime.
	I0830 22:55:13.966585 1225867 cache_images.go:84] Images are preloaded, skipping loading
	I0830 22:55:13.966642 1225867 ssh_runner.go:195] Run: sudo crictl info
	I0830 22:55:14.012144 1225867 cni.go:84] Creating CNI manager for ""
	I0830 22:55:14.012166 1225867 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
	I0830 22:55:14.012198 1225867 kubeadm.go:87] Using pod CIDR: 10.244.0.0/16
	I0830 22:55:14.012221 1225867 kubeadm.go:176] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.28.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-015166 NodeName:addons-015166 DNSDomain:cluster.local CRISocket:/run/containerd/containerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc
/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
	I0830 22:55:14.012366 1225867 kubeadm.go:181] kubeadm config:
	apiVersion: kubeadm.k8s.io/v1beta3
	kind: InitConfiguration
	localAPIEndpoint:
	  advertiseAddress: 192.168.49.2
	  bindPort: 8443
	bootstrapTokens:
	  - groups:
	      - system:bootstrappers:kubeadm:default-node-token
	    ttl: 24h0m0s
	    usages:
	      - signing
	      - authentication
	nodeRegistration:
	  criSocket: unix:///run/containerd/containerd.sock
	  name: "addons-015166"
	  kubeletExtraArgs:
	    node-ip: 192.168.49.2
	  taints: []
	---
	apiVersion: kubeadm.k8s.io/v1beta3
	kind: ClusterConfiguration
	apiServer:
	  certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
	  extraArgs:
	    enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
	controllerManager:
	  extraArgs:
	    allocate-node-cidrs: "true"
	    leader-elect: "false"
	scheduler:
	  extraArgs:
	    leader-elect: "false"
	certificatesDir: /var/lib/minikube/certs
	clusterName: mk
	controlPlaneEndpoint: control-plane.minikube.internal:8443
	etcd:
	  local:
	    dataDir: /var/lib/minikube/etcd
	    extraArgs:
	      proxy-refresh-interval: "70000"
	kubernetesVersion: v1.28.1
	networking:
	  dnsDomain: cluster.local
	  podSubnet: "10.244.0.0/16"
	  serviceSubnet: 10.96.0.0/12
	---
	apiVersion: kubelet.config.k8s.io/v1beta1
	kind: KubeletConfiguration
	authentication:
	  x509:
	    clientCAFile: /var/lib/minikube/certs/ca.crt
	cgroupDriver: cgroupfs
	hairpinMode: hairpin-veth
	runtimeRequestTimeout: 15m
	clusterDomain: "cluster.local"
	# disable disk resource management by default
	imageGCHighThresholdPercent: 100
	evictionHard:
	  nodefs.available: "0%!"(MISSING)
	  nodefs.inodesFree: "0%!"(MISSING)
	  imagefs.available: "0%!"(MISSING)
	failSwapOn: false
	staticPodPath: /etc/kubernetes/manifests
	---
	apiVersion: kubeproxy.config.k8s.io/v1alpha1
	kind: KubeProxyConfiguration
	clusterCIDR: "10.244.0.0/16"
	metricsBindAddress: 0.0.0.0:10249
	conntrack:
	  maxPerCore: 0
	# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
	  tcpEstablishedTimeout: 0s
	# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
	  tcpCloseWaitTimeout: 0s
	
	I0830 22:55:14.012442 1225867 kubeadm.go:976] kubelet [Unit]
	Wants=containerd.service
	
	[Service]
	ExecStart=
	ExecStart=/var/lib/minikube/binaries/v1.28.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --container-runtime-endpoint=unix:///run/containerd/containerd.sock --hostname-override=addons-015166 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
	
	[Install]
	 config:
	{KubernetesVersion:v1.28.1 ClusterName:addons-015166 Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:}
	I0830 22:55:14.012512 1225867 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.28.1
	I0830 22:55:14.023704 1225867 binaries.go:44] Found k8s binaries, skipping transfer
	I0830 22:55:14.023823 1225867 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
	I0830 22:55:14.035057 1225867 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (385 bytes)
	I0830 22:55:14.057565 1225867 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
	I0830 22:55:14.079947 1225867 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2102 bytes)
	I0830 22:55:14.106237 1225867 ssh_runner.go:195] Run: grep 192.168.49.2	control-plane.minikube.internal$ /etc/hosts
	I0830 22:55:14.111013 1225867 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2	control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
	I0830 22:55:14.125227 1225867 certs.go:56] Setting up /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166 for IP: 192.168.49.2
	I0830 22:55:14.125314 1225867 certs.go:190] acquiring lock for shared ca certs: {Name:mk74152f2fceba9f1b7e69c343f47396a1a7db69 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:14.125490 1225867 certs.go:204] generating minikubeCA CA: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.key
	I0830 22:55:14.406524 1225867 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt ...
	I0830 22:55:14.406556 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt: {Name:mk787bb5c72e8c82d69bc86b70d04d11e9f252b3 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:14.406746 1225867 crypto.go:164] Writing key to /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.key ...
	I0830 22:55:14.406758 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.key: {Name:mk45a54d8920d59a9a35fe00274fa515138e47c8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:14.406844 1225867 certs.go:204] generating proxyClientCA CA: /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.key
	I0830 22:55:14.887947 1225867 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.crt ...
	I0830 22:55:14.887976 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.crt: {Name:mk24022c1ad5a0a1f65eb6225d9626b933a5293e Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:14.888158 1225867 crypto.go:164] Writing key to /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.key ...
	I0830 22:55:14.888171 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.key: {Name:mk13508630b0862cd4f3e1b6f73136f2e3a5ebc3 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:14.888673 1225867 certs.go:319] generating minikube-user signed cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.key
	I0830 22:55:14.888710 1225867 crypto.go:68] Generating cert /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt with IP's: []
	I0830 22:55:16.002209 1225867 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt ...
	I0830 22:55:16.002240 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: {Name:mk8634d6e54e4dc420545b2ff8453fbd7319a900 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:16.002425 1225867 crypto.go:164] Writing key to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.key ...
	I0830 22:55:16.002440 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.key: {Name:mk76fc9967556eee915958b419b8f6aafeef30e4 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:16.002521 1225867 certs.go:319] generating minikube signed cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.key.dd3b5fb2
	I0830 22:55:16.002539 1225867 crypto.go:68] Generating cert /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.crt.dd3b5fb2 with IP's: [192.168.49.2 10.96.0.1 127.0.0.1 10.0.0.1]
	I0830 22:55:16.270110 1225867 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.crt.dd3b5fb2 ...
	I0830 22:55:16.270155 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.crt.dd3b5fb2: {Name:mkdef870db4805a3ecc27a682ecbc4af078e9b97 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:16.270404 1225867 crypto.go:164] Writing key to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.key.dd3b5fb2 ...
	I0830 22:55:16.270426 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.key.dd3b5fb2: {Name:mke40cb79e775ea06e515253244caa9b1a3a79d9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:16.270869 1225867 certs.go:337] copying /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.crt.dd3b5fb2 -> /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.crt
	I0830 22:55:16.270960 1225867 certs.go:341] copying /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.key.dd3b5fb2 -> /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.key
	I0830 22:55:16.271011 1225867 certs.go:319] generating aggregator signed cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.key
	I0830 22:55:16.271032 1225867 crypto.go:68] Generating cert /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.crt with IP's: []
	I0830 22:55:16.569561 1225867 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.crt ...
	I0830 22:55:16.569606 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.crt: {Name:mk242a9ae46fff1b1ad3f33882eddd63e3f408af Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:16.569834 1225867 crypto.go:164] Writing key to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.key ...
	I0830 22:55:16.569847 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.key: {Name:mk2fe9db02fac21431e503a5e366107e3e13bdbb Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:16.570063 1225867 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem (1679 bytes)
	I0830 22:55:16.570107 1225867 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem (1082 bytes)
	I0830 22:55:16.570145 1225867 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem (1123 bytes)
	I0830 22:55:16.570175 1225867 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem (1679 bytes)
	I0830 22:55:16.570938 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1399 bytes)
	I0830 22:55:16.601990 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
	I0830 22:55:16.632298 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
	I0830 22:55:16.661682 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
	I0830 22:55:16.692437 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
	I0830 22:55:16.722693 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
	I0830 22:55:16.752134 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
	I0830 22:55:16.781679 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
	I0830 22:55:16.812785 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
	I0830 22:55:16.841975 1225867 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
	I0830 22:55:16.863339 1225867 ssh_runner.go:195] Run: openssl version
	I0830 22:55:16.870397 1225867 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
	I0830 22:55:16.882321 1225867 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
	I0830 22:55:16.887002 1225867 certs.go:480] hashing: -rw-r--r-- 1 root root 1111 Aug 30 22:55 /usr/share/ca-certificates/minikubeCA.pem
	I0830 22:55:16.887093 1225867 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
	I0830 22:55:16.895850 1225867 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
	I0830 22:55:16.907883 1225867 ssh_runner.go:195] Run: ls /var/lib/minikube/certs/etcd
	I0830 22:55:16.912460 1225867 certs.go:353] certs directory doesn't exist, likely first start: ls /var/lib/minikube/certs/etcd: Process exited with status 2
	stdout:
	
	stderr:
	ls: cannot access '/var/lib/minikube/certs/etcd': No such file or directory
	I0830 22:55:16.912520 1225867 kubeadm.go:404] StartCluster: {Name:addons-015166 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:4000 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.28.1 ClusterName:addons-015166 Namespace:default APIServerName:minikubeCA APIServerNames:[] APISe
rverIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.28.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false Disa
bleMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0}
	I0830 22:55:16.912604 1225867 cri.go:54] listing CRI containers in root /run/containerd/runc/k8s.io: {State:paused Name: Namespaces:[kube-system]}
	I0830 22:55:16.912666 1225867 ssh_runner.go:195] Run: sudo -s eval "crictl ps -a --quiet --label io.kubernetes.pod.namespace=kube-system"
	I0830 22:55:16.955838 1225867 cri.go:89] found id: ""
	I0830 22:55:16.955955 1225867 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
	I0830 22:55:16.966655 1225867 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
	I0830 22:55:16.977806 1225867 kubeadm.go:226] ignoring SystemVerification for kubeadm because of docker driver
	I0830 22:55:16.977872 1225867 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
	I0830 22:55:16.988737 1225867 kubeadm.go:152] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
	stdout:
	
	stderr:
	ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
	ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
	ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
	ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
	I0830 22:55:16.988814 1225867 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.28.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml  --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
	I0830 22:55:17.050422 1225867 kubeadm.go:322] [init] Using Kubernetes version: v1.28.1
	I0830 22:55:17.050925 1225867 kubeadm.go:322] [preflight] Running pre-flight checks
	I0830 22:55:17.104432 1225867 kubeadm.go:322] [preflight] The system verification failed. Printing the output from the verification:
	I0830 22:55:17.104498 1225867 kubeadm.go:322] KERNEL_VERSION: 5.15.0-1043-aws
	I0830 22:55:17.104532 1225867 kubeadm.go:322] OS: Linux
	I0830 22:55:17.104578 1225867 kubeadm.go:322] CGROUPS_CPU: enabled
	I0830 22:55:17.104623 1225867 kubeadm.go:322] CGROUPS_CPUACCT: enabled
	I0830 22:55:17.104668 1225867 kubeadm.go:322] CGROUPS_CPUSET: enabled
	I0830 22:55:17.104714 1225867 kubeadm.go:322] CGROUPS_DEVICES: enabled
	I0830 22:55:17.104759 1225867 kubeadm.go:322] CGROUPS_FREEZER: enabled
	I0830 22:55:17.104807 1225867 kubeadm.go:322] CGROUPS_MEMORY: enabled
	I0830 22:55:17.104849 1225867 kubeadm.go:322] CGROUPS_PIDS: enabled
	I0830 22:55:17.104894 1225867 kubeadm.go:322] CGROUPS_HUGETLB: enabled
	I0830 22:55:17.104937 1225867 kubeadm.go:322] CGROUPS_BLKIO: enabled
	I0830 22:55:17.203456 1225867 kubeadm.go:322] [preflight] Pulling images required for setting up a Kubernetes cluster
	I0830 22:55:17.203560 1225867 kubeadm.go:322] [preflight] This might take a minute or two, depending on the speed of your internet connection
	I0830 22:55:17.203652 1225867 kubeadm.go:322] [preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
	I0830 22:55:17.457521 1225867 kubeadm.go:322] [certs] Using certificateDir folder "/var/lib/minikube/certs"
	I0830 22:55:17.462085 1225867 out.go:204]   - Generating certificates and keys ...
	I0830 22:55:17.462182 1225867 kubeadm.go:322] [certs] Using existing ca certificate authority
	I0830 22:55:17.462260 1225867 kubeadm.go:322] [certs] Using existing apiserver certificate and key on disk
	I0830 22:55:18.024124 1225867 kubeadm.go:322] [certs] Generating "apiserver-kubelet-client" certificate and key
	I0830 22:55:18.606799 1225867 kubeadm.go:322] [certs] Generating "front-proxy-ca" certificate and key
	I0830 22:55:19.265465 1225867 kubeadm.go:322] [certs] Generating "front-proxy-client" certificate and key
	I0830 22:55:19.633049 1225867 kubeadm.go:322] [certs] Generating "etcd/ca" certificate and key
	I0830 22:55:19.837853 1225867 kubeadm.go:322] [certs] Generating "etcd/server" certificate and key
	I0830 22:55:19.838241 1225867 kubeadm.go:322] [certs] etcd/server serving cert is signed for DNS names [addons-015166 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
	I0830 22:55:20.764219 1225867 kubeadm.go:322] [certs] Generating "etcd/peer" certificate and key
	I0830 22:55:20.764417 1225867 kubeadm.go:322] [certs] etcd/peer serving cert is signed for DNS names [addons-015166 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
	I0830 22:55:21.086131 1225867 kubeadm.go:322] [certs] Generating "etcd/healthcheck-client" certificate and key
	I0830 22:55:21.381481 1225867 kubeadm.go:322] [certs] Generating "apiserver-etcd-client" certificate and key
	I0830 22:55:21.584542 1225867 kubeadm.go:322] [certs] Generating "sa" key and public key
	I0830 22:55:21.584920 1225867 kubeadm.go:322] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
	I0830 22:55:22.449037 1225867 kubeadm.go:322] [kubeconfig] Writing "admin.conf" kubeconfig file
	I0830 22:55:23.179316 1225867 kubeadm.go:322] [kubeconfig] Writing "kubelet.conf" kubeconfig file
	I0830 22:55:23.721825 1225867 kubeadm.go:322] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
	I0830 22:55:24.035825 1225867 kubeadm.go:322] [kubeconfig] Writing "scheduler.conf" kubeconfig file
	I0830 22:55:24.036484 1225867 kubeadm.go:322] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
	I0830 22:55:24.039365 1225867 kubeadm.go:322] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
	I0830 22:55:24.042000 1225867 out.go:204]   - Booting up control plane ...
	I0830 22:55:24.042152 1225867 kubeadm.go:322] [control-plane] Creating static Pod manifest for "kube-apiserver"
	I0830 22:55:24.042226 1225867 kubeadm.go:322] [control-plane] Creating static Pod manifest for "kube-controller-manager"
	I0830 22:55:24.042288 1225867 kubeadm.go:322] [control-plane] Creating static Pod manifest for "kube-scheduler"
	I0830 22:55:24.057277 1225867 kubeadm.go:322] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
	I0830 22:55:24.058150 1225867 kubeadm.go:322] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
	I0830 22:55:24.058520 1225867 kubeadm.go:322] [kubelet-start] Starting the kubelet
	I0830 22:55:24.175924 1225867 kubeadm.go:322] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
	I0830 22:55:33.678382 1225867 kubeadm.go:322] [apiclient] All control plane components are healthy after 9.502992 seconds
	I0830 22:55:33.678661 1225867 kubeadm.go:322] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
	I0830 22:55:33.707287 1225867 kubeadm.go:322] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
	I0830 22:55:34.237707 1225867 kubeadm.go:322] [upload-certs] Skipping phase. Please see --upload-certs
	I0830 22:55:34.237909 1225867 kubeadm.go:322] [mark-control-plane] Marking the node addons-015166 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
	I0830 22:55:34.749015 1225867 kubeadm.go:322] [bootstrap-token] Using token: 2gijd0.r8muobyvp5fgq8vy
	I0830 22:55:34.751364 1225867 out.go:204]   - Configuring RBAC rules ...
	I0830 22:55:34.751484 1225867 kubeadm.go:322] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
	I0830 22:55:34.756530 1225867 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
	I0830 22:55:34.766834 1225867 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
	I0830 22:55:34.771173 1225867 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
	I0830 22:55:34.775694 1225867 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
	I0830 22:55:34.780041 1225867 kubeadm.go:322] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
	I0830 22:55:34.795221 1225867 kubeadm.go:322] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
	I0830 22:55:35.042070 1225867 kubeadm.go:322] [addons] Applied essential addon: CoreDNS
	I0830 22:55:35.162967 1225867 kubeadm.go:322] [addons] Applied essential addon: kube-proxy
	I0830 22:55:35.166025 1225867 kubeadm.go:322] 
	I0830 22:55:35.166096 1225867 kubeadm.go:322] Your Kubernetes control-plane has initialized successfully!
	I0830 22:55:35.166102 1225867 kubeadm.go:322] 
	I0830 22:55:35.166179 1225867 kubeadm.go:322] To start using your cluster, you need to run the following as a regular user:
	I0830 22:55:35.166184 1225867 kubeadm.go:322] 
	I0830 22:55:35.166208 1225867 kubeadm.go:322]   mkdir -p $HOME/.kube
	I0830 22:55:35.168140 1225867 kubeadm.go:322]   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
	I0830 22:55:35.168211 1225867 kubeadm.go:322]   sudo chown $(id -u):$(id -g) $HOME/.kube/config
	I0830 22:55:35.168216 1225867 kubeadm.go:322] 
	I0830 22:55:35.168268 1225867 kubeadm.go:322] Alternatively, if you are the root user, you can run:
	I0830 22:55:35.168272 1225867 kubeadm.go:322] 
	I0830 22:55:35.168317 1225867 kubeadm.go:322]   export KUBECONFIG=/etc/kubernetes/admin.conf
	I0830 22:55:35.168322 1225867 kubeadm.go:322] 
	I0830 22:55:35.168371 1225867 kubeadm.go:322] You should now deploy a pod network to the cluster.
	I0830 22:55:35.168442 1225867 kubeadm.go:322] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
	I0830 22:55:35.168506 1225867 kubeadm.go:322]   https://kubernetes.io/docs/concepts/cluster-administration/addons/
	I0830 22:55:35.168511 1225867 kubeadm.go:322] 
	I0830 22:55:35.168881 1225867 kubeadm.go:322] You can now join any number of control-plane nodes by copying certificate authorities
	I0830 22:55:35.168959 1225867 kubeadm.go:322] and service account keys on each node and then running the following as root:
	I0830 22:55:35.168964 1225867 kubeadm.go:322] 
	I0830 22:55:35.169905 1225867 kubeadm.go:322]   kubeadm join control-plane.minikube.internal:8443 --token 2gijd0.r8muobyvp5fgq8vy \
	I0830 22:55:35.170012 1225867 kubeadm.go:322] 	--discovery-token-ca-cert-hash sha256:0d761a60bc6e6ab59708493c3270ff221f9632510a1573cc948daf3f99318f4a \
	I0830 22:55:35.170291 1225867 kubeadm.go:322] 	--control-plane 
	I0830 22:55:35.170302 1225867 kubeadm.go:322] 
	I0830 22:55:35.170599 1225867 kubeadm.go:322] Then you can join any number of worker nodes by running the following on each as root:
	I0830 22:55:35.170609 1225867 kubeadm.go:322] 
	I0830 22:55:35.170928 1225867 kubeadm.go:322] kubeadm join control-plane.minikube.internal:8443 --token 2gijd0.r8muobyvp5fgq8vy \
	I0830 22:55:35.171290 1225867 kubeadm.go:322] 	--discovery-token-ca-cert-hash sha256:0d761a60bc6e6ab59708493c3270ff221f9632510a1573cc948daf3f99318f4a 
	I0830 22:55:35.178278 1225867 kubeadm.go:322] 	[WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1043-aws\n", err: exit status 1
	I0830 22:55:35.178391 1225867 kubeadm.go:322] 	[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
	I0830 22:55:35.178410 1225867 cni.go:84] Creating CNI manager for ""
	I0830 22:55:35.178422 1225867 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
	I0830 22:55:35.180748 1225867 out.go:177] * Configuring CNI (Container Networking Interface) ...
	I0830 22:55:35.182853 1225867 ssh_runner.go:195] Run: stat /opt/cni/bin/portmap
	I0830 22:55:35.188531 1225867 cni.go:182] applying CNI manifest using /var/lib/minikube/binaries/v1.28.1/kubectl ...
	I0830 22:55:35.188549 1225867 ssh_runner.go:362] scp memory --> /var/tmp/minikube/cni.yaml (2438 bytes)
	I0830 22:55:35.213341 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml
	I0830 22:55:36.230270 1225867 ssh_runner.go:235] Completed: sudo /var/lib/minikube/binaries/v1.28.1/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml: (1.01689296s)
	I0830 22:55:36.230303 1225867 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
	I0830 22:55:36.230414 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:36.230483 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl label nodes minikube.k8s.io/version=v1.31.2 minikube.k8s.io/commit=dcfed3f069eb419c2ffae8f904d3fba5b9405fc5 minikube.k8s.io/name=addons-015166 minikube.k8s.io/updated_at=2023_08_30T22_55_36_0700 minikube.k8s.io/primary=true --all --overwrite --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:36.438156 1225867 ops.go:34] apiserver oom_adj: -16
	I0830 22:55:36.438266 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:36.532650 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:37.123727 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:37.624076 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:38.124610 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:38.624126 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:39.124288 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:39.624563 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:40.124314 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:40.623586 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:41.123610 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:41.623572 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:42.123886 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:42.623580 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:43.123616 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:43.624546 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:44.123617 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:44.624238 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:45.124003 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:45.624575 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:46.123844 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:46.624285 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:47.124391 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:47.623709 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:47.759873 1225867 kubeadm.go:1081] duration metric: took 11.529500958s to wait for elevateKubeSystemPrivileges.
	I0830 22:55:47.759897 1225867 kubeadm.go:406] StartCluster complete in 30.847380706s
	I0830 22:55:47.759912 1225867 settings.go:142] acquiring lock: {Name:mk7f59b3a5ac74e6581a6632908a77e182a8ae89 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:47.760022 1225867 settings.go:150] Updating kubeconfig:  /home/jenkins/minikube-integration/17114-1219981/kubeconfig
	I0830 22:55:47.760406 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/kubeconfig: {Name:mkdc9959048bf1aa0b8afb8aa74417552569b184 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:47.761055 1225867 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
	I0830 22:55:47.761404 1225867 config.go:182] Loaded profile config "addons-015166": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
	I0830 22:55:47.761516 1225867 addons.go:499] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false helm-tiller:false inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false volumesnapshots:true]
	I0830 22:55:47.761585 1225867 addons.go:69] Setting volumesnapshots=true in profile "addons-015166"
	I0830 22:55:47.761598 1225867 addons.go:231] Setting addon volumesnapshots=true in "addons-015166"
	I0830 22:55:47.761635 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.762112 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.763539 1225867 addons.go:69] Setting ingress-dns=true in profile "addons-015166"
	I0830 22:55:47.763561 1225867 addons.go:231] Setting addon ingress-dns=true in "addons-015166"
	I0830 22:55:47.763631 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.764082 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.764449 1225867 addons.go:69] Setting inspektor-gadget=true in profile "addons-015166"
	I0830 22:55:47.764490 1225867 addons.go:231] Setting addon inspektor-gadget=true in "addons-015166"
	I0830 22:55:47.764532 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.764976 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.765056 1225867 addons.go:69] Setting metrics-server=true in profile "addons-015166"
	I0830 22:55:47.765066 1225867 addons.go:231] Setting addon metrics-server=true in "addons-015166"
	I0830 22:55:47.765093 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.765491 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.765552 1225867 addons.go:69] Setting registry=true in profile "addons-015166"
	I0830 22:55:47.765562 1225867 addons.go:231] Setting addon registry=true in "addons-015166"
	I0830 22:55:47.765591 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.765932 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.765986 1225867 addons.go:69] Setting storage-provisioner=true in profile "addons-015166"
	I0830 22:55:47.766004 1225867 addons.go:231] Setting addon storage-provisioner=true in "addons-015166"
	I0830 22:55:47.766028 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.766395 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.766615 1225867 addons.go:69] Setting default-storageclass=true in profile "addons-015166"
	I0830 22:55:47.766632 1225867 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-015166"
	I0830 22:55:47.766891 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.766966 1225867 addons.go:69] Setting cloud-spanner=true in profile "addons-015166"
	I0830 22:55:47.766980 1225867 addons.go:231] Setting addon cloud-spanner=true in "addons-015166"
	I0830 22:55:47.767009 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.767371 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.767430 1225867 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-015166"
	I0830 22:55:47.767455 1225867 addons.go:231] Setting addon csi-hostpath-driver=true in "addons-015166"
	I0830 22:55:47.767487 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.767874 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.768024 1225867 addons.go:69] Setting gcp-auth=true in profile "addons-015166"
	I0830 22:55:47.768041 1225867 mustload.go:65] Loading cluster: addons-015166
	I0830 22:55:47.768202 1225867 config.go:182] Loaded profile config "addons-015166": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
	I0830 22:55:47.768433 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.774369 1225867 addons.go:69] Setting ingress=true in profile "addons-015166"
	I0830 22:55:47.774407 1225867 addons.go:231] Setting addon ingress=true in "addons-015166"
	I0830 22:55:47.774468 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.774970 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.866361 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
	I0830 22:55:47.868604 1225867 addons.go:423] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
	I0830 22:55:47.868675 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
	I0830 22:55:47.868760 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:47.871046 1225867 out.go:177]   - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v20230407
	I0830 22:55:47.873423 1225867 out.go:177]   - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v20230407
	I0830 22:55:47.876181 1225867 out.go:177]   - Using image registry.k8s.io/ingress-nginx/controller:v1.8.1
	I0830 22:55:47.902163 1225867 addons.go:423] installing /etc/kubernetes/addons/ingress-deploy.yaml
	I0830 22:55:47.902184 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16083 bytes)
	I0830 22:55:47.902246 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:47.930720 1225867 out.go:177]   - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.2
	I0830 22:55:47.939097 1225867 addons.go:423] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
	I0830 22:55:47.939126 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
	I0830 22:55:47.939205 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:47.924432 1225867 kapi.go:248] "coredns" deployment in "kube-system" namespace and "addons-015166" context rescaled to 1 replicas
	I0830 22:55:47.949268 1225867 start.go:223] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.28.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
	I0830 22:55:47.924486 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.978045 1225867 out.go:177]   - Using image registry.k8s.io/metrics-server/metrics-server:v0.6.4
	I0830 22:55:47.975471 1225867 out.go:177] * Verifying Kubernetes components...
	I0830 22:55:47.990408 1225867 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0830 22:55:47.990232 1225867 addons.go:423] installing /etc/kubernetes/addons/metrics-apiservice.yaml
	I0830 22:55:47.990657 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
	I0830 22:55:47.990732 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.055254 1225867 out.go:177]   - Using image docker.io/registry:2.8.1
	I0830 22:55:48.057439 1225867 out.go:177]   - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.5
	I0830 22:55:48.061927 1225867 addons.go:423] installing /etc/kubernetes/addons/registry-rc.yaml
	I0830 22:55:48.061950 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (798 bytes)
	I0830 22:55:48.062025 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.065705 1225867 out.go:177]   - Using image gcr.io/k8s-minikube/storage-provisioner:v5
	I0830 22:55:48.068108 1225867 addons.go:423] installing /etc/kubernetes/addons/storage-provisioner.yaml
	I0830 22:55:48.068127 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
	I0830 22:55:48.068207 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.137859 1225867 out.go:177]   - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.19.0
	I0830 22:55:48.154734 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-namespace.yaml
	I0830 22:55:48.154780 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
	I0830 22:55:48.154868 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.170296 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.170819 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.180504 1225867 addons.go:231] Setting addon default-storageclass=true in "addons-015166"
	I0830 22:55:48.180547 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:48.181015 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:48.186789 1225867 out.go:177]   - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.9
	I0830 22:55:48.189555 1225867 addons.go:423] installing /etc/kubernetes/addons/deployment.yaml
	I0830 22:55:48.189578 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1003 bytes)
	I0830 22:55:48.189647 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.221327 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.265441 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.289346 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
	I0830 22:55:48.295255 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
	I0830 22:55:48.295313 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.300151 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
	I0830 22:55:48.305095 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
	I0830 22:55:48.307086 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
	I0830 22:55:48.311483 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
	I0830 22:55:48.313420 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
	I0830 22:55:48.312395 1225867 node_ready.go:35] waiting up to 6m0s for node "addons-015166" to be "Ready" ...
	I0830 22:55:48.312780 1225867 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^        forward . \/etc\/resolv.conf.*/i \        hosts {\n           192.168.49.1 host.minikube.internal\n           fallthrough\n        }' -e '/^        errors *$/i \        log' | sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
	I0830 22:55:48.317871 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
	I0830 22:55:48.322251 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
	I0830 22:55:48.322267 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
	I0830 22:55:48.322331 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.322163 1225867 node_ready.go:49] node "addons-015166" has status "Ready":"True"
	I0830 22:55:48.322542 1225867 node_ready.go:38] duration metric: took 6.810594ms waiting for node "addons-015166" to be "Ready" ...
	I0830 22:55:48.322555 1225867 pod_ready.go:35] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
	I0830 22:55:48.335715 1225867 pod_ready.go:78] waiting up to 6m0s for pod "coredns-5dd5756b68-dshxx" in "kube-system" namespace to be "Ready" ...
	I0830 22:55:48.376180 1225867 addons.go:423] installing /etc/kubernetes/addons/storageclass.yaml
	I0830 22:55:48.376201 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
	I0830 22:55:48.376266 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.376561 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.377731 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.378574 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.438451 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.446622 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.814877 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
	I0830 22:55:48.897046 1225867 addons.go:423] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
	I0830 22:55:48.897074 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
	I0830 22:55:48.946872 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
	I0830 22:55:48.953415 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-hostpath.yaml
	I0830 22:55:48.953442 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
	I0830 22:55:48.959907 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
	I0830 22:55:49.043597 1225867 addons.go:423] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
	I0830 22:55:49.043667 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
	I0830 22:55:49.058981 1225867 addons.go:423] installing /etc/kubernetes/addons/registry-svc.yaml
	I0830 22:55:49.059055 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
	I0830 22:55:49.075467 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
	I0830 22:55:49.088732 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
	I0830 22:55:49.132079 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
	I0830 22:55:49.132145 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
	I0830 22:55:49.140655 1225867 addons.go:423] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
	I0830 22:55:49.140720 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
	I0830 22:55:49.148109 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
	I0830 22:55:49.148175 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
	I0830 22:55:49.263803 1225867 addons.go:423] installing /etc/kubernetes/addons/registry-proxy.yaml
	I0830 22:55:49.263869 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
	I0830 22:55:49.267128 1225867 addons.go:423] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
	I0830 22:55:49.267193 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
	I0830 22:55:49.340540 1225867 addons.go:423] installing /etc/kubernetes/addons/metrics-server-service.yaml
	I0830 22:55:49.340610 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
	I0830 22:55:49.437995 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
	I0830 22:55:49.476902 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-role.yaml
	I0830 22:55:49.476971 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
	I0830 22:55:49.483940 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
	I0830 22:55:49.484006 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
	I0830 22:55:49.525606 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
	I0830 22:55:49.640406 1225867 addons.go:423] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
	I0830 22:55:49.640475 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
	I0830 22:55:49.770742 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-rolebinding.yaml
	I0830 22:55:49.770814 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
	I0830 22:55:49.785792 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
	I0830 22:55:49.785859 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
	I0830 22:55:49.860921 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
	I0830 22:55:49.860947 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
	I0830 22:55:49.995753 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-clusterrole.yaml
	I0830 22:55:49.995816 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
	I0830 22:55:50.014478 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
	I0830 22:55:50.014541 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
	I0830 22:55:50.053028 1225867 addons.go:423] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
	I0830 22:55:50.053091 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
	I0830 22:55:50.318773 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
	I0830 22:55:50.340134 1225867 addons.go:423] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
	I0830 22:55:50.340241 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
	I0830 22:55:50.366625 1225867 pod_ready.go:97] error getting pod "coredns-5dd5756b68-dshxx" in "kube-system" namespace (skipping!): pods "coredns-5dd5756b68-dshxx" not found
	I0830 22:55:50.366726 1225867 pod_ready.go:81] duration metric: took 2.025980767s waiting for pod "coredns-5dd5756b68-dshxx" in "kube-system" namespace to be "Ready" ...
	E0830 22:55:50.366765 1225867 pod_ready.go:66] WaitExtra: waitPodCondition: error getting pod "coredns-5dd5756b68-dshxx" in "kube-system" namespace (skipping!): pods "coredns-5dd5756b68-dshxx" not found
	I0830 22:55:50.366821 1225867 pod_ready.go:78] waiting up to 6m0s for pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace to be "Ready" ...
	I0830 22:55:50.398806 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
	I0830 22:55:50.398839 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
	I0830 22:55:50.554084 1225867 addons.go:423] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
	I0830 22:55:50.554106 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
	I0830 22:55:50.566085 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-crd.yaml
	I0830 22:55:50.566109 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
	I0830 22:55:50.599660 1225867 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^        forward . \/etc\/resolv.conf.*/i \        hosts {\n           192.168.49.1 host.minikube.internal\n           fallthrough\n        }' -e '/^        errors *$/i \        log' | sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (2.283972982s)
	I0830 22:55:50.599689 1225867 start.go:901] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
	I0830 22:55:50.666399 1225867 addons.go:423] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
	I0830 22:55:50.666423 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
	I0830 22:55:50.689928 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-daemonset.yaml
	I0830 22:55:50.689959 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7741 bytes)
	I0830 22:55:50.823094 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
	I0830 22:55:50.866239 1225867 addons.go:423] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
	I0830 22:55:50.866273 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
	I0830 22:55:51.015705 1225867 addons.go:423] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
	I0830 22:55:51.015729 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
	I0830 22:55:51.263530 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
	I0830 22:55:52.404820 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:55:54.234856 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (5.274922593s)
	I0830 22:55:54.234933 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (5.159432692s)
	I0830 22:55:54.234978 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (5.146216237s)
	I0830 22:55:54.235051 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (4.796980513s)
	I0830 22:55:54.235065 1225867 addons.go:467] Verifying addon metrics-server=true in "addons-015166"
	I0830 22:55:54.235099 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (4.709426444s)
	I0830 22:55:54.235110 1225867 addons.go:467] Verifying addon registry=true in "addons-015166"
	I0830 22:55:54.238813 1225867 out.go:177] * Verifying registry addon...
	I0830 22:55:54.234809 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (5.287897087s)
	I0830 22:55:54.235644 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (3.91679112s)
	I0830 22:55:54.235712 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (3.412575007s)
	I0830 22:55:54.236266 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (5.421359393s)
	I0830 22:55:54.241913 1225867 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
	W0830 22:55:54.242141 1225867 addons.go:449] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
	stdout:
	customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
	customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
	customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
	serviceaccount/snapshot-controller created
	clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
	clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
	role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
	rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
	deployment.apps/snapshot-controller created
	
	stderr:
	error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
	ensure CRDs are installed first
	I0830 22:55:54.242176 1225867 retry.go:31] will retry after 184.541676ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
	stdout:
	customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
	customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
	customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
	serviceaccount/snapshot-controller created
	clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
	clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
	role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
	rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
	deployment.apps/snapshot-controller created
	
	stderr:
	error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
	ensure CRDs are installed first
	I0830 22:55:54.242237 1225867 addons.go:467] Verifying addon ingress=true in "addons-015166"
	I0830 22:55:54.245239 1225867 out.go:177] * Verifying ingress addon...
	I0830 22:55:54.247926 1225867 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
	I0830 22:55:54.249908 1225867 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
	I0830 22:55:54.249938 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:54.261190 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:54.261817 1225867 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
	I0830 22:55:54.261835 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:54.268885 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:54.405371 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:55:54.427121 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
	I0830 22:55:54.778469 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:54.786900 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:54.793796 1225867 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
	I0830 22:55:54.793886 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:54.828644 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:55.267179 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:55.280977 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:55.296761 1225867 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
	I0830 22:55:55.327299 1225867 addons.go:231] Setting addon gcp-auth=true in "addons-015166"
	I0830 22:55:55.327359 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:55.327918 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:55.352583 1225867 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
	I0830 22:55:55.352670 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:55.374845 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:55.767516 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:55.777285 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:56.299547 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:56.302722 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:56.424990 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (1.997763516s)
	I0830 22:55:56.425068 1225867 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (1.072457794s)
	I0830 22:55:56.427658 1225867 out.go:177]   - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v20230407
	I0830 22:55:56.425304 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (5.161710113s)
	I0830 22:55:56.430062 1225867 out.go:177]   - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.0
	I0830 22:55:56.427746 1225867 addons.go:467] Verifying addon csi-hostpath-driver=true in "addons-015166"
	I0830 22:55:56.432423 1225867 addons.go:423] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
	I0830 22:55:56.432449 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
	I0830 22:55:56.434497 1225867 out.go:177] * Verifying csi-hostpath-driver addon...
	I0830 22:55:56.437396 1225867 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
	I0830 22:55:56.455551 1225867 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
	I0830 22:55:56.455580 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:56.470229 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:56.492601 1225867 addons.go:423] installing /etc/kubernetes/addons/gcp-auth-service.yaml
	I0830 22:55:56.492626 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
	I0830 22:55:56.516496 1225867 addons.go:423] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
	I0830 22:55:56.516522 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5412 bytes)
	I0830 22:55:56.543629 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
	I0830 22:55:56.767813 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:56.773866 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:56.897307 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:55:56.980255 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:57.266890 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:57.273715 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:57.485012 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:57.516885 1225867 addons.go:467] Verifying addon gcp-auth=true in "addons-015166"
	I0830 22:55:57.519304 1225867 out.go:177] * Verifying gcp-auth addon...
	I0830 22:55:57.522402 1225867 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
	I0830 22:55:57.534109 1225867 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
	I0830 22:55:57.534185 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:55:57.539988 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:55:57.767176 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:57.773604 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:57.977692 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:58.044564 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:55:58.266854 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:58.273347 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:58.476886 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:58.544516 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:55:58.767168 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:58.773896 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:58.976803 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:59.044735 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:55:59.266608 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:59.274093 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:59.394400 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:55:59.476417 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:59.544491 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:55:59.766486 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:59.772988 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:59.977944 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:00.044863 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:00.272506 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:00.291366 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:00.477876 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:00.546324 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:00.783673 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:00.789453 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:00.977521 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:01.044042 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:01.266658 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:01.274191 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:01.397550 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:01.477476 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:01.545085 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:01.770139 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:01.774682 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:01.978367 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:02.043803 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:02.266704 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:02.273294 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:02.477010 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:02.543786 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:02.766699 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:02.775223 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:02.977988 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:03.043528 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:03.267828 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:03.273775 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:03.401539 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:03.476703 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:03.544877 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:03.767235 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:03.774032 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:03.977426 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:04.044422 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:04.267547 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:04.274547 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:04.477777 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:04.544878 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:04.767360 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:04.774187 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:04.976896 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:05.043959 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:05.267296 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:05.273750 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:05.476744 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:05.544255 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:05.766516 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:05.773737 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:05.895415 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:05.977079 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:06.044323 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:06.266691 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:06.276208 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:06.476758 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:06.544641 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:06.767445 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:06.773721 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:06.976378 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:07.045358 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:07.266540 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:07.273549 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:07.477481 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:07.544376 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:07.766405 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:07.774498 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:07.895764 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:07.985469 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:08.044956 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:08.267521 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:08.274403 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:08.476857 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:08.546039 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:08.767756 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:08.778787 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:08.977764 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:09.044178 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:09.267323 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:09.273912 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:09.479015 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:09.543692 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:09.766019 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:09.773580 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:09.976513 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:10.043497 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:10.267341 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:10.274015 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:10.396113 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:10.476941 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:10.544487 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:10.767035 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:10.773336 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:10.976672 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:11.043923 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:11.265740 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:11.274137 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:11.476144 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:11.543770 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:11.766912 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:11.773513 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:11.976200 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:12.043511 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:12.267120 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:12.273847 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:12.477090 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:12.544292 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:12.766440 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:12.773359 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:12.895089 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:12.976296 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:13.043887 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:13.266980 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:13.273361 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:13.475858 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:13.544494 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:13.766272 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:13.773428 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:13.978315 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:14.043878 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:14.266503 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:14.274148 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:14.476315 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:14.543759 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:14.766091 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:14.773828 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:14.977510 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:15.043632 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:15.266525 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:15.273772 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:15.395371 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:15.476334 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:15.543879 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:15.767711 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:15.773302 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:15.977260 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:16.044662 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:16.266240 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:16.274393 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:16.475789 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:16.544641 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:16.767168 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:16.773413 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:16.976486 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:17.043747 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:17.266745 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:17.273951 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:17.395437 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:17.476315 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:17.543874 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:17.766915 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:17.773435 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:17.976421 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:18.043687 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:18.266633 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:18.273376 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:18.475909 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:18.543391 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:18.766877 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:18.773834 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:18.976696 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:19.044826 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:19.266480 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:19.273030 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:19.397777 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:19.476752 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:19.543817 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:19.766590 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:19.773380 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:19.975520 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:20.044891 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:20.267489 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:20.274400 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:20.478285 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:20.544836 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:20.767722 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:20.774902 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:20.978003 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:21.044544 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:21.268015 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:21.274693 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:21.478432 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:21.544083 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:21.766400 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:21.773925 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:21.895589 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:21.976842 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:22.044792 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:22.266831 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:22.273878 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:22.476429 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:22.546750 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:22.766369 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:22.774125 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:22.975768 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:23.044569 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:23.267429 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:23.274260 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:23.480766 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:23.545634 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:23.768538 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:23.773978 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:23.897190 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:23.977235 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:24.046709 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:24.267621 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:24.275111 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:24.479318 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:24.544228 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:24.769472 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:24.776857 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:24.981234 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:25.046396 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:25.270812 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:25.283040 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:25.481571 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:25.548658 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:25.768807 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:25.776295 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:25.897836 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:25.977667 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:26.044630 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:26.273034 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:26.278753 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:26.492832 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:26.546405 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:26.792821 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:26.797170 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:26.906725 1225867 pod_ready.go:92] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"True"
	I0830 22:56:26.906798 1225867 pod_ready.go:81] duration metric: took 36.539955184s waiting for pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.906834 1225867 pod_ready.go:78] waiting up to 6m0s for pod "etcd-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.916288 1225867 pod_ready.go:92] pod "etcd-addons-015166" in "kube-system" namespace has status "Ready":"True"
	I0830 22:56:26.916351 1225867 pod_ready.go:81] duration metric: took 9.489853ms waiting for pod "etcd-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.916379 1225867 pod_ready.go:78] waiting up to 6m0s for pod "kube-apiserver-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.926984 1225867 pod_ready.go:92] pod "kube-apiserver-addons-015166" in "kube-system" namespace has status "Ready":"True"
	I0830 22:56:26.927048 1225867 pod_ready.go:81] duration metric: took 10.649032ms waiting for pod "kube-apiserver-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.927074 1225867 pod_ready.go:78] waiting up to 6m0s for pod "kube-controller-manager-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.933402 1225867 pod_ready.go:92] pod "kube-controller-manager-addons-015166" in "kube-system" namespace has status "Ready":"True"
	I0830 22:56:26.933469 1225867 pod_ready.go:81] duration metric: took 6.375288ms waiting for pod "kube-controller-manager-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.933496 1225867 pod_ready.go:78] waiting up to 6m0s for pod "kube-proxy-9qnj5" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.946697 1225867 pod_ready.go:92] pod "kube-proxy-9qnj5" in "kube-system" namespace has status "Ready":"True"
	I0830 22:56:26.946773 1225867 pod_ready.go:81] duration metric: took 13.255929ms waiting for pod "kube-proxy-9qnj5" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.946806 1225867 pod_ready.go:78] waiting up to 6m0s for pod "kube-scheduler-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.981050 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:27.046948 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:27.270302 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:27.274360 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:27.293822 1225867 pod_ready.go:92] pod "kube-scheduler-addons-015166" in "kube-system" namespace has status "Ready":"True"
	I0830 22:56:27.293899 1225867 pod_ready.go:81] duration metric: took 347.072461ms waiting for pod "kube-scheduler-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:27.293929 1225867 pod_ready.go:38] duration metric: took 38.971361863s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
	I0830 22:56:27.293977 1225867 api_server.go:52] waiting for apiserver process to appear ...
	I0830 22:56:27.294070 1225867 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 22:56:27.329184 1225867 api_server.go:72] duration metric: took 39.379876034s to wait for apiserver process to appear ...
	I0830 22:56:27.329250 1225867 api_server.go:88] waiting for apiserver healthz status ...
	I0830 22:56:27.329290 1225867 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
	I0830 22:56:27.338596 1225867 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
	ok
	I0830 22:56:27.343208 1225867 api_server.go:141] control plane version: v1.28.1
	I0830 22:56:27.343240 1225867 api_server.go:131] duration metric: took 13.97098ms to wait for apiserver health ...
	I0830 22:56:27.343249 1225867 system_pods.go:43] waiting for kube-system pods to appear ...
	I0830 22:56:27.489102 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:27.504869 1225867 system_pods.go:59] 17 kube-system pods found
	I0830 22:56:27.504958 1225867 system_pods.go:61] "coredns-5dd5756b68-zqzzk" [c01939c6-3560-4bf9-b812-23063cd44277] Running
	I0830 22:56:27.504982 1225867 system_pods.go:61] "csi-hostpath-attacher-0" [050a2cb5-a1c8-4111-a731-ee8003f40a09] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
	I0830 22:56:27.505022 1225867 system_pods.go:61] "csi-hostpath-resizer-0" [fb7d5be5-8065-416b-af32-66d5b72b2341] Running
	I0830 22:56:27.505052 1225867 system_pods.go:61] "csi-hostpathplugin-c4qmz" [844fd9be-03a2-438d-a95d-2f912c388b13] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
	I0830 22:56:27.505072 1225867 system_pods.go:61] "etcd-addons-015166" [23e627bd-c55b-4c2d-9e17-f630fe9dbfb9] Running
	I0830 22:56:27.505091 1225867 system_pods.go:61] "kindnet-gqq4g" [d9dc7990-48db-4173-825e-1ff962f290f6] Running
	I0830 22:56:27.505109 1225867 system_pods.go:61] "kube-apiserver-addons-015166" [e95fbf46-11fb-4bcd-9e7e-07d02e2c7070] Running
	I0830 22:56:27.505303 1225867 system_pods.go:61] "kube-controller-manager-addons-015166" [618ee0b5-f535-45dd-8322-ee5256e71267] Running
	I0830 22:56:27.505341 1225867 system_pods.go:61] "kube-ingress-dns-minikube" [f0723a40-2d31-4e36-9657-0c4e02b37524] Running / Ready:ContainersNotReady (containers with unready status: [minikube-ingress-dns]) / ContainersReady:ContainersNotReady (containers with unready status: [minikube-ingress-dns])
	I0830 22:56:27.505359 1225867 system_pods.go:61] "kube-proxy-9qnj5" [d2ae843c-d8ed-458f-9f49-8dbd3e30c9d2] Running
	I0830 22:56:27.505375 1225867 system_pods.go:61] "kube-scheduler-addons-015166" [fbed79f6-fc5f-4759-aed0-5b85db74727c] Running
	I0830 22:56:27.505397 1225867 system_pods.go:61] "metrics-server-7c66d45ddc-mdsn9" [41e36ea8-5467-4930-a72f-892ee5740e66] Running / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
	I0830 22:56:27.505426 1225867 system_pods.go:61] "registry-gcmd4" [5548c52c-f180-40e9-8007-9ae9bc562acc] Pending / Ready:ContainersNotReady (containers with unready status: [registry]) / ContainersReady:ContainersNotReady (containers with unready status: [registry])
	I0830 22:56:27.505450 1225867 system_pods.go:61] "registry-proxy-r82mq" [e38598dc-2305-4c69-95fc-5496df205f2d] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
	I0830 22:56:27.505468 1225867 system_pods.go:61] "snapshot-controller-58dbcc7b99-2vpgl" [23a2db97-b26d-47cb-afbc-32f39dd6b283] Running
	I0830 22:56:27.505486 1225867 system_pods.go:61] "snapshot-controller-58dbcc7b99-btjhd" [2c8359bc-a1c1-430f-a394-36624245bae5] Running
	I0830 22:56:27.505521 1225867 system_pods.go:61] "storage-provisioner" [160af0dc-ec29-45d3-8494-0a72394086d0] Running
	I0830 22:56:27.505542 1225867 system_pods.go:74] duration metric: took 162.287057ms to wait for pod list to return data ...
	I0830 22:56:27.505574 1225867 default_sa.go:34] waiting for default service account to be created ...
	I0830 22:56:27.545113 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:27.694782 1225867 default_sa.go:45] found service account: "default"
	I0830 22:56:27.694844 1225867 default_sa.go:55] duration metric: took 189.244723ms for default service account to be created ...
	I0830 22:56:27.694868 1225867 system_pods.go:116] waiting for k8s-apps to be running ...
	I0830 22:56:27.767375 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:27.773879 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:27.899274 1225867 system_pods.go:86] 17 kube-system pods found
	I0830 22:56:27.899348 1225867 system_pods.go:89] "coredns-5dd5756b68-zqzzk" [c01939c6-3560-4bf9-b812-23063cd44277] Running
	I0830 22:56:27.899373 1225867 system_pods.go:89] "csi-hostpath-attacher-0" [050a2cb5-a1c8-4111-a731-ee8003f40a09] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
	I0830 22:56:27.899391 1225867 system_pods.go:89] "csi-hostpath-resizer-0" [fb7d5be5-8065-416b-af32-66d5b72b2341] Running
	I0830 22:56:27.899430 1225867 system_pods.go:89] "csi-hostpathplugin-c4qmz" [844fd9be-03a2-438d-a95d-2f912c388b13] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
	I0830 22:56:27.899456 1225867 system_pods.go:89] "etcd-addons-015166" [23e627bd-c55b-4c2d-9e17-f630fe9dbfb9] Running
	I0830 22:56:27.899475 1225867 system_pods.go:89] "kindnet-gqq4g" [d9dc7990-48db-4173-825e-1ff962f290f6] Running
	I0830 22:56:27.899494 1225867 system_pods.go:89] "kube-apiserver-addons-015166" [e95fbf46-11fb-4bcd-9e7e-07d02e2c7070] Running
	I0830 22:56:27.899511 1225867 system_pods.go:89] "kube-controller-manager-addons-015166" [618ee0b5-f535-45dd-8322-ee5256e71267] Running
	I0830 22:56:27.899542 1225867 system_pods.go:89] "kube-ingress-dns-minikube" [f0723a40-2d31-4e36-9657-0c4e02b37524] Running / Ready:ContainersNotReady (containers with unready status: [minikube-ingress-dns]) / ContainersReady:ContainersNotReady (containers with unready status: [minikube-ingress-dns])
	I0830 22:56:27.899567 1225867 system_pods.go:89] "kube-proxy-9qnj5" [d2ae843c-d8ed-458f-9f49-8dbd3e30c9d2] Running
	I0830 22:56:27.899586 1225867 system_pods.go:89] "kube-scheduler-addons-015166" [fbed79f6-fc5f-4759-aed0-5b85db74727c] Running
	I0830 22:56:27.899605 1225867 system_pods.go:89] "metrics-server-7c66d45ddc-mdsn9" [41e36ea8-5467-4930-a72f-892ee5740e66] Running
	I0830 22:56:27.899625 1225867 system_pods.go:89] "registry-gcmd4" [5548c52c-f180-40e9-8007-9ae9bc562acc] Pending / Ready:ContainersNotReady (containers with unready status: [registry]) / ContainersReady:ContainersNotReady (containers with unready status: [registry])
	I0830 22:56:27.899653 1225867 system_pods.go:89] "registry-proxy-r82mq" [e38598dc-2305-4c69-95fc-5496df205f2d] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
	I0830 22:56:27.899676 1225867 system_pods.go:89] "snapshot-controller-58dbcc7b99-2vpgl" [23a2db97-b26d-47cb-afbc-32f39dd6b283] Running
	I0830 22:56:27.899692 1225867 system_pods.go:89] "snapshot-controller-58dbcc7b99-btjhd" [2c8359bc-a1c1-430f-a394-36624245bae5] Running
	I0830 22:56:27.899712 1225867 system_pods.go:89] "storage-provisioner" [160af0dc-ec29-45d3-8494-0a72394086d0] Running
	I0830 22:56:27.899732 1225867 system_pods.go:126] duration metric: took 204.844765ms to wait for k8s-apps to be running ...
	I0830 22:56:27.899759 1225867 system_svc.go:44] waiting for kubelet service to be running ....
	I0830 22:56:27.899835 1225867 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0830 22:56:27.916609 1225867 system_svc.go:56] duration metric: took 16.841041ms WaitForService to wait for kubelet.
	I0830 22:56:27.916634 1225867 kubeadm.go:581] duration metric: took 39.967331751s to wait for : map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] ...
	I0830 22:56:27.916655 1225867 node_conditions.go:102] verifying NodePressure condition ...
	I0830 22:56:27.976660 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:28.044518 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:28.092892 1225867 node_conditions.go:122] node storage ephemeral capacity is 203034800Ki
	I0830 22:56:28.092934 1225867 node_conditions.go:123] node cpu capacity is 2
	I0830 22:56:28.092949 1225867 node_conditions.go:105] duration metric: took 176.289922ms to run NodePressure ...
	I0830 22:56:28.092961 1225867 start.go:228] waiting for startup goroutines ...
	I0830 22:56:28.270225 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:28.274820 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:28.479042 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:28.544903 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:28.766736 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:28.773578 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:28.976393 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:29.044325 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:29.266226 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:29.275002 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:29.475936 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:29.544207 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:29.766134 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:29.773684 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:29.976181 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:30.044738 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:30.267387 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:30.274539 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:30.476774 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:30.544756 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:30.767622 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:30.773933 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:30.978000 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:31.043816 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:31.266048 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:31.273442 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:31.476130 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:31.544812 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:31.767363 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:31.774372 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:31.976700 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:32.044285 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:32.269320 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:32.273801 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:32.481000 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:32.543952 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:32.767346 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:32.774225 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:32.975615 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:33.044610 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:33.266901 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:33.274025 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:33.477189 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:33.543903 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:33.768464 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:33.786263 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:33.981263 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:34.044085 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:34.266833 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:34.274821 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:34.477671 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:34.544281 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:34.773669 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:34.778375 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:34.977711 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:35.044124 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:35.267027 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:35.275931 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:35.476673 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:35.544774 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:35.811651 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:35.815232 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:35.977386 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:36.044181 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:36.266914 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:36.277877 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:36.476786 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:36.546070 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:36.766992 1225867 kapi.go:107] duration metric: took 42.525078256s to wait for kubernetes.io/minikube-addons=registry ...
	I0830 22:56:36.773708 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:36.976455 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:37.043977 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:37.273941 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:37.477803 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:37.544940 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:37.774922 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:37.978460 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:38.046590 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:38.274190 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:38.476371 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:38.544148 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:38.774729 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:38.980986 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:39.047237 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:39.274240 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:39.476328 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:39.544270 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:39.774852 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:39.977383 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:40.044018 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:40.275509 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:40.476284 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:40.544697 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:40.773516 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:40.976349 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:41.044024 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:41.273825 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:41.476367 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:41.543754 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:41.773484 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:41.976504 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:42.044330 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:42.274868 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:42.478019 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:42.543616 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:42.774942 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:42.976907 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:43.051026 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:43.274444 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:43.476356 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:43.543961 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:43.774564 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:43.976537 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:44.043716 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:44.274128 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:44.476358 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:44.546898 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:44.776924 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:44.976748 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:45.044047 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:45.276413 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:45.476416 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:45.544178 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:45.773905 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:45.978165 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:46.045962 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:46.274042 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:46.476390 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:46.552121 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:46.776543 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:46.976425 1225867 kapi.go:107] duration metric: took 50.539026401s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
	I0830 22:56:47.044438 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:47.273862 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:47.544321 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:47.773353 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:48.044600 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:48.273964 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:48.543763 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:48.774907 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:49.044078 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:49.273543 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:49.543619 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:49.774330 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:50.044185 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:50.273996 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:50.543756 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:50.773583 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:51.044464 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:51.274687 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:51.544578 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:51.774091 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:52.043965 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:52.274442 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:52.544167 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:52.773888 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:53.044845 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:53.274033 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:53.543589 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:53.774187 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:54.044107 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:54.274322 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:54.544093 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:54.773751 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:55.044841 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:55.274072 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:55.544103 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:55.774139 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:56.044179 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:56.273308 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:56.544027 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:56.774282 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:57.044212 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:57.274277 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:57.544323 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:57.773719 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:58.044918 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:58.274653 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:58.544995 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:58.774668 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:59.044463 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:59.274163 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:59.544410 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:59.774205 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:00.044601 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:00.280437 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:00.546593 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:00.774467 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:01.044829 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:01.274182 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:01.544666 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:01.774151 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:02.044517 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:02.276259 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:02.544812 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:02.773802 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:03.044896 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:03.273817 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:03.544243 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:03.774880 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:04.045417 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:04.274225 1225867 kapi.go:107] duration metric: took 1m10.026296115s to wait for app.kubernetes.io/name=ingress-nginx ...
	I0830 22:57:04.544385 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:05.044284 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:05.545557 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:06.043662 1225867 kapi.go:107] duration metric: took 1m8.521256432s to wait for kubernetes.io/minikube-addons=gcp-auth ...
	I0830 22:57:06.045448 1225867 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-015166 cluster.
	I0830 22:57:06.047054 1225867 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
	I0830 22:57:06.048742 1225867 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
	I0830 22:57:06.050496 1225867 out.go:177] * Enabled addons: default-storageclass, storage-provisioner, cloud-spanner, metrics-server, ingress-dns, inspektor-gadget, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
	I0830 22:57:06.052223 1225867 addons.go:502] enable addons completed in 1m18.290701031s: enabled=[default-storageclass storage-provisioner cloud-spanner metrics-server ingress-dns inspektor-gadget volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
	I0830 22:57:06.052267 1225867 start.go:233] waiting for cluster config update ...
	I0830 22:57:06.052289 1225867 start.go:242] writing updated cluster config ...
	I0830 22:57:06.052588 1225867 ssh_runner.go:195] Run: rm -f paused
	I0830 22:57:06.117785 1225867 start.go:600] kubectl: 1.28.1, cluster: 1.28.1 (minor skew: 0)
	I0830 22:57:06.120005 1225867 out.go:177] * Done! kubectl is now configured to use "addons-015166" cluster and "default" namespace by default
	
	* 
	* ==> container status <==
	* CONTAINER           IMAGE               CREATED              STATE               NAME                                     ATTEMPT             POD ID              POD
	f2d482f83a864       ab73c7fd67234       3 seconds ago        Running             task-pv-container                        0                   0f21fb582e7ce       task-pv-pod-restore
	d7ea4da00dd68       13753a81eccfd       5 seconds ago        Exited              hello-world-app                          2                   06de2ec44f201       hello-world-app-5d77478584-8crfr
	fb0ee897246c4       fa0c6bb795403       33 seconds ago       Running             nginx                                    0                   791ce5d6d6c24       nginx
	35ffc83ed1eed       71e15c1ff4390       42 seconds ago       Running             headlamp                                 0                   e1abfce443c9d       headlamp-699c48fb74-k524x
	ee04b41f0bf03       2a5f29343eb03       About a minute ago   Running             gcp-auth                                 0                   0860552fad985       gcp-auth-d4c87556c-zsxt6
	1cb36a600adf3       8f2588812ab29       About a minute ago   Exited              patch                                    2                   80a49b91b38cc       ingress-nginx-admission-patch-zc5xm
	4f3b71078fed8       ee6d597e62dc8       About a minute ago   Running             csi-snapshotter                          0                   c357a73a66972       csi-hostpathplugin-c4qmz
	1739a952fc4f0       642ded511e141       About a minute ago   Running             csi-provisioner                          0                   c357a73a66972       csi-hostpathplugin-c4qmz
	5f5359cdcba5e       922312104da8a       About a minute ago   Running             liveness-probe                           0                   c357a73a66972       csi-hostpathplugin-c4qmz
	0f8e9a2b059d9       08f6b2990811a       About a minute ago   Running             hostpath                                 0                   c357a73a66972       csi-hostpathplugin-c4qmz
	3bec36b5d47fd       0107d56dbc0be       About a minute ago   Running             node-driver-registrar                    0                   c357a73a66972       csi-hostpathplugin-c4qmz
	304d7ff1318df       8f2588812ab29       About a minute ago   Exited              create                                   0                   5f8a77a81af5f       ingress-nginx-admission-create-qzcpj
	da433db2dae6c       1461903ec4fe9       About a minute ago   Running             csi-external-health-monitor-controller   0                   c357a73a66972       csi-hostpathplugin-c4qmz
	43ed6efb3f346       9a80d518f102c       About a minute ago   Running             csi-attacher                             0                   4b15c7809c10e       csi-hostpath-attacher-0
	f9a18adb6c2d8       97e04611ad434       About a minute ago   Running             coredns                                  0                   28e7d1bd6036e       coredns-5dd5756b68-zqzzk
	c6c98d79d0f4b       4d1e5c3e97420       About a minute ago   Running             volume-snapshot-controller               0                   90ed02ff215ea       snapshot-controller-58dbcc7b99-btjhd
	e2ce17bf210af       487fa743e1e22       About a minute ago   Running             csi-resizer                              0                   d9c9a0211a6e6       csi-hostpath-resizer-0
	5ae349ee77243       4d1e5c3e97420       About a minute ago   Running             volume-snapshot-controller               0                   bb9cbb600cfa5       snapshot-controller-58dbcc7b99-2vpgl
	f7dfbb5ce2daa       3d00e9f7bc4a0       About a minute ago   Running             gadget                                   0                   06819613aceb9       gadget-khchg
	9a2c72414e97c       ba04bb24b9575       2 minutes ago        Running             storage-provisioner                      0                   ed686676ba58e       storage-provisioner
	d13b68d97b05e       b18bf71b941ba       2 minutes ago        Running             kindnet-cni                              0                   c8a137d0b9f79       kindnet-gqq4g
	c2ca89ac2f5fe       812f5241df7fd       2 minutes ago        Running             kube-proxy                               0                   327ad56d71b9d       kube-proxy-9qnj5
	b1afaa34c0101       8b6e1980b7584       2 minutes ago        Running             kube-controller-manager                  0                   0d6a50dd99362       kube-controller-manager-addons-015166
	4bb20589784e1       b29fb62480892       2 minutes ago        Running             kube-apiserver                           0                   2581802b3bf3c       kube-apiserver-addons-015166
	b355595c7ff7a       9cdd6470f48c8       2 minutes ago        Running             etcd                                     0                   ebedfe81ebaf1       etcd-addons-015166
	f8c96f4a21ae6       b4a5a57e99492       2 minutes ago        Running             kube-scheduler                           0                   deb30374baa14       kube-scheduler-addons-015166
	
	* 
	* ==> containerd <==
	* Aug 30 22:58:00 addons-015166 containerd[745]: time="2023-08-30T22:58:00.526311314Z" level=info msg="TearDown network for sandbox \"cff36fbccd28f3e8647fb38663667bbf882927df5354afb2b12b10678b9cdf24\" successfully"
	Aug 30 22:58:00 addons-015166 containerd[745]: time="2023-08-30T22:58:00.526519634Z" level=info msg="StopPodSandbox for \"cff36fbccd28f3e8647fb38663667bbf882927df5354afb2b12b10678b9cdf24\" returns successfully"
	Aug 30 22:58:01 addons-015166 containerd[745]: time="2023-08-30T22:58:01.019456677Z" level=info msg="RemoveContainer for \"80f748d268e87bd588af505f507cdf45cbf7dfa96f0034073a28745085101bae\""
	Aug 30 22:58:01 addons-015166 containerd[745]: time="2023-08-30T22:58:01.026760940Z" level=info msg="RemoveContainer for \"80f748d268e87bd588af505f507cdf45cbf7dfa96f0034073a28745085101bae\" returns successfully"
	Aug 30 22:58:01 addons-015166 containerd[745]: time="2023-08-30T22:58:01.031302679Z" level=error msg="ContainerStatus for \"80f748d268e87bd588af505f507cdf45cbf7dfa96f0034073a28745085101bae\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"80f748d268e87bd588af505f507cdf45cbf7dfa96f0034073a28745085101bae\": not found"
	Aug 30 22:58:01 addons-015166 containerd[745]: time="2023-08-30T22:58:01.038911935Z" level=info msg="RemoveContainer for \"dcec37596a0cbda8ab1539c81b507254842bebded4556dc1a624d0f69ed6dd74\""
	Aug 30 22:58:01 addons-015166 containerd[745]: time="2023-08-30T22:58:01.044422863Z" level=info msg="RemoveContainer for \"dcec37596a0cbda8ab1539c81b507254842bebded4556dc1a624d0f69ed6dd74\" returns successfully"
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.260986899Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:task-pv-pod-restore,Uid:6b24d1c0-53ae-4f1b-a0cc-17528e535b21,Namespace:default,Attempt:0,}"
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.336589243Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.336736517Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.336771503Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.336986912Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/0f21fb582e7ceae848eddabf489a42b5b79220f218abc1a8cb6b02dfb23a5e10 pid=9402 runtime=io.containerd.runc.v2
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.444804075Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:task-pv-pod-restore,Uid:6b24d1c0-53ae-4f1b-a0cc-17528e535b21,Namespace:default,Attempt:0,} returns sandbox id \"0f21fb582e7ceae848eddabf489a42b5b79220f218abc1a8cb6b02dfb23a5e10\""
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.448995160Z" level=info msg="PullImage \"docker.io/nginx:latest\""
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.451327055Z" level=error msg="failed to decode hosts.toml" error="invalid `host` tree"
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.603480632Z" level=error msg="failed to decode hosts.toml" error="invalid `host` tree"
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.620781497Z" level=info msg="ImageUpdate event &ImageUpdate{Name:docker.io/library/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}"
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.624132197Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:ab73c7fd672341e41ec600081253d0b99ea31d0c1acdfb46a1485004472da7ac,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}"
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.626879895Z" level=info msg="ImageUpdate event &ImageUpdate{Name:docker.io/library/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}"
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.630003830Z" level=info msg="ImageUpdate event &ImageUpdate{Name:docker.io/library/nginx@sha256:104c7c5c54f2685f0f46f3be607ce60da7085da3eaa5ad22d3d9f01594295e9c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}"
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.631156288Z" level=info msg="PullImage \"docker.io/nginx:latest\" returns image reference \"sha256:ab73c7fd672341e41ec600081253d0b99ea31d0c1acdfb46a1485004472da7ac\""
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.638577565Z" level=info msg="CreateContainer within sandbox \"0f21fb582e7ceae848eddabf489a42b5b79220f218abc1a8cb6b02dfb23a5e10\" for container &ContainerMetadata{Name:task-pv-container,Attempt:0,}"
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.662061931Z" level=info msg="CreateContainer within sandbox \"0f21fb582e7ceae848eddabf489a42b5b79220f218abc1a8cb6b02dfb23a5e10\" for &ContainerMetadata{Name:task-pv-container,Attempt:0,} returns container id \"f2d482f83a8644efe536a5ac8f3e7fd28837efbf4d624a3194c311003b528b95\""
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.663053922Z" level=info msg="StartContainer for \"f2d482f83a8644efe536a5ac8f3e7fd28837efbf4d624a3194c311003b528b95\""
	Aug 30 22:58:02 addons-015166 containerd[745]: time="2023-08-30T22:58:02.741945895Z" level=info msg="StartContainer for \"f2d482f83a8644efe536a5ac8f3e7fd28837efbf4d624a3194c311003b528b95\" returns successfully"
	
	* 
	* ==> coredns [f9a18adb6c2d878bcc606e4e339ca5c43300434fd084ea18a7b2aa71a357f2e0] <==
	* [INFO] 10.244.0.16:54471 - 61910 "AAAA IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000056993s
	[INFO] 10.244.0.16:54471 - 53720 "A IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.001429094s
	[INFO] 10.244.0.16:34592 - 3236 "A IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.001982685s
	[INFO] 10.244.0.16:54471 - 46983 "AAAA IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.00162737s
	[INFO] 10.244.0.16:34592 - 57754 "AAAA IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.001570206s
	[INFO] 10.244.0.16:54471 - 43173 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.00025207s
	[INFO] 10.244.0.16:34592 - 47315 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000272131s
	[INFO] 10.244.0.16:35779 - 5561 "A IN hello-world-app.default.svc.cluster.local.ingress-nginx.svc.cluster.local. udp 91 false 512" NXDOMAIN qr,aa,rd 184 0.000099159s
	[INFO] 10.244.0.16:39162 - 40460 "A IN hello-world-app.default.svc.cluster.local.ingress-nginx.svc.cluster.local. udp 91 false 512" NXDOMAIN qr,aa,rd 184 0.000297969s
	[INFO] 10.244.0.16:39162 - 32630 "AAAA IN hello-world-app.default.svc.cluster.local.ingress-nginx.svc.cluster.local. udp 91 false 512" NXDOMAIN qr,aa,rd 184 0.00010336s
	[INFO] 10.244.0.16:35779 - 60214 "AAAA IN hello-world-app.default.svc.cluster.local.ingress-nginx.svc.cluster.local. udp 91 false 512" NXDOMAIN qr,aa,rd 184 0.000096534s
	[INFO] 10.244.0.16:35779 - 57296 "A IN hello-world-app.default.svc.cluster.local.svc.cluster.local. udp 77 false 512" NXDOMAIN qr,aa,rd 170 0.000148119s
	[INFO] 10.244.0.16:39162 - 6278 "A IN hello-world-app.default.svc.cluster.local.svc.cluster.local. udp 77 false 512" NXDOMAIN qr,aa,rd 170 0.000057338s
	[INFO] 10.244.0.16:39162 - 52990 "AAAA IN hello-world-app.default.svc.cluster.local.svc.cluster.local. udp 77 false 512" NXDOMAIN qr,aa,rd 170 0.000049182s
	[INFO] 10.244.0.16:35779 - 24319 "AAAA IN hello-world-app.default.svc.cluster.local.svc.cluster.local. udp 77 false 512" NXDOMAIN qr,aa,rd 170 0.000145149s
	[INFO] 10.244.0.16:35779 - 11713 "A IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000088353s
	[INFO] 10.244.0.16:39162 - 3921 "A IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000196808s
	[INFO] 10.244.0.16:39162 - 27022 "AAAA IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000067979s
	[INFO] 10.244.0.16:35779 - 38494 "AAAA IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000037612s
	[INFO] 10.244.0.16:35779 - 24234 "A IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.002085832s
	[INFO] 10.244.0.16:39162 - 4374 "A IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.002105294s
	[INFO] 10.244.0.16:35779 - 47799 "AAAA IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.001371256s
	[INFO] 10.244.0.16:39162 - 5174 "AAAA IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.00152159s
	[INFO] 10.244.0.16:39162 - 32423 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000064632s
	[INFO] 10.244.0.16:35779 - 9069 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000162741s
	
	* 
	* ==> describe nodes <==
	* Name:               addons-015166
	Roles:              control-plane
	Labels:             beta.kubernetes.io/arch=arm64
	                    beta.kubernetes.io/os=linux
	                    kubernetes.io/arch=arm64
	                    kubernetes.io/hostname=addons-015166
	                    kubernetes.io/os=linux
	                    minikube.k8s.io/commit=dcfed3f069eb419c2ffae8f904d3fba5b9405fc5
	                    minikube.k8s.io/name=addons-015166
	                    minikube.k8s.io/primary=true
	                    minikube.k8s.io/updated_at=2023_08_30T22_55_36_0700
	                    minikube.k8s.io/version=v1.31.2
	                    node-role.kubernetes.io/control-plane=
	                    node.kubernetes.io/exclude-from-external-load-balancers=
	                    topology.hostpath.csi/node=addons-015166
	Annotations:        csi.volume.kubernetes.io/nodeid: {"hostpath.csi.k8s.io":"addons-015166"}
	                    kubeadm.alpha.kubernetes.io/cri-socket: unix:///run/containerd/containerd.sock
	                    node.alpha.kubernetes.io/ttl: 0
	                    volumes.kubernetes.io/controller-managed-attach-detach: true
	CreationTimestamp:  Wed, 30 Aug 2023 22:55:31 +0000
	Taints:             <none>
	Unschedulable:      false
	Lease:
	  HolderIdentity:  addons-015166
	  AcquireTime:     <unset>
	  RenewTime:       Wed, 30 Aug 2023 22:57:57 +0000
	Conditions:
	  Type             Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message
	  ----             ------  -----------------                 ------------------                ------                       -------
	  MemoryPressure   False   Wed, 30 Aug 2023 22:57:37 +0000   Wed, 30 Aug 2023 22:55:27 +0000   KubeletHasSufficientMemory   kubelet has sufficient memory available
	  DiskPressure     False   Wed, 30 Aug 2023 22:57:37 +0000   Wed, 30 Aug 2023 22:55:27 +0000   KubeletHasNoDiskPressure     kubelet has no disk pressure
	  PIDPressure      False   Wed, 30 Aug 2023 22:57:37 +0000   Wed, 30 Aug 2023 22:55:27 +0000   KubeletHasSufficientPID      kubelet has sufficient PID available
	  Ready            True    Wed, 30 Aug 2023 22:57:37 +0000   Wed, 30 Aug 2023 22:55:35 +0000   KubeletReady                 kubelet is posting ready status
	Addresses:
	  InternalIP:  192.168.49.2
	  Hostname:    addons-015166
	Capacity:
	  cpu:                2
	  ephemeral-storage:  203034800Ki
	  hugepages-1Gi:      0
	  hugepages-2Mi:      0
	  hugepages-32Mi:     0
	  hugepages-64Ki:     0
	  memory:             8022572Ki
	  pods:               110
	Allocatable:
	  cpu:                2
	  ephemeral-storage:  203034800Ki
	  hugepages-1Gi:      0
	  hugepages-2Mi:      0
	  hugepages-32Mi:     0
	  hugepages-64Ki:     0
	  memory:             8022572Ki
	  pods:               110
	System Info:
	  Machine ID:                 8d46025cfb614346be2593e1a340a0fd
	  System UUID:                5ab26e7d-ae38-4bb3-8872-d7eafeec3dbd
	  Boot ID:                    98673563-8173-4281-afb4-eac1dfafdc23
	  Kernel Version:             5.15.0-1043-aws
	  OS Image:                   Ubuntu 22.04.3 LTS
	  Operating System:           linux
	  Architecture:               arm64
	  Container Runtime Version:  containerd://1.6.22
	  Kubelet Version:            v1.28.1
	  Kube-Proxy Version:         v1.28.1
	PodCIDR:                      10.244.0.0/24
	PodCIDRs:                     10.244.0.0/24
	Non-terminated Pods:          (19 in total)
	  Namespace                   Name                                     CPU Requests  CPU Limits  Memory Requests  Memory Limits  Age
	  ---------                   ----                                     ------------  ----------  ---------------  -------------  ---
	  default                     hello-world-app-5d77478584-8crfr         0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         26s
	  default                     nginx                                    0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         36s
	  default                     task-pv-pod-restore                      0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         4s
	  gadget                      gadget-khchg                             0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m12s
	  gcp-auth                    gcp-auth-d4c87556c-zsxt6                 0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m8s
	  headlamp                    headlamp-699c48fb74-k524x                0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         46s
	  kube-system                 coredns-5dd5756b68-zqzzk                 100m (5%!)(MISSING)     0 (0%!)(MISSING)      70Mi (0%!)(MISSING)        170Mi (2%!)(MISSING)     2m18s
	  kube-system                 csi-hostpath-attacher-0                  0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m9s
	  kube-system                 csi-hostpath-resizer-0                   0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m9s
	  kube-system                 csi-hostpathplugin-c4qmz                 0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m9s
	  kube-system                 etcd-addons-015166                       100m (5%!)(MISSING)     0 (0%!)(MISSING)      100Mi (1%!)(MISSING)       0 (0%!)(MISSING)         2m30s
	  kube-system                 kindnet-gqq4g                            100m (5%!)(MISSING)     100m (5%!)(MISSING)   50Mi (0%!)(MISSING)        50Mi (0%!)(MISSING)      2m19s
	  kube-system                 kube-apiserver-addons-015166             250m (12%!)(MISSING)    0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m30s
	  kube-system                 kube-controller-manager-addons-015166    200m (10%!)(MISSING)    0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m30s
	  kube-system                 kube-proxy-9qnj5                         0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m19s
	  kube-system                 kube-scheduler-addons-015166             100m (5%!)(MISSING)     0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m30s
	  kube-system                 snapshot-controller-58dbcc7b99-2vpgl     0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m12s
	  kube-system                 snapshot-controller-58dbcc7b99-btjhd     0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m12s
	  kube-system                 storage-provisioner                      0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m13s
	Allocated resources:
	  (Total limits may be over 100 percent, i.e., overcommitted.)
	  Resource           Requests    Limits
	  --------           --------    ------
	  cpu                850m (42%!)(MISSING)  100m (5%!)(MISSING)
	  memory             220Mi (2%!)(MISSING)  220Mi (2%!)(MISSING)
	  ephemeral-storage  0 (0%!)(MISSING)      0 (0%!)(MISSING)
	  hugepages-1Gi      0 (0%!)(MISSING)      0 (0%!)(MISSING)
	  hugepages-2Mi      0 (0%!)(MISSING)      0 (0%!)(MISSING)
	  hugepages-32Mi     0 (0%!)(MISSING)      0 (0%!)(MISSING)
	  hugepages-64Ki     0 (0%!)(MISSING)      0 (0%!)(MISSING)
	Events:
	  Type    Reason                   Age                    From             Message
	  ----    ------                   ----                   ----             -------
	  Normal  Starting                 2m17s                  kube-proxy       
	  Normal  Starting                 2m40s                  kubelet          Starting kubelet.
	  Normal  NodeAllocatableEnforced  2m40s                  kubelet          Updated Node Allocatable limit across pods
	  Normal  NodeHasSufficientMemory  2m39s (x8 over 2m40s)  kubelet          Node addons-015166 status is now: NodeHasSufficientMemory
	  Normal  NodeHasNoDiskPressure    2m39s (x8 over 2m40s)  kubelet          Node addons-015166 status is now: NodeHasNoDiskPressure
	  Normal  NodeHasSufficientPID     2m39s (x7 over 2m40s)  kubelet          Node addons-015166 status is now: NodeHasSufficientPID
	  Normal  Starting                 2m30s                  kubelet          Starting kubelet.
	  Normal  NodeHasSufficientMemory  2m30s                  kubelet          Node addons-015166 status is now: NodeHasSufficientMemory
	  Normal  NodeHasNoDiskPressure    2m30s                  kubelet          Node addons-015166 status is now: NodeHasNoDiskPressure
	  Normal  NodeHasSufficientPID     2m30s                  kubelet          Node addons-015166 status is now: NodeHasSufficientPID
	  Normal  NodeNotReady             2m30s                  kubelet          Node addons-015166 status is now: NodeNotReady
	  Normal  NodeAllocatableEnforced  2m30s                  kubelet          Updated Node Allocatable limit across pods
	  Normal  NodeReady                2m30s                  kubelet          Node addons-015166 status is now: NodeReady
	  Normal  RegisteredNode           2m19s                  node-controller  Node addons-015166 event: Registered Node addons-015166 in Controller
	
	* 
	* ==> dmesg <==
	* [  +0.001068] FS-Cache: O-key=[8] 'a53f5c0100000000'
	[  +0.000743] FS-Cache: N-cookie c=00000054 [p=0000004b fl=2 nc=0 na=1]
	[  +0.000990] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=0000000052a3ffac
	[  +0.001181] FS-Cache: N-key=[8] 'a53f5c0100000000'
	[  +0.003620] FS-Cache: Duplicate cookie detected
	[  +0.000757] FS-Cache: O-cookie c=0000004e [p=0000004b fl=226 nc=0 na=1]
	[  +0.000989] FS-Cache: O-cookie d=00000000d8a48a2b{9p.inode} n=00000000cfc10e18
	[  +0.001078] FS-Cache: O-key=[8] 'a53f5c0100000000'
	[  +0.000892] FS-Cache: N-cookie c=00000055 [p=0000004b fl=2 nc=0 na=1]
	[  +0.000999] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=00000000ec866464
	[  +0.001154] FS-Cache: N-key=[8] 'a53f5c0100000000'
	[  +3.285800] FS-Cache: Duplicate cookie detected
	[  +0.000913] FS-Cache: O-cookie c=0000004c [p=0000004b fl=226 nc=0 na=1]
	[  +0.001105] FS-Cache: O-cookie d=00000000d8a48a2b{9p.inode} n=00000000185770a2
	[  +0.001225] FS-Cache: O-key=[8] 'a43f5c0100000000'
	[  +0.000833] FS-Cache: N-cookie c=00000057 [p=0000004b fl=2 nc=0 na=1]
	[  +0.001080] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=000000006d053276
	[  +0.001194] FS-Cache: N-key=[8] 'a43f5c0100000000'
	[  +0.414572] FS-Cache: Duplicate cookie detected
	[  +0.000724] FS-Cache: O-cookie c=00000051 [p=0000004b fl=226 nc=0 na=1]
	[  +0.000967] FS-Cache: O-cookie d=00000000d8a48a2b{9p.inode} n=000000001a64e3e4
	[  +0.001029] FS-Cache: O-key=[8] 'aa3f5c0100000000'
	[  +0.000731] FS-Cache: N-cookie c=00000058 [p=0000004b fl=2 nc=0 na=1]
	[  +0.000927] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=00000000a41d18fb
	[  +0.001092] FS-Cache: N-key=[8] 'aa3f5c0100000000'
	
	* 
	* ==> etcd [b355595c7ff7afc32de0f59644494b8cdeec14e940ea30b97ee8745babc29b13] <==
	* {"level":"info","ts":"2023-08-30T22:55:26.858137Z","caller":"fileutil/purge.go:44","msg":"started to purge file","dir":"/var/lib/minikube/etcd/member/snap","suffix":"snap.db","max":5,"interval":"30s"}
	{"level":"info","ts":"2023-08-30T22:55:26.861394Z","caller":"fileutil/purge.go:44","msg":"started to purge file","dir":"/var/lib/minikube/etcd/member/snap","suffix":"snap","max":5,"interval":"30s"}
	{"level":"info","ts":"2023-08-30T22:55:26.861509Z","caller":"fileutil/purge.go:44","msg":"started to purge file","dir":"/var/lib/minikube/etcd/member/wal","suffix":"wal","max":5,"interval":"30s"}
	{"level":"info","ts":"2023-08-30T22:55:26.862175Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc switched to configuration voters=(12593026477526642892)"}
	{"level":"info","ts":"2023-08-30T22:55:26.862531Z","caller":"membership/cluster.go:421","msg":"added member","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","added-peer-id":"aec36adc501070cc","added-peer-peer-urls":["https://192.168.49.2:2380"]}
	{"level":"info","ts":"2023-08-30T22:55:26.862426Z","caller":"embed/etcd.go:597","msg":"serving peer traffic","address":"192.168.49.2:2380"}
	{"level":"info","ts":"2023-08-30T22:55:26.862736Z","caller":"embed/etcd.go:569","msg":"cmux::serve","address":"192.168.49.2:2380"}
	{"level":"info","ts":"2023-08-30T22:55:27.02517Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc is starting a new election at term 1"}
	{"level":"info","ts":"2023-08-30T22:55:27.025387Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became pre-candidate at term 1"}
	{"level":"info","ts":"2023-08-30T22:55:27.02549Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgPreVoteResp from aec36adc501070cc at term 1"}
	{"level":"info","ts":"2023-08-30T22:55:27.025592Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became candidate at term 2"}
	{"level":"info","ts":"2023-08-30T22:55:27.025672Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2"}
	{"level":"info","ts":"2023-08-30T22:55:27.025757Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
	{"level":"info","ts":"2023-08-30T22:55:27.025799Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
	{"level":"info","ts":"2023-08-30T22:55:27.029329Z","caller":"etcdserver/server.go:2062","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-015166 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
	{"level":"info","ts":"2023-08-30T22:55:27.029692Z","caller":"etcdserver/server.go:2571","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
	{"level":"info","ts":"2023-08-30T22:55:27.029898Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
	{"level":"info","ts":"2023-08-30T22:55:27.031194Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
	{"level":"info","ts":"2023-08-30T22:55:27.031375Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
	{"level":"info","ts":"2023-08-30T22:55:27.032076Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
	{"level":"info","ts":"2023-08-30T22:55:27.032167Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
	{"level":"info","ts":"2023-08-30T22:55:27.037239Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
	{"level":"info","ts":"2023-08-30T22:55:27.037364Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
	{"level":"info","ts":"2023-08-30T22:55:27.037432Z","caller":"etcdserver/server.go:2595","msg":"cluster version is updated","cluster-version":"3.5"}
	{"level":"info","ts":"2023-08-30T22:55:27.061609Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
	
	* 
	* ==> gcp-auth [ee04b41f0bf0363aaa39bc186c74821e142bcd0766c8a043300f4da588108c3a] <==
	* 2023/08/30 22:57:05 GCP Auth Webhook started!
	2023/08/30 22:57:16 Ready to marshal response ...
	2023/08/30 22:57:16 Ready to write response ...
	2023/08/30 22:57:18 Ready to marshal response ...
	2023/08/30 22:57:18 Ready to write response ...
	2023/08/30 22:57:19 Ready to marshal response ...
	2023/08/30 22:57:19 Ready to write response ...
	2023/08/30 22:57:19 Ready to marshal response ...
	2023/08/30 22:57:19 Ready to write response ...
	2023/08/30 22:57:29 Ready to marshal response ...
	2023/08/30 22:57:29 Ready to write response ...
	2023/08/30 22:57:39 Ready to marshal response ...
	2023/08/30 22:57:39 Ready to write response ...
	2023/08/30 22:57:39 Ready to marshal response ...
	2023/08/30 22:57:39 Ready to write response ...
	2023/08/30 22:58:01 Ready to marshal response ...
	2023/08/30 22:58:01 Ready to write response ...
	
	* 
	* ==> kernel <==
	*  22:58:06 up  7:40,  0 users,  load average: 1.53, 2.07, 2.24
	Linux addons-015166 5.15.0-1043-aws #48~20.04.1-Ubuntu SMP Wed Aug 16 18:32:42 UTC 2023 aarch64 aarch64 aarch64 GNU/Linux
	PRETTY_NAME="Ubuntu 22.04.3 LTS"
	
	* 
	* ==> kindnet [d13b68d97b05e1f052a3d690ae64cf0abaf5eca3b8e292e66001baf603e1d752] <==
	* I0830 22:55:48.596511       1 main.go:146] kindnetd IP family: "ipv4"
	I0830 22:55:48.596520       1 main.go:150] noMask IPv4 subnets: [10.244.0.0/16]
	I0830 22:56:18.910999       1 main.go:191] Failed to get nodes, retrying after error: Get "https://10.96.0.1:443/api/v1/nodes": dial tcp 10.96.0.1:443: i/o timeout
	I0830 22:56:18.925096       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:56:18.925171       1 main.go:227] handling current node
	I0830 22:56:28.984719       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:56:28.984743       1 main.go:227] handling current node
	I0830 22:56:38.999912       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:56:39.000043       1 main.go:227] handling current node
	I0830 22:56:49.004071       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:56:49.004099       1 main.go:227] handling current node
	I0830 22:56:59.018012       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:56:59.018039       1 main.go:227] handling current node
	I0830 22:57:09.022553       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:57:09.022631       1 main.go:227] handling current node
	I0830 22:57:19.046083       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:57:19.046135       1 main.go:227] handling current node
	I0830 22:57:29.057239       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:57:29.057333       1 main.go:227] handling current node
	I0830 22:57:39.070064       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:57:39.070098       1 main.go:227] handling current node
	I0830 22:57:49.078100       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:57:49.078128       1 main.go:227] handling current node
	I0830 22:57:59.089194       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:57:59.089223       1 main.go:227] handling current node
	
	* 
	* ==> kube-apiserver [4bb20589784e15f56ca13bf9ddcc489be39f6c6fba30fe528b12a4eca03b529c] <==
	* E0830 22:56:27.602293       1 available_controller.go:460] v1beta1.metrics.k8s.io failed with: failing or missing response from https://10.104.116.236:443/apis/metrics.k8s.io/v1beta1: Get "https://10.104.116.236:443/apis/metrics.k8s.io/v1beta1": dial tcp 10.104.116.236:443: connect: connection refused
	W0830 22:56:27.602441       1 handler_proxy.go:93] no RequestInfo found in the context
	E0830 22:56:27.602518       1 controller.go:143] Error updating APIService "v1beta1.metrics.k8s.io" with err: failed to download v1beta1.metrics.k8s.io: failed to retrieve openAPI spec, http error: ResponseCode: 503, Body: service unavailable
	, Header: map[Content-Type:[text/plain; charset=utf-8] X-Content-Type-Options:[nosniff]]
	E0830 22:56:27.603347       1 available_controller.go:460] v1beta1.metrics.k8s.io failed with: failing or missing response from https://10.104.116.236:443/apis/metrics.k8s.io/v1beta1: Get "https://10.104.116.236:443/apis/metrics.k8s.io/v1beta1": dial tcp 10.104.116.236:443: connect: connection refused
	I0830 22:56:27.603604       1 handler_discovery.go:337] DiscoveryManager: Failed to download discovery for kube-system/metrics-server:443: 503 error trying to reach service: dial tcp 10.104.116.236:443: connect: connection refused
	I0830 22:56:27.603617       1 handler.go:232] Adding GroupVersion metrics.k8s.io v1beta1 to ResourceManager
	E0830 22:56:27.609875       1 available_controller.go:460] v1beta1.metrics.k8s.io failed with: failing or missing response from https://10.104.116.236:443/apis/metrics.k8s.io/v1beta1: Get "https://10.104.116.236:443/apis/metrics.k8s.io/v1beta1": dial tcp 10.104.116.236:443: connect: connection refused
	I0830 22:56:27.732413       1 handler.go:232] Adding GroupVersion metrics.k8s.io v1beta1 to ResourceManager
	I0830 22:56:31.573927       1 handler.go:232] Adding GroupVersion metrics.k8s.io v1beta1 to ResourceManager
	http2: server: error reading preface from client 192.168.49.1:55888: read tcp 192.168.49.2:8443->192.168.49.1:55888: read: connection reset by peer
	E0830 22:57:17.541660       1 controller.go:159] removing "v1beta1.metrics.k8s.io" from AggregationController failed with: resource not found
	I0830 22:57:19.004417       1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.102.174.55"}
	E0830 22:57:19.976348       1 watch.go:287] unable to encode watch object *v1.WatchEvent: http2: stream closed (&streaming.encoderWithAllocator{writer:responsewriter.outerWithCloseNotifyAndFlush{UserProvidedDecorator:(*metrics.ResponseWriterDelegator)(0x400b446b40), InnerCloseNotifierFlusher:struct { httpsnoop.Unwrapper; http.ResponseWriter; http.Flusher; http.CloseNotifier; http.Pusher }{Unwrapper:(*httpsnoop.rw)(0x400e4fb270), ResponseWriter:(*httpsnoop.rw)(0x400e4fb270), Flusher:(*httpsnoop.rw)(0x400e4fb270), CloseNotifier:(*httpsnoop.rw)(0x400e4fb270), Pusher:(*httpsnoop.rw)(0x400e4fb270)}}, encoder:(*versioning.codec)(0x400db13d60), memAllocator:(*runtime.Allocator)(0x400f030b58)})
	E0830 22:57:28.615265       1 handler_proxy.go:137] error resolving kube-system/metrics-server: service "metrics-server" not found
	W0830 22:57:28.615295       1 handler_proxy.go:93] no RequestInfo found in the context
	E0830 22:57:28.615328       1 controller.go:113] loading OpenAPI spec for "v1beta1.metrics.k8s.io" failed with: failed to retrieve openAPI spec, http error: ResponseCode: 503, Body: service unavailable
	, Header: map[Content-Type:[text/plain; charset=utf-8] X-Content-Type-Options:[nosniff]]
	I0830 22:57:28.615336       1 controller.go:126] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Rate Limited Requeue.
	I0830 22:57:29.596407       1 controller.go:624] quota admission added evaluator for: ingresses.networking.k8s.io
	I0830 22:57:29.905689       1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.108.157.67"}
	I0830 22:57:39.936735       1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.103.138.78"}
	I0830 22:57:50.776153       1 controller.go:624] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
	E0830 22:57:57.272802       1 authentication.go:70] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"ingress-nginx\" not found]"
	E0830 22:57:57.753413       1 authentication.go:70] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"ingress-nginx\" not found]"
	
	* 
	* ==> kube-controller-manager [b1afaa34c01017de27552235b2fb893c1bf242a2a0ca49aa160cad3ce5fe48a3] <==
	* I0830 22:57:19.149494       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="headlamp/headlamp-699c48fb74" duration="98.854292ms"
	I0830 22:57:19.185428       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="headlamp/headlamp-699c48fb74" duration="35.873654ms"
	I0830 22:57:19.221411       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="headlamp/headlamp-699c48fb74" duration="35.926602ms"
	I0830 22:57:19.221734       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="headlamp/headlamp-699c48fb74" duration="277.645µs"
	I0830 22:57:21.114706       1 replica_set.go:676] "Finished syncing" kind="ReplicationController" key="kube-system/registry" duration="10.108µs"
	I0830 22:57:21.758121       1 event.go:307] "Event occurred" object="default/hpvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'hostpath.csi.k8s.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
	I0830 22:57:23.888380       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="headlamp/headlamp-699c48fb74" duration="42.437µs"
	I0830 22:57:23.929118       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="headlamp/headlamp-699c48fb74" duration="17.056639ms"
	I0830 22:57:23.929257       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="headlamp/headlamp-699c48fb74" duration="55.287µs"
	I0830 22:57:31.465200       1 event.go:307] "Event occurred" object="default/hpvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'hostpath.csi.k8s.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
	I0830 22:57:38.450971       1 event.go:307] "Event occurred" object="default/hpvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'hostpath.csi.k8s.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
	I0830 22:57:39.683098       1 event.go:307] "Event occurred" object="default/hello-world-app" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set hello-world-app-5d77478584 to 1"
	I0830 22:57:39.732730       1 event.go:307] "Event occurred" object="default/hello-world-app-5d77478584" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: hello-world-app-5d77478584-8crfr"
	I0830 22:57:39.743235       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="59.488006ms"
	I0830 22:57:39.792360       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="48.95346ms"
	I0830 22:57:39.793399       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="35.832µs"
	I0830 22:57:45.979402       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="33.485µs"
	I0830 22:57:46.982547       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="44.496µs"
	I0830 22:57:47.981852       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="44.955µs"
	I0830 22:57:52.430561       1 event.go:307] "Event occurred" object="default/hpvc-restore" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'hostpath.csi.k8s.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
	I0830 22:57:57.142408       1 job_controller.go:562] "enqueueing job" key="ingress-nginx/ingress-nginx-admission-create"
	I0830 22:57:57.156767       1 job_controller.go:562] "enqueueing job" key="ingress-nginx/ingress-nginx-admission-patch"
	I0830 22:57:57.157722       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="ingress-nginx/ingress-nginx-controller-5dcd45b5bf" duration="4.85µs"
	I0830 22:58:00.916651       1 event.go:307] "Event occurred" object="default/hpvc-restore" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'hostpath.csi.k8s.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
	I0830 22:58:01.036638       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="67.455µs"
	
	* 
	* ==> kube-proxy [c2ca89ac2f5fe88c90af377524e8eb9b6261c0306583f26fef191aa3353bf792] <==
	* I0830 22:55:48.546978       1 server_others.go:69] "Using iptables proxy"
	I0830 22:55:48.567373       1 node.go:141] Successfully retrieved node IP: 192.168.49.2
	I0830 22:55:48.644154       1 server.go:632] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
	I0830 22:55:48.649866       1 server_others.go:152] "Using iptables Proxier"
	I0830 22:55:48.649911       1 server_others.go:421] "Detect-local-mode set to ClusterCIDR, but no cluster CIDR for family" ipFamily="IPv6"
	I0830 22:55:48.649920       1 server_others.go:438] "Defaulting to no-op detect-local"
	I0830 22:55:48.649971       1 proxier.go:251] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses"
	I0830 22:55:48.650284       1 server.go:846] "Version info" version="v1.28.1"
	I0830 22:55:48.650295       1 server.go:848] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
	I0830 22:55:48.652838       1 config.go:188] "Starting service config controller"
	I0830 22:55:48.652854       1 shared_informer.go:311] Waiting for caches to sync for service config
	I0830 22:55:48.652899       1 config.go:97] "Starting endpoint slice config controller"
	I0830 22:55:48.652904       1 shared_informer.go:311] Waiting for caches to sync for endpoint slice config
	I0830 22:55:48.661697       1 config.go:315] "Starting node config controller"
	I0830 22:55:48.661717       1 shared_informer.go:311] Waiting for caches to sync for node config
	I0830 22:55:48.753705       1 shared_informer.go:318] Caches are synced for endpoint slice config
	I0830 22:55:48.753789       1 shared_informer.go:318] Caches are synced for service config
	I0830 22:55:48.761966       1 shared_informer.go:318] Caches are synced for node config
	
	* 
	* ==> kube-scheduler [f8c96f4a21ae6c59bb1386e79dc6453d8a3bf23bfdf1ddae4df793a07bf1bdc2] <==
	* W0830 22:55:31.754771       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
	E0830 22:55:31.754789       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
	W0830 22:55:31.754844       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
	E0830 22:55:31.754853       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
	W0830 22:55:31.754893       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
	E0830 22:55:31.754909       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
	W0830 22:55:31.755836       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
	E0830 22:55:31.755951       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
	W0830 22:55:32.583799       1 reflector.go:535] pkg/server/dynamiccertificates/configmap_cafile_content.go:206: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
	E0830 22:55:32.583832       1 reflector.go:147] pkg/server/dynamiccertificates/configmap_cafile_content.go:206: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
	W0830 22:55:32.597610       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
	E0830 22:55:32.597655       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
	W0830 22:55:32.659885       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
	E0830 22:55:32.659927       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
	W0830 22:55:32.689468       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
	E0830 22:55:32.689511       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
	W0830 22:55:32.706333       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
	E0830 22:55:32.706369       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
	W0830 22:55:32.783443       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
	E0830 22:55:32.783668       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
	W0830 22:55:32.795179       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
	E0830 22:55:32.795225       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
	W0830 22:55:32.849315       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
	E0830 22:55:32.849529       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
	I0830 22:55:34.745079       1 shared_informer.go:318] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
	
	* 
	* ==> kubelet <==
	* Aug 30 22:58:01 addons-015166 kubelet[1354]: I0830 22:58:01.014832    1354 scope.go:117] "RemoveContainer" containerID="80f748d268e87bd588af505f507cdf45cbf7dfa96f0034073a28745085101bae"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: I0830 22:58:01.020968    1354 scope.go:117] "RemoveContainer" containerID="d7ea4da00dd6813494e731791191d666ebcce4f54bef73015e06138e6e9f348f"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: E0830 22:58:01.021666    1354 pod_workers.go:1300] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"hello-world-app\" with CrashLoopBackOff: \"back-off 20s restarting failed container=hello-world-app pod=hello-world-app-5d77478584-8crfr_default(e2b4ef71-b8ee-4e7b-972e-86d452125a20)\"" pod="default/hello-world-app-5d77478584-8crfr" podUID="e2b4ef71-b8ee-4e7b-972e-86d452125a20"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: I0830 22:58:01.029057    1354 scope.go:117] "RemoveContainer" containerID="80f748d268e87bd588af505f507cdf45cbf7dfa96f0034073a28745085101bae"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: E0830 22:58:01.033195    1354 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"80f748d268e87bd588af505f507cdf45cbf7dfa96f0034073a28745085101bae\": not found" containerID="80f748d268e87bd588af505f507cdf45cbf7dfa96f0034073a28745085101bae"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: I0830 22:58:01.033260    1354 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"80f748d268e87bd588af505f507cdf45cbf7dfa96f0034073a28745085101bae"} err="failed to get container status \"80f748d268e87bd588af505f507cdf45cbf7dfa96f0034073a28745085101bae\": rpc error: code = NotFound desc = an error occurred when try to find container \"80f748d268e87bd588af505f507cdf45cbf7dfa96f0034073a28745085101bae\": not found"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: I0830 22:58:01.033276    1354 scope.go:117] "RemoveContainer" containerID="dcec37596a0cbda8ab1539c81b507254842bebded4556dc1a624d0f69ed6dd74"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: I0830 22:58:01.154248    1354 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="d0f38189-0671-4529-ad81-669148906048" path="/var/lib/kubelet/pods/d0f38189-0671-4529-ad81-669148906048/volumes"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: I0830 22:58:01.940624    1354 topology_manager.go:215] "Topology Admit Handler" podUID="6b24d1c0-53ae-4f1b-a0cc-17528e535b21" podNamespace="default" podName="task-pv-pod-restore"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: E0830 22:58:01.941251    1354 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="f0723a40-2d31-4e36-9657-0c4e02b37524" containerName="minikube-ingress-dns"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: E0830 22:58:01.941357    1354 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="49b3a806-f215-4cb0-aa88-6644cff09819" containerName="task-pv-container"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: E0830 22:58:01.941444    1354 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="f0723a40-2d31-4e36-9657-0c4e02b37524" containerName="minikube-ingress-dns"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: E0830 22:58:01.941509    1354 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="f0723a40-2d31-4e36-9657-0c4e02b37524" containerName="minikube-ingress-dns"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: E0830 22:58:01.941575    1354 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d0f38189-0671-4529-ad81-669148906048" containerName="controller"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: I0830 22:58:01.941684    1354 memory_manager.go:346] "RemoveStaleState removing state" podUID="f0723a40-2d31-4e36-9657-0c4e02b37524" containerName="minikube-ingress-dns"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: I0830 22:58:01.941761    1354 memory_manager.go:346] "RemoveStaleState removing state" podUID="f0723a40-2d31-4e36-9657-0c4e02b37524" containerName="minikube-ingress-dns"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: I0830 22:58:01.941825    1354 memory_manager.go:346] "RemoveStaleState removing state" podUID="f0723a40-2d31-4e36-9657-0c4e02b37524" containerName="minikube-ingress-dns"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: I0830 22:58:01.941897    1354 memory_manager.go:346] "RemoveStaleState removing state" podUID="d0f38189-0671-4529-ad81-669148906048" containerName="controller"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: I0830 22:58:01.941972    1354 memory_manager.go:346] "RemoveStaleState removing state" podUID="49b3a806-f215-4cb0-aa88-6644cff09819" containerName="task-pv-container"
	Aug 30 22:58:01 addons-015166 kubelet[1354]: I0830 22:58:01.942031    1354 memory_manager.go:346] "RemoveStaleState removing state" podUID="f0723a40-2d31-4e36-9657-0c4e02b37524" containerName="minikube-ingress-dns"
	Aug 30 22:58:02 addons-015166 kubelet[1354]: I0830 22:58:02.098008    1354 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/6b24d1c0-53ae-4f1b-a0cc-17528e535b21-gcp-creds\") pod \"task-pv-pod-restore\" (UID: \"6b24d1c0-53ae-4f1b-a0cc-17528e535b21\") " pod="default/task-pv-pod-restore"
	Aug 30 22:58:02 addons-015166 kubelet[1354]: I0830 22:58:02.098309    1354 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np6bd\" (UniqueName: \"kubernetes.io/projected/6b24d1c0-53ae-4f1b-a0cc-17528e535b21-kube-api-access-np6bd\") pod \"task-pv-pod-restore\" (UID: \"6b24d1c0-53ae-4f1b-a0cc-17528e535b21\") " pod="default/task-pv-pod-restore"
	Aug 30 22:58:02 addons-015166 kubelet[1354]: I0830 22:58:02.098434    1354 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e211d7a2-807e-429b-bd2e-11bdf7346038\" (UniqueName: \"kubernetes.io/csi/hostpath.csi.k8s.io^ab40040b-4788-11ee-a776-f655b32dea15\") pod \"task-pv-pod-restore\" (UID: \"6b24d1c0-53ae-4f1b-a0cc-17528e535b21\") " pod="default/task-pv-pod-restore"
	Aug 30 22:58:02 addons-015166 kubelet[1354]: I0830 22:58:02.224640    1354 operation_generator.go:661] "MountVolume.MountDevice succeeded for volume \"pvc-e211d7a2-807e-429b-bd2e-11bdf7346038\" (UniqueName: \"kubernetes.io/csi/hostpath.csi.k8s.io^ab40040b-4788-11ee-a776-f655b32dea15\") pod \"task-pv-pod-restore\" (UID: \"6b24d1c0-53ae-4f1b-a0cc-17528e535b21\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/hostpath.csi.k8s.io/0d86c7dd4a36ed6421c0f70a846c120ea96814544bbd97faffd590c740cd6020/globalmount\"" pod="default/task-pv-pod-restore"
	Aug 30 22:58:03 addons-015166 kubelet[1354]: I0830 22:58:03.042158    1354 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/task-pv-pod-restore" podStartSLOduration=1.857028141 podCreationTimestamp="2023-08-30 22:58:01 +0000 UTC" firstStartedPulling="2023-08-30 22:58:02.446412558 +0000 UTC m=+147.453750745" lastFinishedPulling="2023-08-30 22:58:02.631497876 +0000 UTC m=+147.638836072" observedRunningTime="2023-08-30 22:58:03.041369453 +0000 UTC m=+148.048707649" watchObservedRunningTime="2023-08-30 22:58:03.042113468 +0000 UTC m=+148.049451656"
	
	* 
	* ==> storage-provisioner [9a2c72414e97ce1265d3ce584e33a045d4497ceeda30157f012d7d7c46ff1009] <==
	* I0830 22:55:52.894996       1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
	I0830 22:55:52.913839       1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
	I0830 22:55:52.913943       1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
	I0830 22:55:52.924172       1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
	I0830 22:55:52.925739       1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"2bf5f7d4-2bb7-413e-bc87-109068086fbd", APIVersion:"v1", ResourceVersion:"552", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-015166_d64edb70-8847-44f8-90a3-c93ec07970cd became leader
	I0830 22:55:52.926314       1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-015166_d64edb70-8847-44f8-90a3-c93ec07970cd!
	I0830 22:55:53.028856       1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-015166_d64edb70-8847-44f8-90a3-c93ec07970cd!
	

                                                
                                                
-- /stdout --
helpers_test.go:254: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p addons-015166 -n addons-015166
helpers_test.go:261: (dbg) Run:  kubectl --context addons-015166 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:285: <<< TestAddons/parallel/Ingress FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Ingress (37.98s)

                                                
                                    
x
+
TestAddons/parallel/InspektorGadget (8.59s)

                                                
                                                
=== RUN   TestAddons/parallel/InspektorGadget
=== PAUSE TestAddons/parallel/InspektorGadget

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/InspektorGadget
addons_test.go:814: (dbg) TestAddons/parallel/InspektorGadget: waiting 8m0s for pods matching "k8s-app=gadget" in namespace "gadget" ...
helpers_test.go:344: "gadget-khchg" [46e03991-96b2-427d-8a42-6611037cf432] Running
addons_test.go:814: (dbg) TestAddons/parallel/InspektorGadget: k8s-app=gadget healthy within 5.011863677s
addons_test.go:817: (dbg) Run:  out/minikube-linux-arm64 addons disable inspektor-gadget -p addons-015166
addons_test.go:817: (dbg) Non-zero exit: out/minikube-linux-arm64 addons disable inspektor-gadget -p addons-015166: exit status 11 (652.668467ms)

                                                
                                                
-- stdout --
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	X Exiting due to MK_ADDON_DISABLE_PAUSED: disable failed: check paused: list paused: runc: sudo runc --root /run/containerd/runc/k8s.io list -f json: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T22:58:12Z" level=error msg="stat /run/containerd/runc/k8s.io/43ed6efb3f3466e5878e9ae1a5dc109e5e48f86413e8e247e8142be68e980a71: no such file or directory"
	
	* 
	╭─────────────────────────────────────────────────────────────────────────────────────────────╮
	│                                                                                             │
	│    * If the above advice does not help, please let us know:                                 │
	│      https://github.com/kubernetes/minikube/issues/new/choose                               │
	│                                                                                             │
	│    * Please run `minikube logs --file=logs.txt` and attach logs.txt to the GitHub issue.    │
	│    * Please also attach the following file to the GitHub issue:                             │
	│    * - /tmp/minikube_addons_07218961934993dd21acc63caaf1aa08873c018e_0.log                  │
	│                                                                                             │
	╰─────────────────────────────────────────────────────────────────────────────────────────────╯

                                                
                                                
** /stderr **
addons_test.go:818: failed to disable inspektor-gadget addon: args "out/minikube-linux-arm64 addons disable inspektor-gadget -p addons-015166" : exit status 11
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======>  post-mortem[TestAddons/parallel/InspektorGadget]: docker inspect <======
helpers_test.go:231: (dbg) Run:  docker inspect addons-015166
helpers_test.go:235: (dbg) docker inspect addons-015166:

                                                
                                                
-- stdout --
	[
	    {
	        "Id": "a8fb941f1f0f953c7350d6b1c276cf4d522d308c1599eeae95f89ff74b160ef0",
	        "Created": "2023-08-30T22:55:07.556201355Z",
	        "Path": "/usr/local/bin/entrypoint",
	        "Args": [
	            "/sbin/init"
	        ],
	        "State": {
	            "Status": "running",
	            "Running": true,
	            "Paused": false,
	            "Restarting": false,
	            "OOMKilled": false,
	            "Dead": false,
	            "Pid": 1226334,
	            "ExitCode": 0,
	            "Error": "",
	            "StartedAt": "2023-08-30T22:55:07.889181596Z",
	            "FinishedAt": "0001-01-01T00:00:00Z"
	        },
	        "Image": "sha256:879c6efc994c345ac84dd4ebb4fc5b49dd2a4b340e335879382e51233f79b51a",
	        "ResolvConfPath": "/var/lib/docker/containers/a8fb941f1f0f953c7350d6b1c276cf4d522d308c1599eeae95f89ff74b160ef0/resolv.conf",
	        "HostnamePath": "/var/lib/docker/containers/a8fb941f1f0f953c7350d6b1c276cf4d522d308c1599eeae95f89ff74b160ef0/hostname",
	        "HostsPath": "/var/lib/docker/containers/a8fb941f1f0f953c7350d6b1c276cf4d522d308c1599eeae95f89ff74b160ef0/hosts",
	        "LogPath": "/var/lib/docker/containers/a8fb941f1f0f953c7350d6b1c276cf4d522d308c1599eeae95f89ff74b160ef0/a8fb941f1f0f953c7350d6b1c276cf4d522d308c1599eeae95f89ff74b160ef0-json.log",
	        "Name": "/addons-015166",
	        "RestartCount": 0,
	        "Driver": "overlay2",
	        "Platform": "linux",
	        "MountLabel": "",
	        "ProcessLabel": "",
	        "AppArmorProfile": "unconfined",
	        "ExecIDs": null,
	        "HostConfig": {
	            "Binds": [
	                "addons-015166:/var",
	                "/lib/modules:/lib/modules:ro"
	            ],
	            "ContainerIDFile": "",
	            "LogConfig": {
	                "Type": "json-file",
	                "Config": {}
	            },
	            "NetworkMode": "addons-015166",
	            "PortBindings": {
	                "22/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "2376/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "32443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "5000/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "8443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ]
	            },
	            "RestartPolicy": {
	                "Name": "no",
	                "MaximumRetryCount": 0
	            },
	            "AutoRemove": false,
	            "VolumeDriver": "",
	            "VolumesFrom": null,
	            "ConsoleSize": [
	                0,
	                0
	            ],
	            "CapAdd": null,
	            "CapDrop": null,
	            "CgroupnsMode": "host",
	            "Dns": [],
	            "DnsOptions": [],
	            "DnsSearch": [],
	            "ExtraHosts": null,
	            "GroupAdd": null,
	            "IpcMode": "private",
	            "Cgroup": "",
	            "Links": null,
	            "OomScoreAdj": 0,
	            "PidMode": "",
	            "Privileged": true,
	            "PublishAllPorts": false,
	            "ReadonlyRootfs": false,
	            "SecurityOpt": [
	                "seccomp=unconfined",
	                "apparmor=unconfined",
	                "label=disable"
	            ],
	            "Tmpfs": {
	                "/run": "",
	                "/tmp": ""
	            },
	            "UTSMode": "",
	            "UsernsMode": "",
	            "ShmSize": 67108864,
	            "Runtime": "runc",
	            "Isolation": "",
	            "CpuShares": 0,
	            "Memory": 4194304000,
	            "NanoCpus": 2000000000,
	            "CgroupParent": "",
	            "BlkioWeight": 0,
	            "BlkioWeightDevice": [],
	            "BlkioDeviceReadBps": [],
	            "BlkioDeviceWriteBps": [],
	            "BlkioDeviceReadIOps": [],
	            "BlkioDeviceWriteIOps": [],
	            "CpuPeriod": 0,
	            "CpuQuota": 0,
	            "CpuRealtimePeriod": 0,
	            "CpuRealtimeRuntime": 0,
	            "CpusetCpus": "",
	            "CpusetMems": "",
	            "Devices": [],
	            "DeviceCgroupRules": null,
	            "DeviceRequests": null,
	            "MemoryReservation": 0,
	            "MemorySwap": 8388608000,
	            "MemorySwappiness": null,
	            "OomKillDisable": false,
	            "PidsLimit": null,
	            "Ulimits": null,
	            "CpuCount": 0,
	            "CpuPercent": 0,
	            "IOMaximumIOps": 0,
	            "IOMaximumBandwidth": 0,
	            "MaskedPaths": null,
	            "ReadonlyPaths": null
	        },
	        "GraphDriver": {
	            "Data": {
	                "LowerDir": "/var/lib/docker/overlay2/a319c15f302f306a6e594fea92e83e1a7e485ec90885dcec209d164efbefcd4f-init/diff:/var/lib/docker/overlay2/ecca48d9e9e3ae5c70dec79bae341547fcaf05f85377796ae2cd1184b3d7168f/diff",
	                "MergedDir": "/var/lib/docker/overlay2/a319c15f302f306a6e594fea92e83e1a7e485ec90885dcec209d164efbefcd4f/merged",
	                "UpperDir": "/var/lib/docker/overlay2/a319c15f302f306a6e594fea92e83e1a7e485ec90885dcec209d164efbefcd4f/diff",
	                "WorkDir": "/var/lib/docker/overlay2/a319c15f302f306a6e594fea92e83e1a7e485ec90885dcec209d164efbefcd4f/work"
	            },
	            "Name": "overlay2"
	        },
	        "Mounts": [
	            {
	                "Type": "volume",
	                "Name": "addons-015166",
	                "Source": "/var/lib/docker/volumes/addons-015166/_data",
	                "Destination": "/var",
	                "Driver": "local",
	                "Mode": "z",
	                "RW": true,
	                "Propagation": ""
	            },
	            {
	                "Type": "bind",
	                "Source": "/lib/modules",
	                "Destination": "/lib/modules",
	                "Mode": "ro",
	                "RW": false,
	                "Propagation": "rprivate"
	            }
	        ],
	        "Config": {
	            "Hostname": "addons-015166",
	            "Domainname": "",
	            "User": "",
	            "AttachStdin": false,
	            "AttachStdout": false,
	            "AttachStderr": false,
	            "ExposedPorts": {
	                "22/tcp": {},
	                "2376/tcp": {},
	                "32443/tcp": {},
	                "5000/tcp": {},
	                "8443/tcp": {}
	            },
	            "Tty": true,
	            "OpenStdin": false,
	            "StdinOnce": false,
	            "Env": [
	                "container=docker",
	                "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
	            ],
	            "Cmd": null,
	            "Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec",
	            "Volumes": null,
	            "WorkingDir": "/",
	            "Entrypoint": [
	                "/usr/local/bin/entrypoint",
	                "/sbin/init"
	            ],
	            "OnBuild": null,
	            "Labels": {
	                "created_by.minikube.sigs.k8s.io": "true",
	                "mode.minikube.sigs.k8s.io": "addons-015166",
	                "name.minikube.sigs.k8s.io": "addons-015166",
	                "role.minikube.sigs.k8s.io": ""
	            },
	            "StopSignal": "SIGRTMIN+3"
	        },
	        "NetworkSettings": {
	            "Bridge": "",
	            "SandboxID": "fc88c85ba1ec60a1c5079c503f3fe903dd9ebe6885d4e7935ac796e10deea0f2",
	            "HairpinMode": false,
	            "LinkLocalIPv6Address": "",
	            "LinkLocalIPv6PrefixLen": 0,
	            "Ports": {
	                "22/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34314"
	                    }
	                ],
	                "2376/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34313"
	                    }
	                ],
	                "32443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34310"
	                    }
	                ],
	                "5000/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34312"
	                    }
	                ],
	                "8443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34311"
	                    }
	                ]
	            },
	            "SandboxKey": "/var/run/docker/netns/fc88c85ba1ec",
	            "SecondaryIPAddresses": null,
	            "SecondaryIPv6Addresses": null,
	            "EndpointID": "",
	            "Gateway": "",
	            "GlobalIPv6Address": "",
	            "GlobalIPv6PrefixLen": 0,
	            "IPAddress": "",
	            "IPPrefixLen": 0,
	            "IPv6Gateway": "",
	            "MacAddress": "",
	            "Networks": {
	                "addons-015166": {
	                    "IPAMConfig": {
	                        "IPv4Address": "192.168.49.2"
	                    },
	                    "Links": null,
	                    "Aliases": [
	                        "a8fb941f1f0f",
	                        "addons-015166"
	                    ],
	                    "NetworkID": "e4b094cfd79395f0e8d6b44a20c7acdc584d8c0b0e8f44f096b4b8ac417f3d8a",
	                    "EndpointID": "48a430ee68952ef9f1764d82d16ddd42052970c53df7fb70b2c0a91fe225a23d",
	                    "Gateway": "192.168.49.1",
	                    "IPAddress": "192.168.49.2",
	                    "IPPrefixLen": 24,
	                    "IPv6Gateway": "",
	                    "GlobalIPv6Address": "",
	                    "GlobalIPv6PrefixLen": 0,
	                    "MacAddress": "02:42:c0:a8:31:02",
	                    "DriverOpts": null
	                }
	            }
	        }
	    }
	]

                                                
                                                
-- /stdout --
helpers_test.go:239: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p addons-015166 -n addons-015166
helpers_test.go:244: <<< TestAddons/parallel/InspektorGadget FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======>  post-mortem[TestAddons/parallel/InspektorGadget]: minikube logs <======
helpers_test.go:247: (dbg) Run:  out/minikube-linux-arm64 -p addons-015166 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-arm64 -p addons-015166 logs -n 25: (1.676064133s)
helpers_test.go:252: TestAddons/parallel/InspektorGadget logs: 
-- stdout --
	* 
	* ==> Audit <==
	* |---------|--------------------------------|------------------------|---------|---------|---------------------|---------------------|
	| Command |              Args              |        Profile         |  User   | Version |     Start Time      |      End Time       |
	|---------|--------------------------------|------------------------|---------|---------|---------------------|---------------------|
	| start   | -o=json --download-only        | download-only-635688   | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC |                     |
	|         | -p download-only-635688        |                        |         |         |                     |                     |
	|         | --force --alsologtostderr      |                        |         |         |                     |                     |
	|         | --kubernetes-version=v1.16.0   |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	|         | --driver=docker                |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	| start   | -o=json --download-only        | download-only-635688   | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC |                     |
	|         | -p download-only-635688        |                        |         |         |                     |                     |
	|         | --force --alsologtostderr      |                        |         |         |                     |                     |
	|         | --kubernetes-version=v1.28.1   |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	|         | --driver=docker                |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	| delete  | --all                          | minikube               | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC | 30 Aug 23 22:54 UTC |
	| delete  | -p download-only-635688        | download-only-635688   | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC | 30 Aug 23 22:54 UTC |
	| delete  | -p download-only-635688        | download-only-635688   | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC | 30 Aug 23 22:54 UTC |
	| start   | --download-only -p             | download-docker-266424 | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC |                     |
	|         | download-docker-266424         |                        |         |         |                     |                     |
	|         | --alsologtostderr              |                        |         |         |                     |                     |
	|         | --driver=docker                |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	| delete  | -p download-docker-266424      | download-docker-266424 | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC | 30 Aug 23 22:54 UTC |
	| start   | --download-only -p             | binary-mirror-621146   | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC |                     |
	|         | binary-mirror-621146           |                        |         |         |                     |                     |
	|         | --alsologtostderr              |                        |         |         |                     |                     |
	|         | --binary-mirror                |                        |         |         |                     |                     |
	|         | http://127.0.0.1:44437         |                        |         |         |                     |                     |
	|         | --driver=docker                |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	| delete  | -p binary-mirror-621146        | binary-mirror-621146   | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC | 30 Aug 23 22:54 UTC |
	| start   | -p addons-015166               | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC | 30 Aug 23 22:57 UTC |
	|         | --wait=true --memory=4000      |                        |         |         |                     |                     |
	|         | --alsologtostderr              |                        |         |         |                     |                     |
	|         | --addons=registry              |                        |         |         |                     |                     |
	|         | --addons=metrics-server        |                        |         |         |                     |                     |
	|         | --addons=volumesnapshots       |                        |         |         |                     |                     |
	|         | --addons=csi-hostpath-driver   |                        |         |         |                     |                     |
	|         | --addons=gcp-auth              |                        |         |         |                     |                     |
	|         | --addons=cloud-spanner         |                        |         |         |                     |                     |
	|         | --addons=inspektor-gadget      |                        |         |         |                     |                     |
	|         | --driver=docker                |                        |         |         |                     |                     |
	|         | --container-runtime=containerd |                        |         |         |                     |                     |
	|         | --addons=ingress               |                        |         |         |                     |                     |
	|         | --addons=ingress-dns           |                        |         |         |                     |                     |
	| addons  | disable cloud-spanner -p       | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	|         | addons-015166                  |                        |         |         |                     |                     |
	| addons  | addons-015166 addons           | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	|         | disable metrics-server         |                        |         |         |                     |                     |
	|         | --alsologtostderr -v=1         |                        |         |         |                     |                     |
	| addons  | enable headlamp                | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	|         | -p addons-015166               |                        |         |         |                     |                     |
	|         | --alsologtostderr -v=1         |                        |         |         |                     |                     |
	| ip      | addons-015166 ip               | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	| addons  | addons-015166 addons disable   | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	|         | registry --alsologtostderr     |                        |         |         |                     |                     |
	|         | -v=1                           |                        |         |         |                     |                     |
	| ssh     | addons-015166 ssh curl -s      | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	|         | http://127.0.0.1/ -H 'Host:    |                        |         |         |                     |                     |
	|         | nginx.example.com'             |                        |         |         |                     |                     |
	| ip      | addons-015166 ip               | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	| addons  | addons-015166 addons disable   | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:57 UTC |
	|         | ingress-dns --alsologtostderr  |                        |         |         |                     |                     |
	|         | -v=1                           |                        |         |         |                     |                     |
	| addons  | addons-015166 addons disable   | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:57 UTC | 30 Aug 23 22:58 UTC |
	|         | ingress --alsologtostderr -v=1 |                        |         |         |                     |                     |
	| addons  | addons-015166 addons           | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:58 UTC |                     |
	|         | disable csi-hostpath-driver    |                        |         |         |                     |                     |
	|         | --alsologtostderr -v=1         |                        |         |         |                     |                     |
	| addons  | disable inspektor-gadget -p    | addons-015166          | jenkins | v1.31.2 | 30 Aug 23 22:58 UTC |                     |
	|         | addons-015166                  |                        |         |         |                     |                     |
	|---------|--------------------------------|------------------------|---------|---------|---------------------|---------------------|
	
	* 
	* ==> Last Start <==
	* Log file created at: 2023/08/30 22:54:44
	Running on machine: ip-172-31-31-251
	Binary: Built with gc go1.20.7 for linux/arm64
	Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
	I0830 22:54:44.288127 1225867 out.go:296] Setting OutFile to fd 1 ...
	I0830 22:54:44.288286 1225867 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 22:54:44.288296 1225867 out.go:309] Setting ErrFile to fd 2...
	I0830 22:54:44.288301 1225867 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 22:54:44.288579 1225867 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
	I0830 22:54:44.289010 1225867 out.go:303] Setting JSON to false
	I0830 22:54:44.289885 1225867 start.go:128] hostinfo: {"hostname":"ip-172-31-31-251","uptime":27419,"bootTime":1693408666,"procs":183,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1043-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"982e3628-3742-4b3e-bb63-ac1b07660ec7"}
	I0830 22:54:44.289953 1225867 start.go:138] virtualization:  
	I0830 22:54:44.293534 1225867 out.go:177] * [addons-015166] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	I0830 22:54:44.295256 1225867 out.go:177]   - MINIKUBE_LOCATION=17114
	I0830 22:54:44.297092 1225867 out.go:177]   - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0830 22:54:44.295415 1225867 notify.go:220] Checking for updates...
	I0830 22:54:44.300354 1225867 out.go:177]   - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	I0830 22:54:44.302092 1225867 out.go:177]   - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	I0830 22:54:44.304093 1225867 out.go:177]   - MINIKUBE_BIN=out/minikube-linux-arm64
	I0830 22:54:44.305841 1225867 out.go:177]   - MINIKUBE_FORCE_SYSTEMD=
	I0830 22:54:44.307896 1225867 driver.go:373] Setting default libvirt URI to qemu:///system
	I0830 22:54:44.333491 1225867 docker.go:121] docker version: linux-24.0.5:Docker Engine - Community
	I0830 22:54:44.333600 1225867 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 22:54:44.425503 1225867 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:28 OomKillDisable:true NGoroutines:38 SystemTime:2023-08-30 22:54:44.414894496 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 22:54:44.425636 1225867 docker.go:294] overlay module found
	I0830 22:54:44.429303 1225867 out.go:177] * Using the docker driver based on user configuration
	I0830 22:54:44.430880 1225867 start.go:298] selected driver: docker
	I0830 22:54:44.430905 1225867 start.go:902] validating driver "docker" against <nil>
	I0830 22:54:44.430935 1225867 start.go:913] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
	I0830 22:54:44.431574 1225867 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 22:54:44.505406 1225867 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:28 OomKillDisable:true NGoroutines:38 SystemTime:2023-08-30 22:54:44.495875074 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 22:54:44.505563 1225867 start_flags.go:305] no existing cluster config was found, will generate one from the flags 
	I0830 22:54:44.505779 1225867 start_flags.go:919] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
	I0830 22:54:44.507592 1225867 out.go:177] * Using Docker driver with root privileges
	I0830 22:54:44.509430 1225867 cni.go:84] Creating CNI manager for ""
	I0830 22:54:44.509462 1225867 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
	I0830 22:54:44.509473 1225867 start_flags.go:314] Found "CNI" CNI - setting NetworkPlugin=cni
	I0830 22:54:44.509488 1225867 start_flags.go:319] config:
	{Name:addons-015166 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:4000 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.28.1 ClusterName:addons-015166 Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containe
rd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0}
	I0830 22:54:44.511549 1225867 out.go:177] * Starting control plane node addons-015166 in cluster addons-015166
	I0830 22:54:44.513090 1225867 cache.go:122] Beginning downloading kic base image for docker with containerd
	I0830 22:54:44.514803 1225867 out.go:177] * Pulling base image ...
	I0830 22:54:44.516427 1225867 preload.go:132] Checking if preload exists for k8s version v1.28.1 and runtime containerd
	I0830 22:54:44.516481 1225867 preload.go:148] Found local preload: /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.28.1-containerd-overlay2-arm64.tar.lz4
	I0830 22:54:44.516503 1225867 cache.go:57] Caching tarball of preloaded images
	I0830 22:54:44.516511 1225867 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec in local docker daemon
	I0830 22:54:44.516585 1225867 preload.go:174] Found /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.28.1-containerd-overlay2-arm64.tar.lz4 in cache, skipping download
	I0830 22:54:44.516595 1225867 cache.go:60] Finished verifying existence of preloaded tar for  v1.28.1 on containerd
	I0830 22:54:44.517014 1225867 profile.go:148] Saving config to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/config.json ...
	I0830 22:54:44.517034 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/config.json: {Name:mkc799c1e3aadbb7e65ebbcba8e932055b83a00b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:54:44.536561 1225867 cache.go:150] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec to local cache
	I0830 22:54:44.536714 1225867 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec in local cache directory
	I0830 22:54:44.536745 1225867 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec in local cache directory, skipping pull
	I0830 22:54:44.536750 1225867 image.go:105] gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec exists in cache, skipping pull
	I0830 22:54:44.536758 1225867 cache.go:153] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec as a tarball
	I0830 22:54:44.536776 1225867 cache.go:163] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec from local cache
	I0830 22:55:00.495233 1225867 cache.go:165] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec from cached tarball
	I0830 22:55:00.495270 1225867 cache.go:195] Successfully downloaded all kic artifacts
	I0830 22:55:00.495324 1225867 start.go:365] acquiring machines lock for addons-015166: {Name:mkd1caf7e2d7c8b873baf6be8f0355d0884d5b9c Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
	I0830 22:55:00.495457 1225867 start.go:369] acquired machines lock for "addons-015166" in 114.363µs
	I0830 22:55:00.495484 1225867 start.go:93] Provisioning new machine with config: &{Name:addons-015166 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:4000 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.28.1 ClusterName:addons-015166 Namespace:default APIServerName:minikubeCA A
PIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.28.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:
false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0} &{Name: IP: Port:8443 KubernetesVersion:v1.28.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
	I0830 22:55:00.495586 1225867 start.go:125] createHost starting for "" (driver="docker")
	I0830 22:55:00.497857 1225867 out.go:204] * Creating docker container (CPUs=2, Memory=4000MB) ...
	I0830 22:55:00.498120 1225867 start.go:159] libmachine.API.Create for "addons-015166" (driver="docker")
	I0830 22:55:00.498152 1225867 client.go:168] LocalClient.Create starting
	I0830 22:55:00.498277 1225867 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem
	I0830 22:55:00.918608 1225867 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem
	I0830 22:55:01.212104 1225867 cli_runner.go:164] Run: docker network inspect addons-015166 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	W0830 22:55:01.231323 1225867 cli_runner.go:211] docker network inspect addons-015166 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
	I0830 22:55:01.231415 1225867 network_create.go:281] running [docker network inspect addons-015166] to gather additional debugging logs...
	I0830 22:55:01.231435 1225867 cli_runner.go:164] Run: docker network inspect addons-015166
	W0830 22:55:01.254522 1225867 cli_runner.go:211] docker network inspect addons-015166 returned with exit code 1
	I0830 22:55:01.254568 1225867 network_create.go:284] error running [docker network inspect addons-015166]: docker network inspect addons-015166: exit status 1
	stdout:
	[]
	
	stderr:
	Error response from daemon: network addons-015166 not found
	I0830 22:55:01.254581 1225867 network_create.go:286] output of [docker network inspect addons-015166]: -- stdout --
	[]
	
	-- /stdout --
	** stderr ** 
	Error response from daemon: network addons-015166 not found
	
	** /stderr **
	I0830 22:55:01.254649 1225867 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	I0830 22:55:01.274472 1225867 network.go:209] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x40028627d0}
	I0830 22:55:01.274511 1225867 network_create.go:123] attempt to create docker network addons-015166 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
	I0830 22:55:01.274576 1225867 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-015166 addons-015166
	I0830 22:55:01.348031 1225867 network_create.go:107] docker network addons-015166 192.168.49.0/24 created
	I0830 22:55:01.348090 1225867 kic.go:117] calculated static IP "192.168.49.2" for the "addons-015166" container
	I0830 22:55:01.348166 1225867 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
	I0830 22:55:01.365657 1225867 cli_runner.go:164] Run: docker volume create addons-015166 --label name.minikube.sigs.k8s.io=addons-015166 --label created_by.minikube.sigs.k8s.io=true
	I0830 22:55:01.384685 1225867 oci.go:103] Successfully created a docker volume addons-015166
	I0830 22:55:01.384816 1225867 cli_runner.go:164] Run: docker run --rm --name addons-015166-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-015166 --entrypoint /usr/bin/test -v addons-015166:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec -d /var/lib
	I0830 22:55:03.279386 1225867 cli_runner.go:217] Completed: docker run --rm --name addons-015166-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-015166 --entrypoint /usr/bin/test -v addons-015166:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec -d /var/lib: (1.89451146s)
	I0830 22:55:03.279422 1225867 oci.go:107] Successfully prepared a docker volume addons-015166
	I0830 22:55:03.279449 1225867 preload.go:132] Checking if preload exists for k8s version v1.28.1 and runtime containerd
	I0830 22:55:03.279469 1225867 kic.go:190] Starting extracting preloaded images to volume ...
	I0830 22:55:03.279564 1225867 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.28.1-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v addons-015166:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec -I lz4 -xf /preloaded.tar -C /extractDir
	I0830 22:55:07.469719 1225867 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.28.1-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v addons-015166:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec -I lz4 -xf /preloaded.tar -C /extractDir: (4.190097173s)
	I0830 22:55:07.469750 1225867 kic.go:199] duration metric: took 4.190276 seconds to extract preloaded images to volume
	W0830 22:55:07.469900 1225867 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
	I0830 22:55:07.470010 1225867 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
	I0830 22:55:07.539204 1225867 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-015166 --name addons-015166 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-015166 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-015166 --network addons-015166 --ip 192.168.49.2 --volume addons-015166:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec
	I0830 22:55:07.900476 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Running}}
	I0830 22:55:07.925054 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:07.960094 1225867 cli_runner.go:164] Run: docker exec addons-015166 stat /var/lib/dpkg/alternatives/iptables
	I0830 22:55:08.055683 1225867 oci.go:144] the created container "addons-015166" has a running status.
	I0830 22:55:08.055709 1225867 kic.go:221] Creating ssh key for kic: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa...
	I0830 22:55:08.520104 1225867 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
	I0830 22:55:08.558856 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:08.584595 1225867 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
	I0830 22:55:08.584614 1225867 kic_runner.go:114] Args: [docker exec --privileged addons-015166 chown docker:docker /home/docker/.ssh/authorized_keys]
	I0830 22:55:08.686413 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:08.721361 1225867 machine.go:88] provisioning docker machine ...
	I0830 22:55:08.721393 1225867 ubuntu.go:169] provisioning hostname "addons-015166"
	I0830 22:55:08.721459 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:08.750886 1225867 main.go:141] libmachine: Using SSH client type: native
	I0830 22:55:08.751366 1225867 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34314 <nil> <nil>}
	I0830 22:55:08.751384 1225867 main.go:141] libmachine: About to run SSH command:
	sudo hostname addons-015166 && echo "addons-015166" | sudo tee /etc/hostname
	I0830 22:55:08.967147 1225867 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-015166
	
	I0830 22:55:08.967231 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:08.987891 1225867 main.go:141] libmachine: Using SSH client type: native
	I0830 22:55:08.988364 1225867 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34314 <nil> <nil>}
	I0830 22:55:08.988390 1225867 main.go:141] libmachine: About to run SSH command:
	
			if ! grep -xq '.*\saddons-015166' /etc/hosts; then
				if grep -xq '127.0.1.1\s.*' /etc/hosts; then
					sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-015166/g' /etc/hosts;
				else 
					echo '127.0.1.1 addons-015166' | sudo tee -a /etc/hosts; 
				fi
			fi
	I0830 22:55:09.147200 1225867 main.go:141] libmachine: SSH cmd err, output: <nil>: 
	I0830 22:55:09.147226 1225867 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/17114-1219981/.minikube CaCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/17114-1219981/.minikube}
	I0830 22:55:09.147249 1225867 ubuntu.go:177] setting up certificates
	I0830 22:55:09.147258 1225867 provision.go:83] configureAuth start
	I0830 22:55:09.147321 1225867 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-015166
	I0830 22:55:09.173237 1225867 provision.go:138] copyHostCerts
	I0830 22:55:09.173312 1225867 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem (1082 bytes)
	I0830 22:55:09.173433 1225867 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem (1123 bytes)
	I0830 22:55:09.173499 1225867 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem (1679 bytes)
	I0830 22:55:09.173556 1225867 provision.go:112] generating server cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem org=jenkins.addons-015166 san=[192.168.49.2 127.0.0.1 localhost 127.0.0.1 minikube addons-015166]
	I0830 22:55:11.969349 1225867 provision.go:172] copyRemoteCerts
	I0830 22:55:11.969424 1225867 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
	I0830 22:55:11.969473 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:11.987831 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:12.095177 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
	I0830 22:55:12.126863 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
	I0830 22:55:12.157479 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem --> /etc/docker/server.pem (1216 bytes)
	I0830 22:55:12.187894 1225867 provision.go:86] duration metric: configureAuth took 3.040615392s
	I0830 22:55:12.187926 1225867 ubuntu.go:193] setting minikube options for container-runtime
	I0830 22:55:12.188168 1225867 config.go:182] Loaded profile config "addons-015166": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
	I0830 22:55:12.188181 1225867 machine.go:91] provisioned docker machine in 3.466799986s
	I0830 22:55:12.188187 1225867 client.go:171] LocalClient.Create took 11.690030577s
	I0830 22:55:12.188216 1225867 start.go:167] duration metric: libmachine.API.Create for "addons-015166" took 11.690097513s
	I0830 22:55:12.188241 1225867 start.go:300] post-start starting for "addons-015166" (driver="docker")
	I0830 22:55:12.188250 1225867 start.go:329] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
	I0830 22:55:12.188336 1225867 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
	I0830 22:55:12.188381 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:12.206826 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:12.313851 1225867 ssh_runner.go:195] Run: cat /etc/os-release
	I0830 22:55:12.318474 1225867 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
	I0830 22:55:12.318515 1225867 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
	I0830 22:55:12.318527 1225867 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
	I0830 22:55:12.318534 1225867 info.go:137] Remote host: Ubuntu 22.04.3 LTS
	I0830 22:55:12.318544 1225867 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/addons for local assets ...
	I0830 22:55:12.318613 1225867 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/files for local assets ...
	I0830 22:55:12.318641 1225867 start.go:303] post-start completed in 130.394084ms
	I0830 22:55:12.318976 1225867 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-015166
	I0830 22:55:12.337680 1225867 profile.go:148] Saving config to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/config.json ...
	I0830 22:55:12.337979 1225867 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0830 22:55:12.338030 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:12.356290 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:12.459405 1225867 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
	I0830 22:55:12.465439 1225867 start.go:128] duration metric: createHost completed in 11.969838317s
	I0830 22:55:12.465462 1225867 start.go:83] releasing machines lock for "addons-015166", held for 11.969996611s
	I0830 22:55:12.465534 1225867 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-015166
	I0830 22:55:12.483132 1225867 ssh_runner.go:195] Run: cat /version.json
	I0830 22:55:12.483184 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:12.483206 1225867 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
	I0830 22:55:12.483306 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:12.507749 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:12.509782 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:12.610051 1225867 ssh_runner.go:195] Run: systemctl --version
	I0830 22:55:12.743718 1225867 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
	I0830 22:55:12.749595 1225867 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
	I0830 22:55:12.780799 1225867 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
	I0830 22:55:12.780879 1225867 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%!p(MISSING), " -exec sh -c "sudo mv {} {}.mk_disabled" ;
	I0830 22:55:12.815904 1225867 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
	I0830 22:55:12.815930 1225867 start.go:466] detecting cgroup driver to use...
	I0830 22:55:12.815961 1225867 detect.go:196] detected "cgroupfs" cgroup driver on host os
	I0830 22:55:12.816028 1225867 ssh_runner.go:195] Run: sudo systemctl stop -f crio
	I0830 22:55:12.830893 1225867 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
	I0830 22:55:12.844398 1225867 docker.go:196] disabling cri-docker service (if available) ...
	I0830 22:55:12.844462 1225867 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
	I0830 22:55:12.860658 1225867 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
	I0830 22:55:12.877580 1225867 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
	I0830 22:55:12.974862 1225867 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
	I0830 22:55:13.076455 1225867 docker.go:212] disabling docker service ...
	I0830 22:55:13.076559 1225867 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
	I0830 22:55:13.100824 1225867 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
	I0830 22:55:13.115486 1225867 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
	I0830 22:55:13.208014 1225867 ssh_runner.go:195] Run: sudo systemctl mask docker.service
	I0830 22:55:13.312758 1225867 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
	I0830 22:55:13.326519 1225867 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///run/containerd/containerd.sock
	" | sudo tee /etc/crictl.yaml"
	I0830 22:55:13.346444 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.9"|' /etc/containerd/config.toml"
	I0830 22:55:13.358491 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
	I0830 22:55:13.370743 1225867 containerd.go:145] configuring containerd to use "cgroupfs" as cgroup driver...
	I0830 22:55:13.370850 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
	I0830 22:55:13.383098 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
	I0830 22:55:13.395209 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
	I0830 22:55:13.407233 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
	I0830 22:55:13.419163 1225867 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
	I0830 22:55:13.430722 1225867 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
	I0830 22:55:13.442720 1225867 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
	I0830 22:55:13.453316 1225867 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
	I0830 22:55:13.464475 1225867 ssh_runner.go:195] Run: sudo systemctl daemon-reload
	I0830 22:55:13.564609 1225867 ssh_runner.go:195] Run: sudo systemctl restart containerd
	I0830 22:55:13.702714 1225867 start.go:513] Will wait 60s for socket path /run/containerd/containerd.sock
	I0830 22:55:13.702833 1225867 ssh_runner.go:195] Run: stat /run/containerd/containerd.sock
	I0830 22:55:13.707912 1225867 start.go:534] Will wait 60s for crictl version
	I0830 22:55:13.708009 1225867 ssh_runner.go:195] Run: which crictl
	I0830 22:55:13.712710 1225867 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
	I0830 22:55:13.762208 1225867 start.go:550] Version:  0.1.0
	RuntimeName:  containerd
	RuntimeVersion:  1.6.22
	RuntimeApiVersion:  v1
	I0830 22:55:13.762349 1225867 ssh_runner.go:195] Run: containerd --version
	I0830 22:55:13.793307 1225867 ssh_runner.go:195] Run: containerd --version
	I0830 22:55:13.830513 1225867 out.go:177] * Preparing Kubernetes v1.28.1 on containerd 1.6.22 ...
	I0830 22:55:13.832626 1225867 cli_runner.go:164] Run: docker network inspect addons-015166 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	I0830 22:55:13.854629 1225867 ssh_runner.go:195] Run: grep 192.168.49.1	host.minikube.internal$ /etc/hosts
	I0830 22:55:13.859439 1225867 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1	host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
	I0830 22:55:13.874109 1225867 preload.go:132] Checking if preload exists for k8s version v1.28.1 and runtime containerd
	I0830 22:55:13.874184 1225867 ssh_runner.go:195] Run: sudo crictl images --output json
	I0830 22:55:13.920916 1225867 containerd.go:604] all images are preloaded for containerd runtime.
	I0830 22:55:13.920939 1225867 containerd.go:518] Images already preloaded, skipping extraction
	I0830 22:55:13.921000 1225867 ssh_runner.go:195] Run: sudo crictl images --output json
	I0830 22:55:13.966563 1225867 containerd.go:604] all images are preloaded for containerd runtime.
	I0830 22:55:13.966585 1225867 cache_images.go:84] Images are preloaded, skipping loading
	I0830 22:55:13.966642 1225867 ssh_runner.go:195] Run: sudo crictl info
	I0830 22:55:14.012144 1225867 cni.go:84] Creating CNI manager for ""
	I0830 22:55:14.012166 1225867 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
	I0830 22:55:14.012198 1225867 kubeadm.go:87] Using pod CIDR: 10.244.0.0/16
	I0830 22:55:14.012221 1225867 kubeadm.go:176] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.28.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-015166 NodeName:addons-015166 DNSDomain:cluster.local CRISocket:/run/containerd/containerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc
/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
	I0830 22:55:14.012366 1225867 kubeadm.go:181] kubeadm config:
	apiVersion: kubeadm.k8s.io/v1beta3
	kind: InitConfiguration
	localAPIEndpoint:
	  advertiseAddress: 192.168.49.2
	  bindPort: 8443
	bootstrapTokens:
	  - groups:
	      - system:bootstrappers:kubeadm:default-node-token
	    ttl: 24h0m0s
	    usages:
	      - signing
	      - authentication
	nodeRegistration:
	  criSocket: unix:///run/containerd/containerd.sock
	  name: "addons-015166"
	  kubeletExtraArgs:
	    node-ip: 192.168.49.2
	  taints: []
	---
	apiVersion: kubeadm.k8s.io/v1beta3
	kind: ClusterConfiguration
	apiServer:
	  certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
	  extraArgs:
	    enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
	controllerManager:
	  extraArgs:
	    allocate-node-cidrs: "true"
	    leader-elect: "false"
	scheduler:
	  extraArgs:
	    leader-elect: "false"
	certificatesDir: /var/lib/minikube/certs
	clusterName: mk
	controlPlaneEndpoint: control-plane.minikube.internal:8443
	etcd:
	  local:
	    dataDir: /var/lib/minikube/etcd
	    extraArgs:
	      proxy-refresh-interval: "70000"
	kubernetesVersion: v1.28.1
	networking:
	  dnsDomain: cluster.local
	  podSubnet: "10.244.0.0/16"
	  serviceSubnet: 10.96.0.0/12
	---
	apiVersion: kubelet.config.k8s.io/v1beta1
	kind: KubeletConfiguration
	authentication:
	  x509:
	    clientCAFile: /var/lib/minikube/certs/ca.crt
	cgroupDriver: cgroupfs
	hairpinMode: hairpin-veth
	runtimeRequestTimeout: 15m
	clusterDomain: "cluster.local"
	# disable disk resource management by default
	imageGCHighThresholdPercent: 100
	evictionHard:
	  nodefs.available: "0%!"(MISSING)
	  nodefs.inodesFree: "0%!"(MISSING)
	  imagefs.available: "0%!"(MISSING)
	failSwapOn: false
	staticPodPath: /etc/kubernetes/manifests
	---
	apiVersion: kubeproxy.config.k8s.io/v1alpha1
	kind: KubeProxyConfiguration
	clusterCIDR: "10.244.0.0/16"
	metricsBindAddress: 0.0.0.0:10249
	conntrack:
	  maxPerCore: 0
	# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
	  tcpEstablishedTimeout: 0s
	# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
	  tcpCloseWaitTimeout: 0s
	
	I0830 22:55:14.012442 1225867 kubeadm.go:976] kubelet [Unit]
	Wants=containerd.service
	
	[Service]
	ExecStart=
	ExecStart=/var/lib/minikube/binaries/v1.28.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --container-runtime-endpoint=unix:///run/containerd/containerd.sock --hostname-override=addons-015166 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
	
	[Install]
	 config:
	{KubernetesVersion:v1.28.1 ClusterName:addons-015166 Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:}
	I0830 22:55:14.012512 1225867 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.28.1
	I0830 22:55:14.023704 1225867 binaries.go:44] Found k8s binaries, skipping transfer
	I0830 22:55:14.023823 1225867 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
	I0830 22:55:14.035057 1225867 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (385 bytes)
	I0830 22:55:14.057565 1225867 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
	I0830 22:55:14.079947 1225867 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2102 bytes)
	I0830 22:55:14.106237 1225867 ssh_runner.go:195] Run: grep 192.168.49.2	control-plane.minikube.internal$ /etc/hosts
	I0830 22:55:14.111013 1225867 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2	control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
	I0830 22:55:14.125227 1225867 certs.go:56] Setting up /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166 for IP: 192.168.49.2
	I0830 22:55:14.125314 1225867 certs.go:190] acquiring lock for shared ca certs: {Name:mk74152f2fceba9f1b7e69c343f47396a1a7db69 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:14.125490 1225867 certs.go:204] generating minikubeCA CA: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.key
	I0830 22:55:14.406524 1225867 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt ...
	I0830 22:55:14.406556 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt: {Name:mk787bb5c72e8c82d69bc86b70d04d11e9f252b3 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:14.406746 1225867 crypto.go:164] Writing key to /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.key ...
	I0830 22:55:14.406758 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.key: {Name:mk45a54d8920d59a9a35fe00274fa515138e47c8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:14.406844 1225867 certs.go:204] generating proxyClientCA CA: /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.key
	I0830 22:55:14.887947 1225867 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.crt ...
	I0830 22:55:14.887976 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.crt: {Name:mk24022c1ad5a0a1f65eb6225d9626b933a5293e Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:14.888158 1225867 crypto.go:164] Writing key to /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.key ...
	I0830 22:55:14.888171 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.key: {Name:mk13508630b0862cd4f3e1b6f73136f2e3a5ebc3 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:14.888673 1225867 certs.go:319] generating minikube-user signed cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.key
	I0830 22:55:14.888710 1225867 crypto.go:68] Generating cert /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt with IP's: []
	I0830 22:55:16.002209 1225867 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt ...
	I0830 22:55:16.002240 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: {Name:mk8634d6e54e4dc420545b2ff8453fbd7319a900 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:16.002425 1225867 crypto.go:164] Writing key to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.key ...
	I0830 22:55:16.002440 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.key: {Name:mk76fc9967556eee915958b419b8f6aafeef30e4 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:16.002521 1225867 certs.go:319] generating minikube signed cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.key.dd3b5fb2
	I0830 22:55:16.002539 1225867 crypto.go:68] Generating cert /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.crt.dd3b5fb2 with IP's: [192.168.49.2 10.96.0.1 127.0.0.1 10.0.0.1]
	I0830 22:55:16.270110 1225867 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.crt.dd3b5fb2 ...
	I0830 22:55:16.270155 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.crt.dd3b5fb2: {Name:mkdef870db4805a3ecc27a682ecbc4af078e9b97 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:16.270404 1225867 crypto.go:164] Writing key to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.key.dd3b5fb2 ...
	I0830 22:55:16.270426 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.key.dd3b5fb2: {Name:mke40cb79e775ea06e515253244caa9b1a3a79d9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:16.270869 1225867 certs.go:337] copying /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.crt.dd3b5fb2 -> /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.crt
	I0830 22:55:16.270960 1225867 certs.go:341] copying /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.key.dd3b5fb2 -> /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.key
	I0830 22:55:16.271011 1225867 certs.go:319] generating aggregator signed cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.key
	I0830 22:55:16.271032 1225867 crypto.go:68] Generating cert /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.crt with IP's: []
	I0830 22:55:16.569561 1225867 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.crt ...
	I0830 22:55:16.569606 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.crt: {Name:mk242a9ae46fff1b1ad3f33882eddd63e3f408af Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:16.569834 1225867 crypto.go:164] Writing key to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.key ...
	I0830 22:55:16.569847 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.key: {Name:mk2fe9db02fac21431e503a5e366107e3e13bdbb Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:16.570063 1225867 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem (1679 bytes)
	I0830 22:55:16.570107 1225867 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem (1082 bytes)
	I0830 22:55:16.570145 1225867 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem (1123 bytes)
	I0830 22:55:16.570175 1225867 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem (1679 bytes)
	I0830 22:55:16.570938 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1399 bytes)
	I0830 22:55:16.601990 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
	I0830 22:55:16.632298 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
	I0830 22:55:16.661682 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
	I0830 22:55:16.692437 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
	I0830 22:55:16.722693 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
	I0830 22:55:16.752134 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
	I0830 22:55:16.781679 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
	I0830 22:55:16.812785 1225867 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
	I0830 22:55:16.841975 1225867 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
	I0830 22:55:16.863339 1225867 ssh_runner.go:195] Run: openssl version
	I0830 22:55:16.870397 1225867 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
	I0830 22:55:16.882321 1225867 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
	I0830 22:55:16.887002 1225867 certs.go:480] hashing: -rw-r--r-- 1 root root 1111 Aug 30 22:55 /usr/share/ca-certificates/minikubeCA.pem
	I0830 22:55:16.887093 1225867 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
	I0830 22:55:16.895850 1225867 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
	I0830 22:55:16.907883 1225867 ssh_runner.go:195] Run: ls /var/lib/minikube/certs/etcd
	I0830 22:55:16.912460 1225867 certs.go:353] certs directory doesn't exist, likely first start: ls /var/lib/minikube/certs/etcd: Process exited with status 2
	stdout:
	
	stderr:
	ls: cannot access '/var/lib/minikube/certs/etcd': No such file or directory
	I0830 22:55:16.912520 1225867 kubeadm.go:404] StartCluster: {Name:addons-015166 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:4000 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.28.1 ClusterName:addons-015166 Namespace:default APIServerName:minikubeCA APIServerNames:[] APISe
rverIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.28.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false Disa
bleMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0}
	I0830 22:55:16.912604 1225867 cri.go:54] listing CRI containers in root /run/containerd/runc/k8s.io: {State:paused Name: Namespaces:[kube-system]}
	I0830 22:55:16.912666 1225867 ssh_runner.go:195] Run: sudo -s eval "crictl ps -a --quiet --label io.kubernetes.pod.namespace=kube-system"
	I0830 22:55:16.955838 1225867 cri.go:89] found id: ""
	I0830 22:55:16.955955 1225867 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
	I0830 22:55:16.966655 1225867 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
	I0830 22:55:16.977806 1225867 kubeadm.go:226] ignoring SystemVerification for kubeadm because of docker driver
	I0830 22:55:16.977872 1225867 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
	I0830 22:55:16.988737 1225867 kubeadm.go:152] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
	stdout:
	
	stderr:
	ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
	ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
	ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
	ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
	I0830 22:55:16.988814 1225867 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.28.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml  --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
	I0830 22:55:17.050422 1225867 kubeadm.go:322] [init] Using Kubernetes version: v1.28.1
	I0830 22:55:17.050925 1225867 kubeadm.go:322] [preflight] Running pre-flight checks
	I0830 22:55:17.104432 1225867 kubeadm.go:322] [preflight] The system verification failed. Printing the output from the verification:
	I0830 22:55:17.104498 1225867 kubeadm.go:322] KERNEL_VERSION: 5.15.0-1043-aws
	I0830 22:55:17.104532 1225867 kubeadm.go:322] OS: Linux
	I0830 22:55:17.104578 1225867 kubeadm.go:322] CGROUPS_CPU: enabled
	I0830 22:55:17.104623 1225867 kubeadm.go:322] CGROUPS_CPUACCT: enabled
	I0830 22:55:17.104668 1225867 kubeadm.go:322] CGROUPS_CPUSET: enabled
	I0830 22:55:17.104714 1225867 kubeadm.go:322] CGROUPS_DEVICES: enabled
	I0830 22:55:17.104759 1225867 kubeadm.go:322] CGROUPS_FREEZER: enabled
	I0830 22:55:17.104807 1225867 kubeadm.go:322] CGROUPS_MEMORY: enabled
	I0830 22:55:17.104849 1225867 kubeadm.go:322] CGROUPS_PIDS: enabled
	I0830 22:55:17.104894 1225867 kubeadm.go:322] CGROUPS_HUGETLB: enabled
	I0830 22:55:17.104937 1225867 kubeadm.go:322] CGROUPS_BLKIO: enabled
	I0830 22:55:17.203456 1225867 kubeadm.go:322] [preflight] Pulling images required for setting up a Kubernetes cluster
	I0830 22:55:17.203560 1225867 kubeadm.go:322] [preflight] This might take a minute or two, depending on the speed of your internet connection
	I0830 22:55:17.203652 1225867 kubeadm.go:322] [preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
	I0830 22:55:17.457521 1225867 kubeadm.go:322] [certs] Using certificateDir folder "/var/lib/minikube/certs"
	I0830 22:55:17.462085 1225867 out.go:204]   - Generating certificates and keys ...
	I0830 22:55:17.462182 1225867 kubeadm.go:322] [certs] Using existing ca certificate authority
	I0830 22:55:17.462260 1225867 kubeadm.go:322] [certs] Using existing apiserver certificate and key on disk
	I0830 22:55:18.024124 1225867 kubeadm.go:322] [certs] Generating "apiserver-kubelet-client" certificate and key
	I0830 22:55:18.606799 1225867 kubeadm.go:322] [certs] Generating "front-proxy-ca" certificate and key
	I0830 22:55:19.265465 1225867 kubeadm.go:322] [certs] Generating "front-proxy-client" certificate and key
	I0830 22:55:19.633049 1225867 kubeadm.go:322] [certs] Generating "etcd/ca" certificate and key
	I0830 22:55:19.837853 1225867 kubeadm.go:322] [certs] Generating "etcd/server" certificate and key
	I0830 22:55:19.838241 1225867 kubeadm.go:322] [certs] etcd/server serving cert is signed for DNS names [addons-015166 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
	I0830 22:55:20.764219 1225867 kubeadm.go:322] [certs] Generating "etcd/peer" certificate and key
	I0830 22:55:20.764417 1225867 kubeadm.go:322] [certs] etcd/peer serving cert is signed for DNS names [addons-015166 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
	I0830 22:55:21.086131 1225867 kubeadm.go:322] [certs] Generating "etcd/healthcheck-client" certificate and key
	I0830 22:55:21.381481 1225867 kubeadm.go:322] [certs] Generating "apiserver-etcd-client" certificate and key
	I0830 22:55:21.584542 1225867 kubeadm.go:322] [certs] Generating "sa" key and public key
	I0830 22:55:21.584920 1225867 kubeadm.go:322] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
	I0830 22:55:22.449037 1225867 kubeadm.go:322] [kubeconfig] Writing "admin.conf" kubeconfig file
	I0830 22:55:23.179316 1225867 kubeadm.go:322] [kubeconfig] Writing "kubelet.conf" kubeconfig file
	I0830 22:55:23.721825 1225867 kubeadm.go:322] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
	I0830 22:55:24.035825 1225867 kubeadm.go:322] [kubeconfig] Writing "scheduler.conf" kubeconfig file
	I0830 22:55:24.036484 1225867 kubeadm.go:322] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
	I0830 22:55:24.039365 1225867 kubeadm.go:322] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
	I0830 22:55:24.042000 1225867 out.go:204]   - Booting up control plane ...
	I0830 22:55:24.042152 1225867 kubeadm.go:322] [control-plane] Creating static Pod manifest for "kube-apiserver"
	I0830 22:55:24.042226 1225867 kubeadm.go:322] [control-plane] Creating static Pod manifest for "kube-controller-manager"
	I0830 22:55:24.042288 1225867 kubeadm.go:322] [control-plane] Creating static Pod manifest for "kube-scheduler"
	I0830 22:55:24.057277 1225867 kubeadm.go:322] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
	I0830 22:55:24.058150 1225867 kubeadm.go:322] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
	I0830 22:55:24.058520 1225867 kubeadm.go:322] [kubelet-start] Starting the kubelet
	I0830 22:55:24.175924 1225867 kubeadm.go:322] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
	I0830 22:55:33.678382 1225867 kubeadm.go:322] [apiclient] All control plane components are healthy after 9.502992 seconds
	I0830 22:55:33.678661 1225867 kubeadm.go:322] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
	I0830 22:55:33.707287 1225867 kubeadm.go:322] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
	I0830 22:55:34.237707 1225867 kubeadm.go:322] [upload-certs] Skipping phase. Please see --upload-certs
	I0830 22:55:34.237909 1225867 kubeadm.go:322] [mark-control-plane] Marking the node addons-015166 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
	I0830 22:55:34.749015 1225867 kubeadm.go:322] [bootstrap-token] Using token: 2gijd0.r8muobyvp5fgq8vy
	I0830 22:55:34.751364 1225867 out.go:204]   - Configuring RBAC rules ...
	I0830 22:55:34.751484 1225867 kubeadm.go:322] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
	I0830 22:55:34.756530 1225867 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
	I0830 22:55:34.766834 1225867 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
	I0830 22:55:34.771173 1225867 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
	I0830 22:55:34.775694 1225867 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
	I0830 22:55:34.780041 1225867 kubeadm.go:322] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
	I0830 22:55:34.795221 1225867 kubeadm.go:322] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
	I0830 22:55:35.042070 1225867 kubeadm.go:322] [addons] Applied essential addon: CoreDNS
	I0830 22:55:35.162967 1225867 kubeadm.go:322] [addons] Applied essential addon: kube-proxy
	I0830 22:55:35.166025 1225867 kubeadm.go:322] 
	I0830 22:55:35.166096 1225867 kubeadm.go:322] Your Kubernetes control-plane has initialized successfully!
	I0830 22:55:35.166102 1225867 kubeadm.go:322] 
	I0830 22:55:35.166179 1225867 kubeadm.go:322] To start using your cluster, you need to run the following as a regular user:
	I0830 22:55:35.166184 1225867 kubeadm.go:322] 
	I0830 22:55:35.166208 1225867 kubeadm.go:322]   mkdir -p $HOME/.kube
	I0830 22:55:35.168140 1225867 kubeadm.go:322]   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
	I0830 22:55:35.168211 1225867 kubeadm.go:322]   sudo chown $(id -u):$(id -g) $HOME/.kube/config
	I0830 22:55:35.168216 1225867 kubeadm.go:322] 
	I0830 22:55:35.168268 1225867 kubeadm.go:322] Alternatively, if you are the root user, you can run:
	I0830 22:55:35.168272 1225867 kubeadm.go:322] 
	I0830 22:55:35.168317 1225867 kubeadm.go:322]   export KUBECONFIG=/etc/kubernetes/admin.conf
	I0830 22:55:35.168322 1225867 kubeadm.go:322] 
	I0830 22:55:35.168371 1225867 kubeadm.go:322] You should now deploy a pod network to the cluster.
	I0830 22:55:35.168442 1225867 kubeadm.go:322] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
	I0830 22:55:35.168506 1225867 kubeadm.go:322]   https://kubernetes.io/docs/concepts/cluster-administration/addons/
	I0830 22:55:35.168511 1225867 kubeadm.go:322] 
	I0830 22:55:35.168881 1225867 kubeadm.go:322] You can now join any number of control-plane nodes by copying certificate authorities
	I0830 22:55:35.168959 1225867 kubeadm.go:322] and service account keys on each node and then running the following as root:
	I0830 22:55:35.168964 1225867 kubeadm.go:322] 
	I0830 22:55:35.169905 1225867 kubeadm.go:322]   kubeadm join control-plane.minikube.internal:8443 --token 2gijd0.r8muobyvp5fgq8vy \
	I0830 22:55:35.170012 1225867 kubeadm.go:322] 	--discovery-token-ca-cert-hash sha256:0d761a60bc6e6ab59708493c3270ff221f9632510a1573cc948daf3f99318f4a \
	I0830 22:55:35.170291 1225867 kubeadm.go:322] 	--control-plane 
	I0830 22:55:35.170302 1225867 kubeadm.go:322] 
	I0830 22:55:35.170599 1225867 kubeadm.go:322] Then you can join any number of worker nodes by running the following on each as root:
	I0830 22:55:35.170609 1225867 kubeadm.go:322] 
	I0830 22:55:35.170928 1225867 kubeadm.go:322] kubeadm join control-plane.minikube.internal:8443 --token 2gijd0.r8muobyvp5fgq8vy \
	I0830 22:55:35.171290 1225867 kubeadm.go:322] 	--discovery-token-ca-cert-hash sha256:0d761a60bc6e6ab59708493c3270ff221f9632510a1573cc948daf3f99318f4a 
	I0830 22:55:35.178278 1225867 kubeadm.go:322] 	[WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1043-aws\n", err: exit status 1
	I0830 22:55:35.178391 1225867 kubeadm.go:322] 	[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
	I0830 22:55:35.178410 1225867 cni.go:84] Creating CNI manager for ""
	I0830 22:55:35.178422 1225867 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
	I0830 22:55:35.180748 1225867 out.go:177] * Configuring CNI (Container Networking Interface) ...
	I0830 22:55:35.182853 1225867 ssh_runner.go:195] Run: stat /opt/cni/bin/portmap
	I0830 22:55:35.188531 1225867 cni.go:182] applying CNI manifest using /var/lib/minikube/binaries/v1.28.1/kubectl ...
	I0830 22:55:35.188549 1225867 ssh_runner.go:362] scp memory --> /var/tmp/minikube/cni.yaml (2438 bytes)
	I0830 22:55:35.213341 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml
	I0830 22:55:36.230270 1225867 ssh_runner.go:235] Completed: sudo /var/lib/minikube/binaries/v1.28.1/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml: (1.01689296s)
	I0830 22:55:36.230303 1225867 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
	I0830 22:55:36.230414 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:36.230483 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl label nodes minikube.k8s.io/version=v1.31.2 minikube.k8s.io/commit=dcfed3f069eb419c2ffae8f904d3fba5b9405fc5 minikube.k8s.io/name=addons-015166 minikube.k8s.io/updated_at=2023_08_30T22_55_36_0700 minikube.k8s.io/primary=true --all --overwrite --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:36.438156 1225867 ops.go:34] apiserver oom_adj: -16
	I0830 22:55:36.438266 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:36.532650 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:37.123727 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:37.624076 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:38.124610 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:38.624126 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:39.124288 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:39.624563 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:40.124314 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:40.623586 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:41.123610 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:41.623572 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:42.123886 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:42.623580 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:43.123616 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:43.624546 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:44.123617 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:44.624238 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:45.124003 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:45.624575 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:46.123844 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:46.624285 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:47.124391 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:47.623709 1225867 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 22:55:47.759873 1225867 kubeadm.go:1081] duration metric: took 11.529500958s to wait for elevateKubeSystemPrivileges.
	I0830 22:55:47.759897 1225867 kubeadm.go:406] StartCluster complete in 30.847380706s
	I0830 22:55:47.759912 1225867 settings.go:142] acquiring lock: {Name:mk7f59b3a5ac74e6581a6632908a77e182a8ae89 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:47.760022 1225867 settings.go:150] Updating kubeconfig:  /home/jenkins/minikube-integration/17114-1219981/kubeconfig
	I0830 22:55:47.760406 1225867 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/kubeconfig: {Name:mkdc9959048bf1aa0b8afb8aa74417552569b184 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:55:47.761055 1225867 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
	I0830 22:55:47.761404 1225867 config.go:182] Loaded profile config "addons-015166": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
	I0830 22:55:47.761516 1225867 addons.go:499] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false helm-tiller:false inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false volumesnapshots:true]
	I0830 22:55:47.761585 1225867 addons.go:69] Setting volumesnapshots=true in profile "addons-015166"
	I0830 22:55:47.761598 1225867 addons.go:231] Setting addon volumesnapshots=true in "addons-015166"
	I0830 22:55:47.761635 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.762112 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.763539 1225867 addons.go:69] Setting ingress-dns=true in profile "addons-015166"
	I0830 22:55:47.763561 1225867 addons.go:231] Setting addon ingress-dns=true in "addons-015166"
	I0830 22:55:47.763631 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.764082 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.764449 1225867 addons.go:69] Setting inspektor-gadget=true in profile "addons-015166"
	I0830 22:55:47.764490 1225867 addons.go:231] Setting addon inspektor-gadget=true in "addons-015166"
	I0830 22:55:47.764532 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.764976 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.765056 1225867 addons.go:69] Setting metrics-server=true in profile "addons-015166"
	I0830 22:55:47.765066 1225867 addons.go:231] Setting addon metrics-server=true in "addons-015166"
	I0830 22:55:47.765093 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.765491 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.765552 1225867 addons.go:69] Setting registry=true in profile "addons-015166"
	I0830 22:55:47.765562 1225867 addons.go:231] Setting addon registry=true in "addons-015166"
	I0830 22:55:47.765591 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.765932 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.765986 1225867 addons.go:69] Setting storage-provisioner=true in profile "addons-015166"
	I0830 22:55:47.766004 1225867 addons.go:231] Setting addon storage-provisioner=true in "addons-015166"
	I0830 22:55:47.766028 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.766395 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.766615 1225867 addons.go:69] Setting default-storageclass=true in profile "addons-015166"
	I0830 22:55:47.766632 1225867 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-015166"
	I0830 22:55:47.766891 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.766966 1225867 addons.go:69] Setting cloud-spanner=true in profile "addons-015166"
	I0830 22:55:47.766980 1225867 addons.go:231] Setting addon cloud-spanner=true in "addons-015166"
	I0830 22:55:47.767009 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.767371 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.767430 1225867 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-015166"
	I0830 22:55:47.767455 1225867 addons.go:231] Setting addon csi-hostpath-driver=true in "addons-015166"
	I0830 22:55:47.767487 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.767874 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.768024 1225867 addons.go:69] Setting gcp-auth=true in profile "addons-015166"
	I0830 22:55:47.768041 1225867 mustload.go:65] Loading cluster: addons-015166
	I0830 22:55:47.768202 1225867 config.go:182] Loaded profile config "addons-015166": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
	I0830 22:55:47.768433 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.774369 1225867 addons.go:69] Setting ingress=true in profile "addons-015166"
	I0830 22:55:47.774407 1225867 addons.go:231] Setting addon ingress=true in "addons-015166"
	I0830 22:55:47.774468 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.774970 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:47.866361 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
	I0830 22:55:47.868604 1225867 addons.go:423] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
	I0830 22:55:47.868675 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
	I0830 22:55:47.868760 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:47.871046 1225867 out.go:177]   - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v20230407
	I0830 22:55:47.873423 1225867 out.go:177]   - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v20230407
	I0830 22:55:47.876181 1225867 out.go:177]   - Using image registry.k8s.io/ingress-nginx/controller:v1.8.1
	I0830 22:55:47.902163 1225867 addons.go:423] installing /etc/kubernetes/addons/ingress-deploy.yaml
	I0830 22:55:47.902184 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16083 bytes)
	I0830 22:55:47.902246 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:47.930720 1225867 out.go:177]   - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.2
	I0830 22:55:47.939097 1225867 addons.go:423] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
	I0830 22:55:47.939126 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
	I0830 22:55:47.939205 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:47.924432 1225867 kapi.go:248] "coredns" deployment in "kube-system" namespace and "addons-015166" context rescaled to 1 replicas
	I0830 22:55:47.949268 1225867 start.go:223] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.28.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
	I0830 22:55:47.924486 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:47.978045 1225867 out.go:177]   - Using image registry.k8s.io/metrics-server/metrics-server:v0.6.4
	I0830 22:55:47.975471 1225867 out.go:177] * Verifying Kubernetes components...
	I0830 22:55:47.990408 1225867 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0830 22:55:47.990232 1225867 addons.go:423] installing /etc/kubernetes/addons/metrics-apiservice.yaml
	I0830 22:55:47.990657 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
	I0830 22:55:47.990732 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.055254 1225867 out.go:177]   - Using image docker.io/registry:2.8.1
	I0830 22:55:48.057439 1225867 out.go:177]   - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.5
	I0830 22:55:48.061927 1225867 addons.go:423] installing /etc/kubernetes/addons/registry-rc.yaml
	I0830 22:55:48.061950 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (798 bytes)
	I0830 22:55:48.062025 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.065705 1225867 out.go:177]   - Using image gcr.io/k8s-minikube/storage-provisioner:v5
	I0830 22:55:48.068108 1225867 addons.go:423] installing /etc/kubernetes/addons/storage-provisioner.yaml
	I0830 22:55:48.068127 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
	I0830 22:55:48.068207 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.137859 1225867 out.go:177]   - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.19.0
	I0830 22:55:48.154734 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-namespace.yaml
	I0830 22:55:48.154780 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
	I0830 22:55:48.154868 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.170296 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.170819 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.180504 1225867 addons.go:231] Setting addon default-storageclass=true in "addons-015166"
	I0830 22:55:48.180547 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:48.181015 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:48.186789 1225867 out.go:177]   - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.9
	I0830 22:55:48.189555 1225867 addons.go:423] installing /etc/kubernetes/addons/deployment.yaml
	I0830 22:55:48.189578 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1003 bytes)
	I0830 22:55:48.189647 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.221327 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.265441 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.289346 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
	I0830 22:55:48.295255 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
	I0830 22:55:48.295313 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.300151 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
	I0830 22:55:48.305095 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
	I0830 22:55:48.307086 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
	I0830 22:55:48.311483 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
	I0830 22:55:48.313420 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
	I0830 22:55:48.312395 1225867 node_ready.go:35] waiting up to 6m0s for node "addons-015166" to be "Ready" ...
	I0830 22:55:48.312780 1225867 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^        forward . \/etc\/resolv.conf.*/i \        hosts {\n           192.168.49.1 host.minikube.internal\n           fallthrough\n        }' -e '/^        errors *$/i \        log' | sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
	I0830 22:55:48.317871 1225867 out.go:177]   - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
	I0830 22:55:48.322251 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
	I0830 22:55:48.322267 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
	I0830 22:55:48.322331 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.322163 1225867 node_ready.go:49] node "addons-015166" has status "Ready":"True"
	I0830 22:55:48.322542 1225867 node_ready.go:38] duration metric: took 6.810594ms waiting for node "addons-015166" to be "Ready" ...
	I0830 22:55:48.322555 1225867 pod_ready.go:35] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
	I0830 22:55:48.335715 1225867 pod_ready.go:78] waiting up to 6m0s for pod "coredns-5dd5756b68-dshxx" in "kube-system" namespace to be "Ready" ...
	I0830 22:55:48.376180 1225867 addons.go:423] installing /etc/kubernetes/addons/storageclass.yaml
	I0830 22:55:48.376201 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
	I0830 22:55:48.376266 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:48.376561 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.377731 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.378574 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.438451 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.446622 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:48.814877 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
	I0830 22:55:48.897046 1225867 addons.go:423] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
	I0830 22:55:48.897074 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
	I0830 22:55:48.946872 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
	I0830 22:55:48.953415 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-hostpath.yaml
	I0830 22:55:48.953442 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
	I0830 22:55:48.959907 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
	I0830 22:55:49.043597 1225867 addons.go:423] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
	I0830 22:55:49.043667 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
	I0830 22:55:49.058981 1225867 addons.go:423] installing /etc/kubernetes/addons/registry-svc.yaml
	I0830 22:55:49.059055 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
	I0830 22:55:49.075467 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
	I0830 22:55:49.088732 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
	I0830 22:55:49.132079 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
	I0830 22:55:49.132145 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
	I0830 22:55:49.140655 1225867 addons.go:423] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
	I0830 22:55:49.140720 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
	I0830 22:55:49.148109 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
	I0830 22:55:49.148175 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
	I0830 22:55:49.263803 1225867 addons.go:423] installing /etc/kubernetes/addons/registry-proxy.yaml
	I0830 22:55:49.263869 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
	I0830 22:55:49.267128 1225867 addons.go:423] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
	I0830 22:55:49.267193 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
	I0830 22:55:49.340540 1225867 addons.go:423] installing /etc/kubernetes/addons/metrics-server-service.yaml
	I0830 22:55:49.340610 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
	I0830 22:55:49.437995 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
	I0830 22:55:49.476902 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-role.yaml
	I0830 22:55:49.476971 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
	I0830 22:55:49.483940 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
	I0830 22:55:49.484006 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
	I0830 22:55:49.525606 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
	I0830 22:55:49.640406 1225867 addons.go:423] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
	I0830 22:55:49.640475 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
	I0830 22:55:49.770742 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-rolebinding.yaml
	I0830 22:55:49.770814 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
	I0830 22:55:49.785792 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
	I0830 22:55:49.785859 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
	I0830 22:55:49.860921 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
	I0830 22:55:49.860947 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
	I0830 22:55:49.995753 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-clusterrole.yaml
	I0830 22:55:49.995816 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
	I0830 22:55:50.014478 1225867 addons.go:423] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
	I0830 22:55:50.014541 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
	I0830 22:55:50.053028 1225867 addons.go:423] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
	I0830 22:55:50.053091 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
	I0830 22:55:50.318773 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
	I0830 22:55:50.340134 1225867 addons.go:423] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
	I0830 22:55:50.340241 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
	I0830 22:55:50.366625 1225867 pod_ready.go:97] error getting pod "coredns-5dd5756b68-dshxx" in "kube-system" namespace (skipping!): pods "coredns-5dd5756b68-dshxx" not found
	I0830 22:55:50.366726 1225867 pod_ready.go:81] duration metric: took 2.025980767s waiting for pod "coredns-5dd5756b68-dshxx" in "kube-system" namespace to be "Ready" ...
	E0830 22:55:50.366765 1225867 pod_ready.go:66] WaitExtra: waitPodCondition: error getting pod "coredns-5dd5756b68-dshxx" in "kube-system" namespace (skipping!): pods "coredns-5dd5756b68-dshxx" not found
	I0830 22:55:50.366821 1225867 pod_ready.go:78] waiting up to 6m0s for pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace to be "Ready" ...
	I0830 22:55:50.398806 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
	I0830 22:55:50.398839 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
	I0830 22:55:50.554084 1225867 addons.go:423] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
	I0830 22:55:50.554106 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
	I0830 22:55:50.566085 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-crd.yaml
	I0830 22:55:50.566109 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
	I0830 22:55:50.599660 1225867 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^        forward . \/etc\/resolv.conf.*/i \        hosts {\n           192.168.49.1 host.minikube.internal\n           fallthrough\n        }' -e '/^        errors *$/i \        log' | sudo /var/lib/minikube/binaries/v1.28.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (2.283972982s)
	I0830 22:55:50.599689 1225867 start.go:901] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
	I0830 22:55:50.666399 1225867 addons.go:423] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
	I0830 22:55:50.666423 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
	I0830 22:55:50.689928 1225867 addons.go:423] installing /etc/kubernetes/addons/ig-daemonset.yaml
	I0830 22:55:50.689959 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7741 bytes)
	I0830 22:55:50.823094 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
	I0830 22:55:50.866239 1225867 addons.go:423] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
	I0830 22:55:50.866273 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
	I0830 22:55:51.015705 1225867 addons.go:423] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
	I0830 22:55:51.015729 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
	I0830 22:55:51.263530 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
	I0830 22:55:52.404820 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:55:54.234856 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (5.274922593s)
	I0830 22:55:54.234933 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (5.159432692s)
	I0830 22:55:54.234978 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (5.146216237s)
	I0830 22:55:54.235051 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (4.796980513s)
	I0830 22:55:54.235065 1225867 addons.go:467] Verifying addon metrics-server=true in "addons-015166"
	I0830 22:55:54.235099 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (4.709426444s)
	I0830 22:55:54.235110 1225867 addons.go:467] Verifying addon registry=true in "addons-015166"
	I0830 22:55:54.238813 1225867 out.go:177] * Verifying registry addon...
	I0830 22:55:54.234809 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (5.287897087s)
	I0830 22:55:54.235644 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (3.91679112s)
	I0830 22:55:54.235712 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (3.412575007s)
	I0830 22:55:54.236266 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (5.421359393s)
	I0830 22:55:54.241913 1225867 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
	W0830 22:55:54.242141 1225867 addons.go:449] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
	stdout:
	customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
	customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
	customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
	serviceaccount/snapshot-controller created
	clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
	clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
	role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
	rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
	deployment.apps/snapshot-controller created
	
	stderr:
	error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
	ensure CRDs are installed first
	I0830 22:55:54.242176 1225867 retry.go:31] will retry after 184.541676ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
	stdout:
	customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
	customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
	customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
	serviceaccount/snapshot-controller created
	clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
	clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
	role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
	rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
	deployment.apps/snapshot-controller created
	
	stderr:
	error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
	ensure CRDs are installed first
	I0830 22:55:54.242237 1225867 addons.go:467] Verifying addon ingress=true in "addons-015166"
	I0830 22:55:54.245239 1225867 out.go:177] * Verifying ingress addon...
	I0830 22:55:54.247926 1225867 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
	I0830 22:55:54.249908 1225867 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
	I0830 22:55:54.249938 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:54.261190 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:54.261817 1225867 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
	I0830 22:55:54.261835 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:54.268885 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:54.405371 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:55:54.427121 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
	I0830 22:55:54.778469 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:54.786900 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:54.793796 1225867 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
	I0830 22:55:54.793886 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:54.828644 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:55.267179 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:55.280977 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:55.296761 1225867 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
	I0830 22:55:55.327299 1225867 addons.go:231] Setting addon gcp-auth=true in "addons-015166"
	I0830 22:55:55.327359 1225867 host.go:66] Checking if "addons-015166" exists ...
	I0830 22:55:55.327918 1225867 cli_runner.go:164] Run: docker container inspect addons-015166 --format={{.State.Status}}
	I0830 22:55:55.352583 1225867 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
	I0830 22:55:55.352670 1225867 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-015166
	I0830 22:55:55.374845 1225867 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34314 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/addons-015166/id_rsa Username:docker}
	I0830 22:55:55.767516 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:55.777285 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:56.299547 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:56.302722 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:56.424990 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (1.997763516s)
	I0830 22:55:56.425068 1225867 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (1.072457794s)
	I0830 22:55:56.427658 1225867 out.go:177]   - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v20230407
	I0830 22:55:56.425304 1225867 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (5.161710113s)
	I0830 22:55:56.430062 1225867 out.go:177]   - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.0
	I0830 22:55:56.427746 1225867 addons.go:467] Verifying addon csi-hostpath-driver=true in "addons-015166"
	I0830 22:55:56.432423 1225867 addons.go:423] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
	I0830 22:55:56.432449 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
	I0830 22:55:56.434497 1225867 out.go:177] * Verifying csi-hostpath-driver addon...
	I0830 22:55:56.437396 1225867 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
	I0830 22:55:56.455551 1225867 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
	I0830 22:55:56.455580 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:56.470229 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:56.492601 1225867 addons.go:423] installing /etc/kubernetes/addons/gcp-auth-service.yaml
	I0830 22:55:56.492626 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
	I0830 22:55:56.516496 1225867 addons.go:423] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
	I0830 22:55:56.516522 1225867 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5412 bytes)
	I0830 22:55:56.543629 1225867 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
	I0830 22:55:56.767813 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:56.773866 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:56.897307 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:55:56.980255 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:57.266890 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:57.273715 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:57.485012 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:57.516885 1225867 addons.go:467] Verifying addon gcp-auth=true in "addons-015166"
	I0830 22:55:57.519304 1225867 out.go:177] * Verifying gcp-auth addon...
	I0830 22:55:57.522402 1225867 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
	I0830 22:55:57.534109 1225867 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
	I0830 22:55:57.534185 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:55:57.539988 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:55:57.767176 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:57.773604 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:57.977692 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:58.044564 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:55:58.266854 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:58.273347 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:58.476886 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:58.544516 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:55:58.767168 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:58.773896 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:58.976803 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:59.044735 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:55:59.266608 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:59.274093 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:59.394400 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:55:59.476417 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:55:59.544491 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:55:59.766486 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:55:59.772988 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:55:59.977944 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:00.044863 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:00.272506 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:00.291366 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:00.477876 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:00.546324 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:00.783673 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:00.789453 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:00.977521 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:01.044042 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:01.266658 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:01.274191 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:01.397550 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:01.477476 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:01.545085 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:01.770139 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:01.774682 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:01.978367 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:02.043803 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:02.266704 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:02.273294 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:02.477010 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:02.543786 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:02.766699 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:02.775223 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:02.977988 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:03.043528 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:03.267828 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:03.273775 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:03.401539 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:03.476703 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:03.544877 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:03.767235 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:03.774032 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:03.977426 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:04.044422 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:04.267547 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:04.274547 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:04.477777 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:04.544878 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:04.767360 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:04.774187 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:04.976896 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:05.043959 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:05.267296 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:05.273750 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:05.476744 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:05.544255 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:05.766516 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:05.773737 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:05.895415 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:05.977079 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:06.044323 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:06.266691 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:06.276208 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:06.476758 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:06.544641 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:06.767445 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:06.773721 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:06.976378 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:07.045358 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:07.266540 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:07.273549 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:07.477481 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:07.544376 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:07.766405 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:07.774498 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:07.895764 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:07.985469 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:08.044956 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:08.267521 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:08.274403 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:08.476857 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:08.546039 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:08.767756 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:08.778787 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:08.977764 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:09.044178 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:09.267323 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:09.273912 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:09.479015 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:09.543692 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:09.766019 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:09.773580 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:09.976513 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:10.043497 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:10.267341 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:10.274015 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:10.396113 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:10.476941 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:10.544487 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:10.767035 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:10.773336 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:10.976672 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:11.043923 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:11.265740 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:11.274137 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:11.476144 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:11.543770 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:11.766912 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:11.773513 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:11.976200 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:12.043511 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:12.267120 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:12.273847 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:12.477090 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:12.544292 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:12.766440 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:12.773359 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:12.895089 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:12.976296 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:13.043887 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:13.266980 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:13.273361 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:13.475858 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:13.544494 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:13.766272 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:13.773428 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:13.978315 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:14.043878 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:14.266503 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:14.274148 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:14.476315 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:14.543759 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:14.766091 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:14.773828 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:14.977510 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:15.043632 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:15.266525 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:15.273772 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:15.395371 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:15.476334 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:15.543879 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:15.767711 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:15.773302 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:15.977260 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:16.044662 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:16.266240 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:16.274393 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:16.475789 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:16.544641 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:16.767168 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:16.773413 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:16.976486 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:17.043747 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:17.266745 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:17.273951 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:17.395437 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:17.476315 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:17.543874 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:17.766915 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:17.773435 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:17.976421 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:18.043687 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:18.266633 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:18.273376 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:18.475909 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:18.543391 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:18.766877 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:18.773834 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:18.976696 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:19.044826 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:19.266480 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:19.273030 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:19.397777 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:19.476752 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:19.543817 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:19.766590 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:19.773380 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:19.975520 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:20.044891 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:20.267489 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:20.274400 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:20.478285 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:20.544836 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:20.767722 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:20.774902 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:20.978003 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:21.044544 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:21.268015 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:21.274693 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:21.478432 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:21.544083 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:21.766400 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:21.773925 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:21.895589 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:21.976842 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:22.044792 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:22.266831 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:22.273878 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:22.476429 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:22.546750 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:22.766369 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:22.774125 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:22.975768 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:23.044569 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:23.267429 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:23.274260 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:23.480766 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:23.545634 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:23.768538 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:23.773978 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:23.897190 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:23.977235 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:24.046709 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:24.267621 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:24.275111 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:24.479318 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:24.544228 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:24.769472 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:24.776857 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:24.981234 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:25.046396 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:25.270812 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:25.283040 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:25.481571 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:25.548658 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:25.768807 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:25.776295 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:25.897836 1225867 pod_ready.go:102] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"False"
	I0830 22:56:25.977667 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:26.044630 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:26.273034 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:26.278753 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:26.492832 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:26.546405 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:26.792821 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:26.797170 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:26.906725 1225867 pod_ready.go:92] pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace has status "Ready":"True"
	I0830 22:56:26.906798 1225867 pod_ready.go:81] duration metric: took 36.539955184s waiting for pod "coredns-5dd5756b68-zqzzk" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.906834 1225867 pod_ready.go:78] waiting up to 6m0s for pod "etcd-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.916288 1225867 pod_ready.go:92] pod "etcd-addons-015166" in "kube-system" namespace has status "Ready":"True"
	I0830 22:56:26.916351 1225867 pod_ready.go:81] duration metric: took 9.489853ms waiting for pod "etcd-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.916379 1225867 pod_ready.go:78] waiting up to 6m0s for pod "kube-apiserver-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.926984 1225867 pod_ready.go:92] pod "kube-apiserver-addons-015166" in "kube-system" namespace has status "Ready":"True"
	I0830 22:56:26.927048 1225867 pod_ready.go:81] duration metric: took 10.649032ms waiting for pod "kube-apiserver-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.927074 1225867 pod_ready.go:78] waiting up to 6m0s for pod "kube-controller-manager-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.933402 1225867 pod_ready.go:92] pod "kube-controller-manager-addons-015166" in "kube-system" namespace has status "Ready":"True"
	I0830 22:56:26.933469 1225867 pod_ready.go:81] duration metric: took 6.375288ms waiting for pod "kube-controller-manager-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.933496 1225867 pod_ready.go:78] waiting up to 6m0s for pod "kube-proxy-9qnj5" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.946697 1225867 pod_ready.go:92] pod "kube-proxy-9qnj5" in "kube-system" namespace has status "Ready":"True"
	I0830 22:56:26.946773 1225867 pod_ready.go:81] duration metric: took 13.255929ms waiting for pod "kube-proxy-9qnj5" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.946806 1225867 pod_ready.go:78] waiting up to 6m0s for pod "kube-scheduler-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:26.981050 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:27.046948 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:27.270302 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:27.274360 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:27.293822 1225867 pod_ready.go:92] pod "kube-scheduler-addons-015166" in "kube-system" namespace has status "Ready":"True"
	I0830 22:56:27.293899 1225867 pod_ready.go:81] duration metric: took 347.072461ms waiting for pod "kube-scheduler-addons-015166" in "kube-system" namespace to be "Ready" ...
	I0830 22:56:27.293929 1225867 pod_ready.go:38] duration metric: took 38.971361863s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
	I0830 22:56:27.293977 1225867 api_server.go:52] waiting for apiserver process to appear ...
	I0830 22:56:27.294070 1225867 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 22:56:27.329184 1225867 api_server.go:72] duration metric: took 39.379876034s to wait for apiserver process to appear ...
	I0830 22:56:27.329250 1225867 api_server.go:88] waiting for apiserver healthz status ...
	I0830 22:56:27.329290 1225867 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
	I0830 22:56:27.338596 1225867 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
	ok
	I0830 22:56:27.343208 1225867 api_server.go:141] control plane version: v1.28.1
	I0830 22:56:27.343240 1225867 api_server.go:131] duration metric: took 13.97098ms to wait for apiserver health ...
	I0830 22:56:27.343249 1225867 system_pods.go:43] waiting for kube-system pods to appear ...
	I0830 22:56:27.489102 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:27.504869 1225867 system_pods.go:59] 17 kube-system pods found
	I0830 22:56:27.504958 1225867 system_pods.go:61] "coredns-5dd5756b68-zqzzk" [c01939c6-3560-4bf9-b812-23063cd44277] Running
	I0830 22:56:27.504982 1225867 system_pods.go:61] "csi-hostpath-attacher-0" [050a2cb5-a1c8-4111-a731-ee8003f40a09] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
	I0830 22:56:27.505022 1225867 system_pods.go:61] "csi-hostpath-resizer-0" [fb7d5be5-8065-416b-af32-66d5b72b2341] Running
	I0830 22:56:27.505052 1225867 system_pods.go:61] "csi-hostpathplugin-c4qmz" [844fd9be-03a2-438d-a95d-2f912c388b13] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
	I0830 22:56:27.505072 1225867 system_pods.go:61] "etcd-addons-015166" [23e627bd-c55b-4c2d-9e17-f630fe9dbfb9] Running
	I0830 22:56:27.505091 1225867 system_pods.go:61] "kindnet-gqq4g" [d9dc7990-48db-4173-825e-1ff962f290f6] Running
	I0830 22:56:27.505109 1225867 system_pods.go:61] "kube-apiserver-addons-015166" [e95fbf46-11fb-4bcd-9e7e-07d02e2c7070] Running
	I0830 22:56:27.505303 1225867 system_pods.go:61] "kube-controller-manager-addons-015166" [618ee0b5-f535-45dd-8322-ee5256e71267] Running
	I0830 22:56:27.505341 1225867 system_pods.go:61] "kube-ingress-dns-minikube" [f0723a40-2d31-4e36-9657-0c4e02b37524] Running / Ready:ContainersNotReady (containers with unready status: [minikube-ingress-dns]) / ContainersReady:ContainersNotReady (containers with unready status: [minikube-ingress-dns])
	I0830 22:56:27.505359 1225867 system_pods.go:61] "kube-proxy-9qnj5" [d2ae843c-d8ed-458f-9f49-8dbd3e30c9d2] Running
	I0830 22:56:27.505375 1225867 system_pods.go:61] "kube-scheduler-addons-015166" [fbed79f6-fc5f-4759-aed0-5b85db74727c] Running
	I0830 22:56:27.505397 1225867 system_pods.go:61] "metrics-server-7c66d45ddc-mdsn9" [41e36ea8-5467-4930-a72f-892ee5740e66] Running / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
	I0830 22:56:27.505426 1225867 system_pods.go:61] "registry-gcmd4" [5548c52c-f180-40e9-8007-9ae9bc562acc] Pending / Ready:ContainersNotReady (containers with unready status: [registry]) / ContainersReady:ContainersNotReady (containers with unready status: [registry])
	I0830 22:56:27.505450 1225867 system_pods.go:61] "registry-proxy-r82mq" [e38598dc-2305-4c69-95fc-5496df205f2d] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
	I0830 22:56:27.505468 1225867 system_pods.go:61] "snapshot-controller-58dbcc7b99-2vpgl" [23a2db97-b26d-47cb-afbc-32f39dd6b283] Running
	I0830 22:56:27.505486 1225867 system_pods.go:61] "snapshot-controller-58dbcc7b99-btjhd" [2c8359bc-a1c1-430f-a394-36624245bae5] Running
	I0830 22:56:27.505521 1225867 system_pods.go:61] "storage-provisioner" [160af0dc-ec29-45d3-8494-0a72394086d0] Running
	I0830 22:56:27.505542 1225867 system_pods.go:74] duration metric: took 162.287057ms to wait for pod list to return data ...
	I0830 22:56:27.505574 1225867 default_sa.go:34] waiting for default service account to be created ...
	I0830 22:56:27.545113 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:27.694782 1225867 default_sa.go:45] found service account: "default"
	I0830 22:56:27.694844 1225867 default_sa.go:55] duration metric: took 189.244723ms for default service account to be created ...
	I0830 22:56:27.694868 1225867 system_pods.go:116] waiting for k8s-apps to be running ...
	I0830 22:56:27.767375 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:27.773879 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:27.899274 1225867 system_pods.go:86] 17 kube-system pods found
	I0830 22:56:27.899348 1225867 system_pods.go:89] "coredns-5dd5756b68-zqzzk" [c01939c6-3560-4bf9-b812-23063cd44277] Running
	I0830 22:56:27.899373 1225867 system_pods.go:89] "csi-hostpath-attacher-0" [050a2cb5-a1c8-4111-a731-ee8003f40a09] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
	I0830 22:56:27.899391 1225867 system_pods.go:89] "csi-hostpath-resizer-0" [fb7d5be5-8065-416b-af32-66d5b72b2341] Running
	I0830 22:56:27.899430 1225867 system_pods.go:89] "csi-hostpathplugin-c4qmz" [844fd9be-03a2-438d-a95d-2f912c388b13] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
	I0830 22:56:27.899456 1225867 system_pods.go:89] "etcd-addons-015166" [23e627bd-c55b-4c2d-9e17-f630fe9dbfb9] Running
	I0830 22:56:27.899475 1225867 system_pods.go:89] "kindnet-gqq4g" [d9dc7990-48db-4173-825e-1ff962f290f6] Running
	I0830 22:56:27.899494 1225867 system_pods.go:89] "kube-apiserver-addons-015166" [e95fbf46-11fb-4bcd-9e7e-07d02e2c7070] Running
	I0830 22:56:27.899511 1225867 system_pods.go:89] "kube-controller-manager-addons-015166" [618ee0b5-f535-45dd-8322-ee5256e71267] Running
	I0830 22:56:27.899542 1225867 system_pods.go:89] "kube-ingress-dns-minikube" [f0723a40-2d31-4e36-9657-0c4e02b37524] Running / Ready:ContainersNotReady (containers with unready status: [minikube-ingress-dns]) / ContainersReady:ContainersNotReady (containers with unready status: [minikube-ingress-dns])
	I0830 22:56:27.899567 1225867 system_pods.go:89] "kube-proxy-9qnj5" [d2ae843c-d8ed-458f-9f49-8dbd3e30c9d2] Running
	I0830 22:56:27.899586 1225867 system_pods.go:89] "kube-scheduler-addons-015166" [fbed79f6-fc5f-4759-aed0-5b85db74727c] Running
	I0830 22:56:27.899605 1225867 system_pods.go:89] "metrics-server-7c66d45ddc-mdsn9" [41e36ea8-5467-4930-a72f-892ee5740e66] Running
	I0830 22:56:27.899625 1225867 system_pods.go:89] "registry-gcmd4" [5548c52c-f180-40e9-8007-9ae9bc562acc] Pending / Ready:ContainersNotReady (containers with unready status: [registry]) / ContainersReady:ContainersNotReady (containers with unready status: [registry])
	I0830 22:56:27.899653 1225867 system_pods.go:89] "registry-proxy-r82mq" [e38598dc-2305-4c69-95fc-5496df205f2d] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
	I0830 22:56:27.899676 1225867 system_pods.go:89] "snapshot-controller-58dbcc7b99-2vpgl" [23a2db97-b26d-47cb-afbc-32f39dd6b283] Running
	I0830 22:56:27.899692 1225867 system_pods.go:89] "snapshot-controller-58dbcc7b99-btjhd" [2c8359bc-a1c1-430f-a394-36624245bae5] Running
	I0830 22:56:27.899712 1225867 system_pods.go:89] "storage-provisioner" [160af0dc-ec29-45d3-8494-0a72394086d0] Running
	I0830 22:56:27.899732 1225867 system_pods.go:126] duration metric: took 204.844765ms to wait for k8s-apps to be running ...
	I0830 22:56:27.899759 1225867 system_svc.go:44] waiting for kubelet service to be running ....
	I0830 22:56:27.899835 1225867 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0830 22:56:27.916609 1225867 system_svc.go:56] duration metric: took 16.841041ms WaitForService to wait for kubelet.
	I0830 22:56:27.916634 1225867 kubeadm.go:581] duration metric: took 39.967331751s to wait for : map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] ...
	I0830 22:56:27.916655 1225867 node_conditions.go:102] verifying NodePressure condition ...
	I0830 22:56:27.976660 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:28.044518 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:28.092892 1225867 node_conditions.go:122] node storage ephemeral capacity is 203034800Ki
	I0830 22:56:28.092934 1225867 node_conditions.go:123] node cpu capacity is 2
	I0830 22:56:28.092949 1225867 node_conditions.go:105] duration metric: took 176.289922ms to run NodePressure ...
	I0830 22:56:28.092961 1225867 start.go:228] waiting for startup goroutines ...
	I0830 22:56:28.270225 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:28.274820 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:28.479042 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:28.544903 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:28.766736 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:28.773578 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:28.976393 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:29.044325 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:29.266226 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:29.275002 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:29.475936 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:29.544207 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:29.766134 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:29.773684 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:29.976181 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:30.044738 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:30.267387 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:30.274539 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:30.476774 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:30.544756 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:30.767622 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:30.773933 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:30.978000 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:31.043816 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:31.266048 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:31.273442 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:31.476130 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:31.544812 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:31.767363 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:31.774372 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:31.976700 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:32.044285 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:32.269320 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:32.273801 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:32.481000 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:32.543952 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:32.767346 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:32.774225 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:32.975615 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:33.044610 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:33.266901 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:33.274025 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:33.477189 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:33.543903 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:33.768464 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:33.786263 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:33.981263 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:34.044085 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:34.266833 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:34.274821 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:34.477671 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:34.544281 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:34.773669 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:34.778375 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:34.977711 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:35.044124 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:35.267027 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:35.275931 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:35.476673 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:35.544774 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:35.811651 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:35.815232 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:35.977386 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:36.044181 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:36.266914 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0830 22:56:36.277877 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:36.476786 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:36.546070 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:36.766992 1225867 kapi.go:107] duration metric: took 42.525078256s to wait for kubernetes.io/minikube-addons=registry ...
	I0830 22:56:36.773708 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:36.976455 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:37.043977 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:37.273941 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:37.477803 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:37.544940 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:37.774922 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:37.978460 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:38.046590 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:38.274190 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:38.476371 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:38.544148 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:38.774729 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:38.980986 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:39.047237 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:39.274240 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:39.476328 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:39.544270 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:39.774852 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:39.977383 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:40.044018 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:40.275509 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:40.476284 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:40.544697 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:40.773516 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:40.976349 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:41.044024 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:41.273825 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:41.476367 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:41.543754 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:41.773484 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:41.976504 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:42.044330 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:42.274868 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:42.478019 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:42.543616 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:42.774942 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:42.976907 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:43.051026 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:43.274444 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:43.476356 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:43.543961 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:43.774564 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:43.976537 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:44.043716 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:44.274128 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:44.476358 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:44.546898 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:44.776924 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:44.976748 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:45.044047 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:45.276413 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:45.476416 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:45.544178 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:45.773905 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:45.978165 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:46.045962 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:46.274042 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:46.476390 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0830 22:56:46.552121 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:46.776543 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:46.976425 1225867 kapi.go:107] duration metric: took 50.539026401s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
	I0830 22:56:47.044438 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:47.273862 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:47.544321 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:47.773353 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:48.044600 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:48.273964 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:48.543763 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:48.774907 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:49.044078 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:49.273543 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:49.543619 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:49.774330 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:50.044185 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:50.273996 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:50.543756 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:50.773583 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:51.044464 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:51.274687 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:51.544578 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:51.774091 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:52.043965 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:52.274442 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:52.544167 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:52.773888 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:53.044845 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:53.274033 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:53.543589 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:53.774187 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:54.044107 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:54.274322 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:54.544093 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:54.773751 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:55.044841 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:55.274072 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:55.544103 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:55.774139 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:56.044179 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:56.273308 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:56.544027 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:56.774282 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:57.044212 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:57.274277 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:57.544323 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:57.773719 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:58.044918 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:58.274653 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:58.544995 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:58.774668 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:59.044463 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:59.274163 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:56:59.544410 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:56:59.774205 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:00.044601 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:00.280437 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:00.546593 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:00.774467 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:01.044829 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:01.274182 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:01.544666 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:01.774151 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:02.044517 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:02.276259 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:02.544812 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:02.773802 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:03.044896 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:03.273817 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:03.544243 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:03.774880 1225867 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0830 22:57:04.045417 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:04.274225 1225867 kapi.go:107] duration metric: took 1m10.026296115s to wait for app.kubernetes.io/name=ingress-nginx ...
	I0830 22:57:04.544385 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:05.044284 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:05.545557 1225867 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0830 22:57:06.043662 1225867 kapi.go:107] duration metric: took 1m8.521256432s to wait for kubernetes.io/minikube-addons=gcp-auth ...
	I0830 22:57:06.045448 1225867 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-015166 cluster.
	I0830 22:57:06.047054 1225867 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
	I0830 22:57:06.048742 1225867 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
	I0830 22:57:06.050496 1225867 out.go:177] * Enabled addons: default-storageclass, storage-provisioner, cloud-spanner, metrics-server, ingress-dns, inspektor-gadget, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
	I0830 22:57:06.052223 1225867 addons.go:502] enable addons completed in 1m18.290701031s: enabled=[default-storageclass storage-provisioner cloud-spanner metrics-server ingress-dns inspektor-gadget volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
	I0830 22:57:06.052267 1225867 start.go:233] waiting for cluster config update ...
	I0830 22:57:06.052289 1225867 start.go:242] writing updated cluster config ...
	I0830 22:57:06.052588 1225867 ssh_runner.go:195] Run: rm -f paused
	I0830 22:57:06.117785 1225867 start.go:600] kubectl: 1.28.1, cluster: 1.28.1 (minor skew: 0)
	I0830 22:57:06.120005 1225867 out.go:177] * Done! kubectl is now configured to use "addons-015166" cluster and "default" namespace by default
	
	* 
	* ==> container status <==
	* CONTAINER           IMAGE               CREATED              STATE               NAME                         ATTEMPT             POD ID              POD
	d7ea4da00dd68       13753a81eccfd       14 seconds ago       Exited              hello-world-app              2                   06de2ec44f201       hello-world-app-5d77478584-8crfr
	fb0ee897246c4       fa0c6bb795403       42 seconds ago       Running             nginx                        0                   791ce5d6d6c24       nginx
	35ffc83ed1eed       71e15c1ff4390       51 seconds ago       Running             headlamp                     0                   e1abfce443c9d       headlamp-699c48fb74-k524x
	ee04b41f0bf03       2a5f29343eb03       About a minute ago   Running             gcp-auth                     0                   0860552fad985       gcp-auth-d4c87556c-zsxt6
	1cb36a600adf3       8f2588812ab29       About a minute ago   Exited              patch                        2                   80a49b91b38cc       ingress-nginx-admission-patch-zc5xm
	304d7ff1318df       8f2588812ab29       About a minute ago   Exited              create                       0                   5f8a77a81af5f       ingress-nginx-admission-create-qzcpj
	f9a18adb6c2d8       97e04611ad434       About a minute ago   Running             coredns                      0                   28e7d1bd6036e       coredns-5dd5756b68-zqzzk
	c6c98d79d0f4b       4d1e5c3e97420       About a minute ago   Running             volume-snapshot-controller   0                   90ed02ff215ea       snapshot-controller-58dbcc7b99-btjhd
	5ae349ee77243       4d1e5c3e97420       About a minute ago   Running             volume-snapshot-controller   0                   bb9cbb600cfa5       snapshot-controller-58dbcc7b99-2vpgl
	f7dfbb5ce2daa       3d00e9f7bc4a0       2 minutes ago        Running             gadget                       0                   06819613aceb9       gadget-khchg
	9a2c72414e97c       ba04bb24b9575       2 minutes ago        Running             storage-provisioner          0                   ed686676ba58e       storage-provisioner
	d13b68d97b05e       b18bf71b941ba       2 minutes ago        Running             kindnet-cni                  0                   c8a137d0b9f79       kindnet-gqq4g
	c2ca89ac2f5fe       812f5241df7fd       2 minutes ago        Running             kube-proxy                   0                   327ad56d71b9d       kube-proxy-9qnj5
	b1afaa34c0101       8b6e1980b7584       2 minutes ago        Running             kube-controller-manager      0                   0d6a50dd99362       kube-controller-manager-addons-015166
	4bb20589784e1       b29fb62480892       2 minutes ago        Running             kube-apiserver               0                   2581802b3bf3c       kube-apiserver-addons-015166
	b355595c7ff7a       9cdd6470f48c8       2 minutes ago        Running             etcd                         0                   ebedfe81ebaf1       etcd-addons-015166
	f8c96f4a21ae6       b4a5a57e99492       2 minutes ago        Running             kube-scheduler               0                   deb30374baa14       kube-scheduler-addons-015166
	
	* 
	* ==> containerd <==
	* Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.239932764Z" level=error msg="ContainerStatus for \"5f5359cdcba5e18f2c51309a012aa4e5a31bfda48186661dee257e4dd5dcdcfe\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"5f5359cdcba5e18f2c51309a012aa4e5a31bfda48186661dee257e4dd5dcdcfe\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.240660911Z" level=error msg="ContainerStatus for \"0f8e9a2b059d951a3f84c66694048b21ffcf5c5a1819b45499347b75760cc135\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0f8e9a2b059d951a3f84c66694048b21ffcf5c5a1819b45499347b75760cc135\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.241316311Z" level=error msg="ContainerStatus for \"3bec36b5d47fdd1526c0129bcc9db5a468deef3fbe44e594f32bd644488bafc4\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3bec36b5d47fdd1526c0129bcc9db5a468deef3fbe44e594f32bd644488bafc4\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.242130792Z" level=error msg="ContainerStatus for \"da433db2dae6c84c73764f99474c7a81f4b7ce3928d0f1c51810dde311aaaf27\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"da433db2dae6c84c73764f99474c7a81f4b7ce3928d0f1c51810dde311aaaf27\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.243653203Z" level=error msg="ContainerStatus for \"4f3b71078fed853badeefa53db64ff3271fd56f60fc4c18a36fbc1cb42e31e93\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"4f3b71078fed853badeefa53db64ff3271fd56f60fc4c18a36fbc1cb42e31e93\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.244500866Z" level=error msg="ContainerStatus for \"1739a952fc4f06711af2c0152a0edff17e3b6d138943f516e709f4aeecf70bda\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1739a952fc4f06711af2c0152a0edff17e3b6d138943f516e709f4aeecf70bda\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.245252406Z" level=error msg="ContainerStatus for \"5f5359cdcba5e18f2c51309a012aa4e5a31bfda48186661dee257e4dd5dcdcfe\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"5f5359cdcba5e18f2c51309a012aa4e5a31bfda48186661dee257e4dd5dcdcfe\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.245952450Z" level=error msg="ContainerStatus for \"0f8e9a2b059d951a3f84c66694048b21ffcf5c5a1819b45499347b75760cc135\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0f8e9a2b059d951a3f84c66694048b21ffcf5c5a1819b45499347b75760cc135\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.246565455Z" level=error msg="ContainerStatus for \"3bec36b5d47fdd1526c0129bcc9db5a468deef3fbe44e594f32bd644488bafc4\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3bec36b5d47fdd1526c0129bcc9db5a468deef3fbe44e594f32bd644488bafc4\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.249484328Z" level=error msg="ContainerStatus for \"da433db2dae6c84c73764f99474c7a81f4b7ce3928d0f1c51810dde311aaaf27\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"da433db2dae6c84c73764f99474c7a81f4b7ce3928d0f1c51810dde311aaaf27\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.250245566Z" level=error msg="ContainerStatus for \"4f3b71078fed853badeefa53db64ff3271fd56f60fc4c18a36fbc1cb42e31e93\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"4f3b71078fed853badeefa53db64ff3271fd56f60fc4c18a36fbc1cb42e31e93\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.254123248Z" level=error msg="ContainerStatus for \"1739a952fc4f06711af2c0152a0edff17e3b6d138943f516e709f4aeecf70bda\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1739a952fc4f06711af2c0152a0edff17e3b6d138943f516e709f4aeecf70bda\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.255025393Z" level=error msg="ContainerStatus for \"5f5359cdcba5e18f2c51309a012aa4e5a31bfda48186661dee257e4dd5dcdcfe\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"5f5359cdcba5e18f2c51309a012aa4e5a31bfda48186661dee257e4dd5dcdcfe\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.256153744Z" level=error msg="ContainerStatus for \"0f8e9a2b059d951a3f84c66694048b21ffcf5c5a1819b45499347b75760cc135\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0f8e9a2b059d951a3f84c66694048b21ffcf5c5a1819b45499347b75760cc135\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.256876976Z" level=error msg="ContainerStatus for \"3bec36b5d47fdd1526c0129bcc9db5a468deef3fbe44e594f32bd644488bafc4\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3bec36b5d47fdd1526c0129bcc9db5a468deef3fbe44e594f32bd644488bafc4\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.257971917Z" level=error msg="ContainerStatus for \"da433db2dae6c84c73764f99474c7a81f4b7ce3928d0f1c51810dde311aaaf27\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"da433db2dae6c84c73764f99474c7a81f4b7ce3928d0f1c51810dde311aaaf27\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.258679650Z" level=error msg="ContainerStatus for \"4f3b71078fed853badeefa53db64ff3271fd56f60fc4c18a36fbc1cb42e31e93\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"4f3b71078fed853badeefa53db64ff3271fd56f60fc4c18a36fbc1cb42e31e93\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.259414927Z" level=error msg="ContainerStatus for \"1739a952fc4f06711af2c0152a0edff17e3b6d138943f516e709f4aeecf70bda\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1739a952fc4f06711af2c0152a0edff17e3b6d138943f516e709f4aeecf70bda\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.260180317Z" level=error msg="ContainerStatus for \"5f5359cdcba5e18f2c51309a012aa4e5a31bfda48186661dee257e4dd5dcdcfe\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"5f5359cdcba5e18f2c51309a012aa4e5a31bfda48186661dee257e4dd5dcdcfe\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.260896976Z" level=error msg="ContainerStatus for \"0f8e9a2b059d951a3f84c66694048b21ffcf5c5a1819b45499347b75760cc135\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0f8e9a2b059d951a3f84c66694048b21ffcf5c5a1819b45499347b75760cc135\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.261752401Z" level=error msg="ContainerStatus for \"3bec36b5d47fdd1526c0129bcc9db5a468deef3fbe44e594f32bd644488bafc4\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3bec36b5d47fdd1526c0129bcc9db5a468deef3fbe44e594f32bd644488bafc4\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.262845692Z" level=error msg="ContainerStatus for \"da433db2dae6c84c73764f99474c7a81f4b7ce3928d0f1c51810dde311aaaf27\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"da433db2dae6c84c73764f99474c7a81f4b7ce3928d0f1c51810dde311aaaf27\": not found"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.269003889Z" level=info msg="RemoveContainer for \"43ed6efb3f3466e5878e9ae1a5dc109e5e48f86413e8e247e8142be68e980a71\""
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.278449336Z" level=info msg="RemoveContainer for \"43ed6efb3f3466e5878e9ae1a5dc109e5e48f86413e8e247e8142be68e980a71\" returns successfully"
	Aug 30 22:58:13 addons-015166 containerd[745]: time="2023-08-30T22:58:13.282714874Z" level=error msg="ContainerStatus for \"43ed6efb3f3466e5878e9ae1a5dc109e5e48f86413e8e247e8142be68e980a71\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"43ed6efb3f3466e5878e9ae1a5dc109e5e48f86413e8e247e8142be68e980a71\": not found"
	
	* 
	* ==> coredns [f9a18adb6c2d878bcc606e4e339ca5c43300434fd084ea18a7b2aa71a357f2e0] <==
	* [INFO] 10.244.0.16:54471 - 61910 "AAAA IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000056993s
	[INFO] 10.244.0.16:54471 - 53720 "A IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.001429094s
	[INFO] 10.244.0.16:34592 - 3236 "A IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.001982685s
	[INFO] 10.244.0.16:54471 - 46983 "AAAA IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.00162737s
	[INFO] 10.244.0.16:34592 - 57754 "AAAA IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.001570206s
	[INFO] 10.244.0.16:54471 - 43173 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.00025207s
	[INFO] 10.244.0.16:34592 - 47315 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000272131s
	[INFO] 10.244.0.16:35779 - 5561 "A IN hello-world-app.default.svc.cluster.local.ingress-nginx.svc.cluster.local. udp 91 false 512" NXDOMAIN qr,aa,rd 184 0.000099159s
	[INFO] 10.244.0.16:39162 - 40460 "A IN hello-world-app.default.svc.cluster.local.ingress-nginx.svc.cluster.local. udp 91 false 512" NXDOMAIN qr,aa,rd 184 0.000297969s
	[INFO] 10.244.0.16:39162 - 32630 "AAAA IN hello-world-app.default.svc.cluster.local.ingress-nginx.svc.cluster.local. udp 91 false 512" NXDOMAIN qr,aa,rd 184 0.00010336s
	[INFO] 10.244.0.16:35779 - 60214 "AAAA IN hello-world-app.default.svc.cluster.local.ingress-nginx.svc.cluster.local. udp 91 false 512" NXDOMAIN qr,aa,rd 184 0.000096534s
	[INFO] 10.244.0.16:35779 - 57296 "A IN hello-world-app.default.svc.cluster.local.svc.cluster.local. udp 77 false 512" NXDOMAIN qr,aa,rd 170 0.000148119s
	[INFO] 10.244.0.16:39162 - 6278 "A IN hello-world-app.default.svc.cluster.local.svc.cluster.local. udp 77 false 512" NXDOMAIN qr,aa,rd 170 0.000057338s
	[INFO] 10.244.0.16:39162 - 52990 "AAAA IN hello-world-app.default.svc.cluster.local.svc.cluster.local. udp 77 false 512" NXDOMAIN qr,aa,rd 170 0.000049182s
	[INFO] 10.244.0.16:35779 - 24319 "AAAA IN hello-world-app.default.svc.cluster.local.svc.cluster.local. udp 77 false 512" NXDOMAIN qr,aa,rd 170 0.000145149s
	[INFO] 10.244.0.16:35779 - 11713 "A IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000088353s
	[INFO] 10.244.0.16:39162 - 3921 "A IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000196808s
	[INFO] 10.244.0.16:39162 - 27022 "AAAA IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000067979s
	[INFO] 10.244.0.16:35779 - 38494 "AAAA IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000037612s
	[INFO] 10.244.0.16:35779 - 24234 "A IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.002085832s
	[INFO] 10.244.0.16:39162 - 4374 "A IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.002105294s
	[INFO] 10.244.0.16:35779 - 47799 "AAAA IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.001371256s
	[INFO] 10.244.0.16:39162 - 5174 "AAAA IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.00152159s
	[INFO] 10.244.0.16:39162 - 32423 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000064632s
	[INFO] 10.244.0.16:35779 - 9069 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000162741s
	
	* 
	* ==> describe nodes <==
	* Name:               addons-015166
	Roles:              control-plane
	Labels:             beta.kubernetes.io/arch=arm64
	                    beta.kubernetes.io/os=linux
	                    kubernetes.io/arch=arm64
	                    kubernetes.io/hostname=addons-015166
	                    kubernetes.io/os=linux
	                    minikube.k8s.io/commit=dcfed3f069eb419c2ffae8f904d3fba5b9405fc5
	                    minikube.k8s.io/name=addons-015166
	                    minikube.k8s.io/primary=true
	                    minikube.k8s.io/updated_at=2023_08_30T22_55_36_0700
	                    minikube.k8s.io/version=v1.31.2
	                    node-role.kubernetes.io/control-plane=
	                    node.kubernetes.io/exclude-from-external-load-balancers=
	                    topology.hostpath.csi/node=addons-015166
	Annotations:        kubeadm.alpha.kubernetes.io/cri-socket: unix:///run/containerd/containerd.sock
	                    node.alpha.kubernetes.io/ttl: 0
	                    volumes.kubernetes.io/controller-managed-attach-detach: true
	CreationTimestamp:  Wed, 30 Aug 2023 22:55:31 +0000
	Taints:             <none>
	Unschedulable:      false
	Lease:
	  HolderIdentity:  addons-015166
	  AcquireTime:     <unset>
	  RenewTime:       Wed, 30 Aug 2023 22:58:07 +0000
	Conditions:
	  Type             Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message
	  ----             ------  -----------------                 ------------------                ------                       -------
	  MemoryPressure   False   Wed, 30 Aug 2023 22:58:08 +0000   Wed, 30 Aug 2023 22:55:27 +0000   KubeletHasSufficientMemory   kubelet has sufficient memory available
	  DiskPressure     False   Wed, 30 Aug 2023 22:58:08 +0000   Wed, 30 Aug 2023 22:55:27 +0000   KubeletHasNoDiskPressure     kubelet has no disk pressure
	  PIDPressure      False   Wed, 30 Aug 2023 22:58:08 +0000   Wed, 30 Aug 2023 22:55:27 +0000   KubeletHasSufficientPID      kubelet has sufficient PID available
	  Ready            True    Wed, 30 Aug 2023 22:58:08 +0000   Wed, 30 Aug 2023 22:55:35 +0000   KubeletReady                 kubelet is posting ready status
	Addresses:
	  InternalIP:  192.168.49.2
	  Hostname:    addons-015166
	Capacity:
	  cpu:                2
	  ephemeral-storage:  203034800Ki
	  hugepages-1Gi:      0
	  hugepages-2Mi:      0
	  hugepages-32Mi:     0
	  hugepages-64Ki:     0
	  memory:             8022572Ki
	  pods:               110
	Allocatable:
	  cpu:                2
	  ephemeral-storage:  203034800Ki
	  hugepages-1Gi:      0
	  hugepages-2Mi:      0
	  hugepages-32Mi:     0
	  hugepages-64Ki:     0
	  memory:             8022572Ki
	  pods:               110
	System Info:
	  Machine ID:                 8d46025cfb614346be2593e1a340a0fd
	  System UUID:                5ab26e7d-ae38-4bb3-8872-d7eafeec3dbd
	  Boot ID:                    98673563-8173-4281-afb4-eac1dfafdc23
	  Kernel Version:             5.15.0-1043-aws
	  OS Image:                   Ubuntu 22.04.3 LTS
	  Operating System:           linux
	  Architecture:               arm64
	  Container Runtime Version:  containerd://1.6.22
	  Kubelet Version:            v1.28.1
	  Kube-Proxy Version:         v1.28.1
	PodCIDR:                      10.244.0.0/24
	PodCIDRs:                     10.244.0.0/24
	Non-terminated Pods:          (15 in total)
	  Namespace                   Name                                     CPU Requests  CPU Limits  Memory Requests  Memory Limits  Age
	  ---------                   ----                                     ------------  ----------  ---------------  -------------  ---
	  default                     hello-world-app-5d77478584-8crfr         0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         35s
	  default                     nginx                                    0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         45s
	  gadget                      gadget-khchg                             0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m21s
	  gcp-auth                    gcp-auth-d4c87556c-zsxt6                 0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m17s
	  headlamp                    headlamp-699c48fb74-k524x                0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         55s
	  kube-system                 coredns-5dd5756b68-zqzzk                 100m (5%!)(MISSING)     0 (0%!)(MISSING)      70Mi (0%!)(MISSING)        170Mi (2%!)(MISSING)     2m27s
	  kube-system                 etcd-addons-015166                       100m (5%!)(MISSING)     0 (0%!)(MISSING)      100Mi (1%!)(MISSING)       0 (0%!)(MISSING)         2m39s
	  kube-system                 kindnet-gqq4g                            100m (5%!)(MISSING)     100m (5%!)(MISSING)   50Mi (0%!)(MISSING)        50Mi (0%!)(MISSING)      2m28s
	  kube-system                 kube-apiserver-addons-015166             250m (12%!)(MISSING)    0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m39s
	  kube-system                 kube-controller-manager-addons-015166    200m (10%!)(MISSING)    0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m39s
	  kube-system                 kube-proxy-9qnj5                         0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m28s
	  kube-system                 kube-scheduler-addons-015166             100m (5%!)(MISSING)     0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m39s
	  kube-system                 snapshot-controller-58dbcc7b99-2vpgl     0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m21s
	  kube-system                 snapshot-controller-58dbcc7b99-btjhd     0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m21s
	  kube-system                 storage-provisioner                      0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         2m22s
	Allocated resources:
	  (Total limits may be over 100 percent, i.e., overcommitted.)
	  Resource           Requests    Limits
	  --------           --------    ------
	  cpu                850m (42%!)(MISSING)  100m (5%!)(MISSING)
	  memory             220Mi (2%!)(MISSING)  220Mi (2%!)(MISSING)
	  ephemeral-storage  0 (0%!)(MISSING)      0 (0%!)(MISSING)
	  hugepages-1Gi      0 (0%!)(MISSING)      0 (0%!)(MISSING)
	  hugepages-2Mi      0 (0%!)(MISSING)      0 (0%!)(MISSING)
	  hugepages-32Mi     0 (0%!)(MISSING)      0 (0%!)(MISSING)
	  hugepages-64Ki     0 (0%!)(MISSING)      0 (0%!)(MISSING)
	Events:
	  Type    Reason                   Age                    From             Message
	  ----    ------                   ----                   ----             -------
	  Normal  Starting                 2m25s                  kube-proxy       
	  Normal  Starting                 2m49s                  kubelet          Starting kubelet.
	  Normal  NodeAllocatableEnforced  2m49s                  kubelet          Updated Node Allocatable limit across pods
	  Normal  NodeHasSufficientMemory  2m48s (x8 over 2m49s)  kubelet          Node addons-015166 status is now: NodeHasSufficientMemory
	  Normal  NodeHasNoDiskPressure    2m48s (x8 over 2m49s)  kubelet          Node addons-015166 status is now: NodeHasNoDiskPressure
	  Normal  NodeHasSufficientPID     2m48s (x7 over 2m49s)  kubelet          Node addons-015166 status is now: NodeHasSufficientPID
	  Normal  Starting                 2m39s                  kubelet          Starting kubelet.
	  Normal  NodeHasSufficientMemory  2m39s                  kubelet          Node addons-015166 status is now: NodeHasSufficientMemory
	  Normal  NodeHasNoDiskPressure    2m39s                  kubelet          Node addons-015166 status is now: NodeHasNoDiskPressure
	  Normal  NodeHasSufficientPID     2m39s                  kubelet          Node addons-015166 status is now: NodeHasSufficientPID
	  Normal  NodeNotReady             2m39s                  kubelet          Node addons-015166 status is now: NodeNotReady
	  Normal  NodeAllocatableEnforced  2m39s                  kubelet          Updated Node Allocatable limit across pods
	  Normal  NodeReady                2m39s                  kubelet          Node addons-015166 status is now: NodeReady
	  Normal  RegisteredNode           2m28s                  node-controller  Node addons-015166 event: Registered Node addons-015166 in Controller
	
	* 
	* ==> dmesg <==
	* [  +0.001068] FS-Cache: O-key=[8] 'a53f5c0100000000'
	[  +0.000743] FS-Cache: N-cookie c=00000054 [p=0000004b fl=2 nc=0 na=1]
	[  +0.000990] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=0000000052a3ffac
	[  +0.001181] FS-Cache: N-key=[8] 'a53f5c0100000000'
	[  +0.003620] FS-Cache: Duplicate cookie detected
	[  +0.000757] FS-Cache: O-cookie c=0000004e [p=0000004b fl=226 nc=0 na=1]
	[  +0.000989] FS-Cache: O-cookie d=00000000d8a48a2b{9p.inode} n=00000000cfc10e18
	[  +0.001078] FS-Cache: O-key=[8] 'a53f5c0100000000'
	[  +0.000892] FS-Cache: N-cookie c=00000055 [p=0000004b fl=2 nc=0 na=1]
	[  +0.000999] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=00000000ec866464
	[  +0.001154] FS-Cache: N-key=[8] 'a53f5c0100000000'
	[  +3.285800] FS-Cache: Duplicate cookie detected
	[  +0.000913] FS-Cache: O-cookie c=0000004c [p=0000004b fl=226 nc=0 na=1]
	[  +0.001105] FS-Cache: O-cookie d=00000000d8a48a2b{9p.inode} n=00000000185770a2
	[  +0.001225] FS-Cache: O-key=[8] 'a43f5c0100000000'
	[  +0.000833] FS-Cache: N-cookie c=00000057 [p=0000004b fl=2 nc=0 na=1]
	[  +0.001080] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=000000006d053276
	[  +0.001194] FS-Cache: N-key=[8] 'a43f5c0100000000'
	[  +0.414572] FS-Cache: Duplicate cookie detected
	[  +0.000724] FS-Cache: O-cookie c=00000051 [p=0000004b fl=226 nc=0 na=1]
	[  +0.000967] FS-Cache: O-cookie d=00000000d8a48a2b{9p.inode} n=000000001a64e3e4
	[  +0.001029] FS-Cache: O-key=[8] 'aa3f5c0100000000'
	[  +0.000731] FS-Cache: N-cookie c=00000058 [p=0000004b fl=2 nc=0 na=1]
	[  +0.000927] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=00000000a41d18fb
	[  +0.001092] FS-Cache: N-key=[8] 'aa3f5c0100000000'
	
	* 
	* ==> etcd [b355595c7ff7afc32de0f59644494b8cdeec14e940ea30b97ee8745babc29b13] <==
	* {"level":"info","ts":"2023-08-30T22:55:26.858137Z","caller":"fileutil/purge.go:44","msg":"started to purge file","dir":"/var/lib/minikube/etcd/member/snap","suffix":"snap.db","max":5,"interval":"30s"}
	{"level":"info","ts":"2023-08-30T22:55:26.861394Z","caller":"fileutil/purge.go:44","msg":"started to purge file","dir":"/var/lib/minikube/etcd/member/snap","suffix":"snap","max":5,"interval":"30s"}
	{"level":"info","ts":"2023-08-30T22:55:26.861509Z","caller":"fileutil/purge.go:44","msg":"started to purge file","dir":"/var/lib/minikube/etcd/member/wal","suffix":"wal","max":5,"interval":"30s"}
	{"level":"info","ts":"2023-08-30T22:55:26.862175Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc switched to configuration voters=(12593026477526642892)"}
	{"level":"info","ts":"2023-08-30T22:55:26.862531Z","caller":"membership/cluster.go:421","msg":"added member","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","added-peer-id":"aec36adc501070cc","added-peer-peer-urls":["https://192.168.49.2:2380"]}
	{"level":"info","ts":"2023-08-30T22:55:26.862426Z","caller":"embed/etcd.go:597","msg":"serving peer traffic","address":"192.168.49.2:2380"}
	{"level":"info","ts":"2023-08-30T22:55:26.862736Z","caller":"embed/etcd.go:569","msg":"cmux::serve","address":"192.168.49.2:2380"}
	{"level":"info","ts":"2023-08-30T22:55:27.02517Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc is starting a new election at term 1"}
	{"level":"info","ts":"2023-08-30T22:55:27.025387Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became pre-candidate at term 1"}
	{"level":"info","ts":"2023-08-30T22:55:27.02549Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgPreVoteResp from aec36adc501070cc at term 1"}
	{"level":"info","ts":"2023-08-30T22:55:27.025592Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became candidate at term 2"}
	{"level":"info","ts":"2023-08-30T22:55:27.025672Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2"}
	{"level":"info","ts":"2023-08-30T22:55:27.025757Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
	{"level":"info","ts":"2023-08-30T22:55:27.025799Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
	{"level":"info","ts":"2023-08-30T22:55:27.029329Z","caller":"etcdserver/server.go:2062","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-015166 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
	{"level":"info","ts":"2023-08-30T22:55:27.029692Z","caller":"etcdserver/server.go:2571","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
	{"level":"info","ts":"2023-08-30T22:55:27.029898Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
	{"level":"info","ts":"2023-08-30T22:55:27.031194Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
	{"level":"info","ts":"2023-08-30T22:55:27.031375Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
	{"level":"info","ts":"2023-08-30T22:55:27.032076Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
	{"level":"info","ts":"2023-08-30T22:55:27.032167Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
	{"level":"info","ts":"2023-08-30T22:55:27.037239Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
	{"level":"info","ts":"2023-08-30T22:55:27.037364Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
	{"level":"info","ts":"2023-08-30T22:55:27.037432Z","caller":"etcdserver/server.go:2595","msg":"cluster version is updated","cluster-version":"3.5"}
	{"level":"info","ts":"2023-08-30T22:55:27.061609Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
	
	* 
	* ==> gcp-auth [ee04b41f0bf0363aaa39bc186c74821e142bcd0766c8a043300f4da588108c3a] <==
	* 2023/08/30 22:57:05 GCP Auth Webhook started!
	2023/08/30 22:57:16 Ready to marshal response ...
	2023/08/30 22:57:16 Ready to write response ...
	2023/08/30 22:57:18 Ready to marshal response ...
	2023/08/30 22:57:18 Ready to write response ...
	2023/08/30 22:57:19 Ready to marshal response ...
	2023/08/30 22:57:19 Ready to write response ...
	2023/08/30 22:57:19 Ready to marshal response ...
	2023/08/30 22:57:19 Ready to write response ...
	2023/08/30 22:57:29 Ready to marshal response ...
	2023/08/30 22:57:29 Ready to write response ...
	2023/08/30 22:57:39 Ready to marshal response ...
	2023/08/30 22:57:39 Ready to write response ...
	2023/08/30 22:57:39 Ready to marshal response ...
	2023/08/30 22:57:39 Ready to write response ...
	2023/08/30 22:58:01 Ready to marshal response ...
	2023/08/30 22:58:01 Ready to write response ...
	
	* 
	* ==> kernel <==
	*  22:58:14 up  7:40,  0 users,  load average: 2.17, 2.18, 2.27
	Linux addons-015166 5.15.0-1043-aws #48~20.04.1-Ubuntu SMP Wed Aug 16 18:32:42 UTC 2023 aarch64 aarch64 aarch64 GNU/Linux
	PRETTY_NAME="Ubuntu 22.04.3 LTS"
	
	* 
	* ==> kindnet [d13b68d97b05e1f052a3d690ae64cf0abaf5eca3b8e292e66001baf603e1d752] <==
	* I0830 22:56:18.910999       1 main.go:191] Failed to get nodes, retrying after error: Get "https://10.96.0.1:443/api/v1/nodes": dial tcp 10.96.0.1:443: i/o timeout
	I0830 22:56:18.925096       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:56:18.925171       1 main.go:227] handling current node
	I0830 22:56:28.984719       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:56:28.984743       1 main.go:227] handling current node
	I0830 22:56:38.999912       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:56:39.000043       1 main.go:227] handling current node
	I0830 22:56:49.004071       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:56:49.004099       1 main.go:227] handling current node
	I0830 22:56:59.018012       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:56:59.018039       1 main.go:227] handling current node
	I0830 22:57:09.022553       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:57:09.022631       1 main.go:227] handling current node
	I0830 22:57:19.046083       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:57:19.046135       1 main.go:227] handling current node
	I0830 22:57:29.057239       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:57:29.057333       1 main.go:227] handling current node
	I0830 22:57:39.070064       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:57:39.070098       1 main.go:227] handling current node
	I0830 22:57:49.078100       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:57:49.078128       1 main.go:227] handling current node
	I0830 22:57:59.089194       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:57:59.089223       1 main.go:227] handling current node
	I0830 22:58:09.093255       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 22:58:09.093286       1 main.go:227] handling current node
	
	* 
	* ==> kube-apiserver [4bb20589784e15f56ca13bf9ddcc489be39f6c6fba30fe528b12a4eca03b529c] <==
	* E0830 22:56:27.602293       1 available_controller.go:460] v1beta1.metrics.k8s.io failed with: failing or missing response from https://10.104.116.236:443/apis/metrics.k8s.io/v1beta1: Get "https://10.104.116.236:443/apis/metrics.k8s.io/v1beta1": dial tcp 10.104.116.236:443: connect: connection refused
	W0830 22:56:27.602441       1 handler_proxy.go:93] no RequestInfo found in the context
	E0830 22:56:27.602518       1 controller.go:143] Error updating APIService "v1beta1.metrics.k8s.io" with err: failed to download v1beta1.metrics.k8s.io: failed to retrieve openAPI spec, http error: ResponseCode: 503, Body: service unavailable
	, Header: map[Content-Type:[text/plain; charset=utf-8] X-Content-Type-Options:[nosniff]]
	E0830 22:56:27.603347       1 available_controller.go:460] v1beta1.metrics.k8s.io failed with: failing or missing response from https://10.104.116.236:443/apis/metrics.k8s.io/v1beta1: Get "https://10.104.116.236:443/apis/metrics.k8s.io/v1beta1": dial tcp 10.104.116.236:443: connect: connection refused
	I0830 22:56:27.603604       1 handler_discovery.go:337] DiscoveryManager: Failed to download discovery for kube-system/metrics-server:443: 503 error trying to reach service: dial tcp 10.104.116.236:443: connect: connection refused
	I0830 22:56:27.603617       1 handler.go:232] Adding GroupVersion metrics.k8s.io v1beta1 to ResourceManager
	E0830 22:56:27.609875       1 available_controller.go:460] v1beta1.metrics.k8s.io failed with: failing or missing response from https://10.104.116.236:443/apis/metrics.k8s.io/v1beta1: Get "https://10.104.116.236:443/apis/metrics.k8s.io/v1beta1": dial tcp 10.104.116.236:443: connect: connection refused
	I0830 22:56:27.732413       1 handler.go:232] Adding GroupVersion metrics.k8s.io v1beta1 to ResourceManager
	I0830 22:56:31.573927       1 handler.go:232] Adding GroupVersion metrics.k8s.io v1beta1 to ResourceManager
	http2: server: error reading preface from client 192.168.49.1:55888: read tcp 192.168.49.2:8443->192.168.49.1:55888: read: connection reset by peer
	E0830 22:57:17.541660       1 controller.go:159] removing "v1beta1.metrics.k8s.io" from AggregationController failed with: resource not found
	I0830 22:57:19.004417       1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.102.174.55"}
	E0830 22:57:19.976348       1 watch.go:287] unable to encode watch object *v1.WatchEvent: http2: stream closed (&streaming.encoderWithAllocator{writer:responsewriter.outerWithCloseNotifyAndFlush{UserProvidedDecorator:(*metrics.ResponseWriterDelegator)(0x400b446b40), InnerCloseNotifierFlusher:struct { httpsnoop.Unwrapper; http.ResponseWriter; http.Flusher; http.CloseNotifier; http.Pusher }{Unwrapper:(*httpsnoop.rw)(0x400e4fb270), ResponseWriter:(*httpsnoop.rw)(0x400e4fb270), Flusher:(*httpsnoop.rw)(0x400e4fb270), CloseNotifier:(*httpsnoop.rw)(0x400e4fb270), Pusher:(*httpsnoop.rw)(0x400e4fb270)}}, encoder:(*versioning.codec)(0x400db13d60), memAllocator:(*runtime.Allocator)(0x400f030b58)})
	E0830 22:57:28.615265       1 handler_proxy.go:137] error resolving kube-system/metrics-server: service "metrics-server" not found
	W0830 22:57:28.615295       1 handler_proxy.go:93] no RequestInfo found in the context
	E0830 22:57:28.615328       1 controller.go:113] loading OpenAPI spec for "v1beta1.metrics.k8s.io" failed with: failed to retrieve openAPI spec, http error: ResponseCode: 503, Body: service unavailable
	, Header: map[Content-Type:[text/plain; charset=utf-8] X-Content-Type-Options:[nosniff]]
	I0830 22:57:28.615336       1 controller.go:126] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Rate Limited Requeue.
	I0830 22:57:29.596407       1 controller.go:624] quota admission added evaluator for: ingresses.networking.k8s.io
	I0830 22:57:29.905689       1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.108.157.67"}
	I0830 22:57:39.936735       1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.103.138.78"}
	I0830 22:57:50.776153       1 controller.go:624] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
	E0830 22:57:57.272802       1 authentication.go:70] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"ingress-nginx\" not found]"
	E0830 22:57:57.753413       1 authentication.go:70] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"ingress-nginx\" not found]"
	
	* 
	* ==> kube-controller-manager [b1afaa34c01017de27552235b2fb893c1bf242a2a0ca49aa160cad3ce5fe48a3] <==
	* I0830 22:57:21.114706       1 replica_set.go:676] "Finished syncing" kind="ReplicationController" key="kube-system/registry" duration="10.108µs"
	I0830 22:57:21.758121       1 event.go:307] "Event occurred" object="default/hpvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'hostpath.csi.k8s.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
	I0830 22:57:23.888380       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="headlamp/headlamp-699c48fb74" duration="42.437µs"
	I0830 22:57:23.929118       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="headlamp/headlamp-699c48fb74" duration="17.056639ms"
	I0830 22:57:23.929257       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="headlamp/headlamp-699c48fb74" duration="55.287µs"
	I0830 22:57:31.465200       1 event.go:307] "Event occurred" object="default/hpvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'hostpath.csi.k8s.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
	I0830 22:57:38.450971       1 event.go:307] "Event occurred" object="default/hpvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'hostpath.csi.k8s.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
	I0830 22:57:39.683098       1 event.go:307] "Event occurred" object="default/hello-world-app" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set hello-world-app-5d77478584 to 1"
	I0830 22:57:39.732730       1 event.go:307] "Event occurred" object="default/hello-world-app-5d77478584" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: hello-world-app-5d77478584-8crfr"
	I0830 22:57:39.743235       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="59.488006ms"
	I0830 22:57:39.792360       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="48.95346ms"
	I0830 22:57:39.793399       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="35.832µs"
	I0830 22:57:45.979402       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="33.485µs"
	I0830 22:57:46.982547       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="44.496µs"
	I0830 22:57:47.981852       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="44.955µs"
	I0830 22:57:52.430561       1 event.go:307] "Event occurred" object="default/hpvc-restore" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'hostpath.csi.k8s.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
	I0830 22:57:57.142408       1 job_controller.go:562] "enqueueing job" key="ingress-nginx/ingress-nginx-admission-create"
	I0830 22:57:57.156767       1 job_controller.go:562] "enqueueing job" key="ingress-nginx/ingress-nginx-admission-patch"
	I0830 22:57:57.157722       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="ingress-nginx/ingress-nginx-controller-5dcd45b5bf" duration="4.85µs"
	I0830 22:58:00.916651       1 event.go:307] "Event occurred" object="default/hpvc-restore" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'hostpath.csi.k8s.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
	I0830 22:58:01.036638       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="67.455µs"
	I0830 22:58:07.376751       1 namespace_controller.go:182] "Namespace has been deleted" namespace="ingress-nginx"
	I0830 22:58:12.299569       1 stateful_set.go:458] "StatefulSet has been deleted" key="kube-system/csi-hostpath-attacher"
	I0830 22:58:12.419191       1 stateful_set.go:458] "StatefulSet has been deleted" key="kube-system/csi-hostpath-resizer"
	I0830 22:58:14.166442       1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="default/hello-world-app-5d77478584" duration="67.511µs"
	
	* 
	* ==> kube-proxy [c2ca89ac2f5fe88c90af377524e8eb9b6261c0306583f26fef191aa3353bf792] <==
	* I0830 22:55:48.546978       1 server_others.go:69] "Using iptables proxy"
	I0830 22:55:48.567373       1 node.go:141] Successfully retrieved node IP: 192.168.49.2
	I0830 22:55:48.644154       1 server.go:632] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
	I0830 22:55:48.649866       1 server_others.go:152] "Using iptables Proxier"
	I0830 22:55:48.649911       1 server_others.go:421] "Detect-local-mode set to ClusterCIDR, but no cluster CIDR for family" ipFamily="IPv6"
	I0830 22:55:48.649920       1 server_others.go:438] "Defaulting to no-op detect-local"
	I0830 22:55:48.649971       1 proxier.go:251] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses"
	I0830 22:55:48.650284       1 server.go:846] "Version info" version="v1.28.1"
	I0830 22:55:48.650295       1 server.go:848] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
	I0830 22:55:48.652838       1 config.go:188] "Starting service config controller"
	I0830 22:55:48.652854       1 shared_informer.go:311] Waiting for caches to sync for service config
	I0830 22:55:48.652899       1 config.go:97] "Starting endpoint slice config controller"
	I0830 22:55:48.652904       1 shared_informer.go:311] Waiting for caches to sync for endpoint slice config
	I0830 22:55:48.661697       1 config.go:315] "Starting node config controller"
	I0830 22:55:48.661717       1 shared_informer.go:311] Waiting for caches to sync for node config
	I0830 22:55:48.753705       1 shared_informer.go:318] Caches are synced for endpoint slice config
	I0830 22:55:48.753789       1 shared_informer.go:318] Caches are synced for service config
	I0830 22:55:48.761966       1 shared_informer.go:318] Caches are synced for node config
	
	* 
	* ==> kube-scheduler [f8c96f4a21ae6c59bb1386e79dc6453d8a3bf23bfdf1ddae4df793a07bf1bdc2] <==
	* W0830 22:55:31.754771       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
	E0830 22:55:31.754789       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
	W0830 22:55:31.754844       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
	E0830 22:55:31.754853       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
	W0830 22:55:31.754893       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
	E0830 22:55:31.754909       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
	W0830 22:55:31.755836       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
	E0830 22:55:31.755951       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
	W0830 22:55:32.583799       1 reflector.go:535] pkg/server/dynamiccertificates/configmap_cafile_content.go:206: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
	E0830 22:55:32.583832       1 reflector.go:147] pkg/server/dynamiccertificates/configmap_cafile_content.go:206: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
	W0830 22:55:32.597610       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
	E0830 22:55:32.597655       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
	W0830 22:55:32.659885       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
	E0830 22:55:32.659927       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
	W0830 22:55:32.689468       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
	E0830 22:55:32.689511       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
	W0830 22:55:32.706333       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
	E0830 22:55:32.706369       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
	W0830 22:55:32.783443       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
	E0830 22:55:32.783668       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
	W0830 22:55:32.795179       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
	E0830 22:55:32.795225       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
	W0830 22:55:32.849315       1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
	E0830 22:55:32.849529       1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
	I0830 22:55:34.745079       1 shared_informer.go:318] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
	
	* 
	* ==> kubelet <==
	* Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.258313    1354 scope.go:117] "RemoveContainer" containerID="4f3b71078fed853badeefa53db64ff3271fd56f60fc4c18a36fbc1cb42e31e93"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.259044    1354 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"4f3b71078fed853badeefa53db64ff3271fd56f60fc4c18a36fbc1cb42e31e93"} err="failed to get container status \"4f3b71078fed853badeefa53db64ff3271fd56f60fc4c18a36fbc1cb42e31e93\": rpc error: code = NotFound desc = an error occurred when try to find container \"4f3b71078fed853badeefa53db64ff3271fd56f60fc4c18a36fbc1cb42e31e93\": not found"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.259067    1354 scope.go:117] "RemoveContainer" containerID="1739a952fc4f06711af2c0152a0edff17e3b6d138943f516e709f4aeecf70bda"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.259779    1354 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"1739a952fc4f06711af2c0152a0edff17e3b6d138943f516e709f4aeecf70bda"} err="failed to get container status \"1739a952fc4f06711af2c0152a0edff17e3b6d138943f516e709f4aeecf70bda\": rpc error: code = NotFound desc = an error occurred when try to find container \"1739a952fc4f06711af2c0152a0edff17e3b6d138943f516e709f4aeecf70bda\": not found"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.259799    1354 scope.go:117] "RemoveContainer" containerID="5f5359cdcba5e18f2c51309a012aa4e5a31bfda48186661dee257e4dd5dcdcfe"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.260480    1354 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"5f5359cdcba5e18f2c51309a012aa4e5a31bfda48186661dee257e4dd5dcdcfe"} err="failed to get container status \"5f5359cdcba5e18f2c51309a012aa4e5a31bfda48186661dee257e4dd5dcdcfe\": rpc error: code = NotFound desc = an error occurred when try to find container \"5f5359cdcba5e18f2c51309a012aa4e5a31bfda48186661dee257e4dd5dcdcfe\": not found"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.260500    1354 scope.go:117] "RemoveContainer" containerID="0f8e9a2b059d951a3f84c66694048b21ffcf5c5a1819b45499347b75760cc135"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.261311    1354 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"0f8e9a2b059d951a3f84c66694048b21ffcf5c5a1819b45499347b75760cc135"} err="failed to get container status \"0f8e9a2b059d951a3f84c66694048b21ffcf5c5a1819b45499347b75760cc135\": rpc error: code = NotFound desc = an error occurred when try to find container \"0f8e9a2b059d951a3f84c66694048b21ffcf5c5a1819b45499347b75760cc135\": not found"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.261333    1354 scope.go:117] "RemoveContainer" containerID="3bec36b5d47fdd1526c0129bcc9db5a468deef3fbe44e594f32bd644488bafc4"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.262392    1354 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"3bec36b5d47fdd1526c0129bcc9db5a468deef3fbe44e594f32bd644488bafc4"} err="failed to get container status \"3bec36b5d47fdd1526c0129bcc9db5a468deef3fbe44e594f32bd644488bafc4\": rpc error: code = NotFound desc = an error occurred when try to find container \"3bec36b5d47fdd1526c0129bcc9db5a468deef3fbe44e594f32bd644488bafc4\": not found"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.262417    1354 scope.go:117] "RemoveContainer" containerID="da433db2dae6c84c73764f99474c7a81f4b7ce3928d0f1c51810dde311aaaf27"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.263401    1354 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"da433db2dae6c84c73764f99474c7a81f4b7ce3928d0f1c51810dde311aaaf27"} err="failed to get container status \"da433db2dae6c84c73764f99474c7a81f4b7ce3928d0f1c51810dde311aaaf27\": rpc error: code = NotFound desc = an error occurred when try to find container \"da433db2dae6c84c73764f99474c7a81f4b7ce3928d0f1c51810dde311aaaf27\": not found"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.263425    1354 scope.go:117] "RemoveContainer" containerID="43ed6efb3f3466e5878e9ae1a5dc109e5e48f86413e8e247e8142be68e980a71"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.282148    1354 scope.go:117] "RemoveContainer" containerID="43ed6efb3f3466e5878e9ae1a5dc109e5e48f86413e8e247e8142be68e980a71"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: E0830 22:58:13.283125    1354 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"43ed6efb3f3466e5878e9ae1a5dc109e5e48f86413e8e247e8142be68e980a71\": not found" containerID="43ed6efb3f3466e5878e9ae1a5dc109e5e48f86413e8e247e8142be68e980a71"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.283256    1354 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"43ed6efb3f3466e5878e9ae1a5dc109e5e48f86413e8e247e8142be68e980a71"} err="failed to get container status \"43ed6efb3f3466e5878e9ae1a5dc109e5e48f86413e8e247e8142be68e980a71\": rpc error: code = NotFound desc = an error occurred when try to find container \"43ed6efb3f3466e5878e9ae1a5dc109e5e48f86413e8e247e8142be68e980a71\": not found"
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.310003    1354 reconciler_common.go:300] "Volume detached for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/844fd9be-03a2-438d-a95d-2f912c388b13-socket-dir\") on node \"addons-015166\" DevicePath \"\""
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.310247    1354 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-pdlsw\" (UniqueName: \"kubernetes.io/projected/844fd9be-03a2-438d-a95d-2f912c388b13-kube-api-access-pdlsw\") on node \"addons-015166\" DevicePath \"\""
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.310553    1354 reconciler_common.go:300] "Volume detached for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/844fd9be-03a2-438d-a95d-2f912c388b13-mountpoint-dir\") on node \"addons-015166\" DevicePath \"\""
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.310699    1354 reconciler_common.go:300] "Volume detached for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/844fd9be-03a2-438d-a95d-2f912c388b13-plugins-dir\") on node \"addons-015166\" DevicePath \"\""
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.311149    1354 reconciler_common.go:300] "Volume detached for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/844fd9be-03a2-438d-a95d-2f912c388b13-csi-data-dir\") on node \"addons-015166\" DevicePath \"\""
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.311396    1354 reconciler_common.go:300] "Volume detached for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/844fd9be-03a2-438d-a95d-2f912c388b13-registration-dir\") on node \"addons-015166\" DevicePath \"\""
	Aug 30 22:58:13 addons-015166 kubelet[1354]: I0830 22:58:13.311432    1354 reconciler_common.go:300] "Volume detached for volume \"dev-dir\" (UniqueName: \"kubernetes.io/host-path/844fd9be-03a2-438d-a95d-2f912c388b13-dev-dir\") on node \"addons-015166\" DevicePath \"\""
	Aug 30 22:58:14 addons-015166 kubelet[1354]: I0830 22:58:14.151149    1354 scope.go:117] "RemoveContainer" containerID="d7ea4da00dd6813494e731791191d666ebcce4f54bef73015e06138e6e9f348f"
	Aug 30 22:58:14 addons-015166 kubelet[1354]: E0830 22:58:14.151426    1354 pod_workers.go:1300] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"hello-world-app\" with CrashLoopBackOff: \"back-off 20s restarting failed container=hello-world-app pod=hello-world-app-5d77478584-8crfr_default(e2b4ef71-b8ee-4e7b-972e-86d452125a20)\"" pod="default/hello-world-app-5d77478584-8crfr" podUID="e2b4ef71-b8ee-4e7b-972e-86d452125a20"
	
	* 
	* ==> storage-provisioner [9a2c72414e97ce1265d3ce584e33a045d4497ceeda30157f012d7d7c46ff1009] <==
	* I0830 22:55:52.894996       1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
	I0830 22:55:52.913839       1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
	I0830 22:55:52.913943       1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
	I0830 22:55:52.924172       1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
	I0830 22:55:52.925739       1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"2bf5f7d4-2bb7-413e-bc87-109068086fbd", APIVersion:"v1", ResourceVersion:"552", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-015166_d64edb70-8847-44f8-90a3-c93ec07970cd became leader
	I0830 22:55:52.926314       1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-015166_d64edb70-8847-44f8-90a3-c93ec07970cd!
	I0830 22:55:53.028856       1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-015166_d64edb70-8847-44f8-90a3-c93ec07970cd!
	

                                                
                                                
-- /stdout --
helpers_test.go:254: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p addons-015166 -n addons-015166
helpers_test.go:261: (dbg) Run:  kubectl --context addons-015166 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:285: <<< TestAddons/parallel/InspektorGadget FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/InspektorGadget (8.59s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageLoadDaemon (4.1s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageLoadDaemon
functional_test.go:354: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image load --daemon gcr.io/google-containers/addon-resizer:functional-479614 --alsologtostderr
functional_test.go:354: (dbg) Done: out/minikube-linux-arm64 -p functional-479614 image load --daemon gcr.io/google-containers/addon-resizer:functional-479614 --alsologtostderr: (3.796801077s)
functional_test.go:447: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image ls
functional_test.go:442: expected "gcr.io/google-containers/addon-resizer:functional-479614" to be loaded into minikube but the image is not there
--- FAIL: TestFunctional/parallel/ImageCommands/ImageLoadDaemon (4.10s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageReloadDaemon (3.52s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageReloadDaemon
functional_test.go:364: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image load --daemon gcr.io/google-containers/addon-resizer:functional-479614 --alsologtostderr
E0830 23:02:47.108554 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
functional_test.go:364: (dbg) Done: out/minikube-linux-arm64 -p functional-479614 image load --daemon gcr.io/google-containers/addon-resizer:functional-479614 --alsologtostderr: (3.269170556s)
functional_test.go:447: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image ls
functional_test.go:442: expected "gcr.io/google-containers/addon-resizer:functional-479614" to be loaded into minikube but the image is not there
--- FAIL: TestFunctional/parallel/ImageCommands/ImageReloadDaemon (3.52s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageTagAndLoadDaemon (5.34s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageTagAndLoadDaemon
functional_test.go:234: (dbg) Run:  docker pull gcr.io/google-containers/addon-resizer:1.8.9
functional_test.go:234: (dbg) Done: docker pull gcr.io/google-containers/addon-resizer:1.8.9: (1.735188243s)
functional_test.go:239: (dbg) Run:  docker tag gcr.io/google-containers/addon-resizer:1.8.9 gcr.io/google-containers/addon-resizer:functional-479614
functional_test.go:244: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image load --daemon gcr.io/google-containers/addon-resizer:functional-479614 --alsologtostderr
functional_test.go:244: (dbg) Done: out/minikube-linux-arm64 -p functional-479614 image load --daemon gcr.io/google-containers/addon-resizer:functional-479614 --alsologtostderr: (3.268446631s)
functional_test.go:447: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image ls
functional_test.go:442: expected "gcr.io/google-containers/addon-resizer:functional-479614" to be loaded into minikube but the image is not there
--- FAIL: TestFunctional/parallel/ImageCommands/ImageTagAndLoadDaemon (5.34s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageSaveToFile (0.7s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageSaveToFile
functional_test.go:379: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image save gcr.io/google-containers/addon-resizer:functional-479614 /home/jenkins/workspace/Docker_Linux_containerd_arm64/addon-resizer-save.tar --alsologtostderr
functional_test.go:385: expected "/home/jenkins/workspace/Docker_Linux_containerd_arm64/addon-resizer-save.tar" to exist after `image save`, but doesn't exist
--- FAIL: TestFunctional/parallel/ImageCommands/ImageSaveToFile (0.70s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageLoadFromFile (0.29s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageLoadFromFile
functional_test.go:408: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image load /home/jenkins/workspace/Docker_Linux_containerd_arm64/addon-resizer-save.tar --alsologtostderr
functional_test.go:410: loading image into minikube from file: <nil>

                                                
                                                
** stderr ** 
	I0830 23:02:57.187070 1254810 out.go:296] Setting OutFile to fd 1 ...
	I0830 23:02:57.187694 1254810 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:02:57.187704 1254810 out.go:309] Setting ErrFile to fd 2...
	I0830 23:02:57.187710 1254810 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:02:57.188107 1254810 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
	I0830 23:02:57.188866 1254810 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
	I0830 23:02:57.188991 1254810 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
	I0830 23:02:57.189559 1254810 cli_runner.go:164] Run: docker container inspect functional-479614 --format={{.State.Status}}
	I0830 23:02:57.209703 1254810 ssh_runner.go:195] Run: systemctl --version
	I0830 23:02:57.209811 1254810 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" functional-479614
	I0830 23:02:57.255247 1254810 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34329 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/functional-479614/id_rsa Username:docker}
	I0830 23:02:57.375476 1254810 cache_images.go:286] Loading image from: /home/jenkins/workspace/Docker_Linux_containerd_arm64/addon-resizer-save.tar
	W0830 23:02:57.375571 1254810 cache_images.go:254] Failed to load cached images for profile functional-479614. make sure the profile is running. loading images: stat /home/jenkins/workspace/Docker_Linux_containerd_arm64/addon-resizer-save.tar: no such file or directory
	I0830 23:02:57.375594 1254810 cache_images.go:262] succeeded pushing to: 
	I0830 23:02:57.375599 1254810 cache_images.go:263] failed pushing to: functional-479614

                                                
                                                
** /stderr **
--- FAIL: TestFunctional/parallel/ImageCommands/ImageLoadFromFile (0.29s)

                                                
                                    
x
+
TestIngressAddonLegacy/serial/ValidateIngressAddons (54.97s)

                                                
                                                
=== RUN   TestIngressAddonLegacy/serial/ValidateIngressAddons
addons_test.go:183: (dbg) Run:  kubectl --context ingress-addon-legacy-235029 wait --for=condition=ready --namespace=ingress-nginx pod --selector=app.kubernetes.io/component=controller --timeout=90s
addons_test.go:183: (dbg) Done: kubectl --context ingress-addon-legacy-235029 wait --for=condition=ready --namespace=ingress-nginx pod --selector=app.kubernetes.io/component=controller --timeout=90s: (10.123648417s)
addons_test.go:208: (dbg) Run:  kubectl --context ingress-addon-legacy-235029 replace --force -f testdata/nginx-ingress-v1beta1.yaml
addons_test.go:221: (dbg) Run:  kubectl --context ingress-addon-legacy-235029 replace --force -f testdata/nginx-pod-svc.yaml
addons_test.go:226: (dbg) TestIngressAddonLegacy/serial/ValidateIngressAddons: waiting 8m0s for pods matching "run=nginx" in namespace "default" ...
helpers_test.go:344: "nginx" [d67063a3-12f1-4ab5-9a70-2f7b3754ef9e] Pending / Ready:ContainersNotReady (containers with unready status: [nginx]) / ContainersReady:ContainersNotReady (containers with unready status: [nginx])
helpers_test.go:344: "nginx" [d67063a3-12f1-4ab5-9a70-2f7b3754ef9e] Running
addons_test.go:226: (dbg) TestIngressAddonLegacy/serial/ValidateIngressAddons: run=nginx healthy within 8.017195499s
addons_test.go:238: (dbg) Run:  out/minikube-linux-arm64 -p ingress-addon-legacy-235029 ssh "curl -s http://127.0.0.1/ -H 'Host: nginx.example.com'"
addons_test.go:262: (dbg) Run:  kubectl --context ingress-addon-legacy-235029 replace --force -f testdata/ingress-dns-example-v1beta1.yaml
addons_test.go:267: (dbg) Run:  out/minikube-linux-arm64 -p ingress-addon-legacy-235029 ip
addons_test.go:273: (dbg) Run:  nslookup hello-john.test 192.168.49.2
addons_test.go:273: (dbg) Non-zero exit: nslookup hello-john.test 192.168.49.2: exit status 1 (15.021304822s)

                                                
                                                
-- stdout --
	;; connection timed out; no servers could be reached
	
	

                                                
                                                
-- /stdout --
addons_test.go:275: failed to nslookup hello-john.test host. args "nslookup hello-john.test 192.168.49.2" : exit status 1
addons_test.go:279: unexpected output from nslookup. stdout: ;; connection timed out; no servers could be reached

                                                
                                                

                                                
                                                

                                                
                                                
stderr: 
addons_test.go:282: (dbg) Run:  out/minikube-linux-arm64 -p ingress-addon-legacy-235029 addons disable ingress-dns --alsologtostderr -v=1
addons_test.go:282: (dbg) Done: out/minikube-linux-arm64 -p ingress-addon-legacy-235029 addons disable ingress-dns --alsologtostderr -v=1: (10.293652648s)
addons_test.go:287: (dbg) Run:  out/minikube-linux-arm64 -p ingress-addon-legacy-235029 addons disable ingress --alsologtostderr -v=1
addons_test.go:287: (dbg) Done: out/minikube-linux-arm64 -p ingress-addon-legacy-235029 addons disable ingress --alsologtostderr -v=1: (7.582197945s)
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======>  post-mortem[TestIngressAddonLegacy/serial/ValidateIngressAddons]: docker inspect <======
helpers_test.go:231: (dbg) Run:  docker inspect ingress-addon-legacy-235029
helpers_test.go:235: (dbg) docker inspect ingress-addon-legacy-235029:

                                                
                                                
-- stdout --
	[
	    {
	        "Id": "d0754b4eaea7ed9494c78aee30c140d34b81e7b084b69787ecf813e2d00561f7",
	        "Created": "2023-08-30T23:04:00.104532781Z",
	        "Path": "/usr/local/bin/entrypoint",
	        "Args": [
	            "/sbin/init"
	        ],
	        "State": {
	            "Status": "running",
	            "Running": true,
	            "Paused": false,
	            "Restarting": false,
	            "OOMKilled": false,
	            "Dead": false,
	            "Pid": 1259413,
	            "ExitCode": 0,
	            "Error": "",
	            "StartedAt": "2023-08-30T23:04:00.700571734Z",
	            "FinishedAt": "0001-01-01T00:00:00Z"
	        },
	        "Image": "sha256:879c6efc994c345ac84dd4ebb4fc5b49dd2a4b340e335879382e51233f79b51a",
	        "ResolvConfPath": "/var/lib/docker/containers/d0754b4eaea7ed9494c78aee30c140d34b81e7b084b69787ecf813e2d00561f7/resolv.conf",
	        "HostnamePath": "/var/lib/docker/containers/d0754b4eaea7ed9494c78aee30c140d34b81e7b084b69787ecf813e2d00561f7/hostname",
	        "HostsPath": "/var/lib/docker/containers/d0754b4eaea7ed9494c78aee30c140d34b81e7b084b69787ecf813e2d00561f7/hosts",
	        "LogPath": "/var/lib/docker/containers/d0754b4eaea7ed9494c78aee30c140d34b81e7b084b69787ecf813e2d00561f7/d0754b4eaea7ed9494c78aee30c140d34b81e7b084b69787ecf813e2d00561f7-json.log",
	        "Name": "/ingress-addon-legacy-235029",
	        "RestartCount": 0,
	        "Driver": "overlay2",
	        "Platform": "linux",
	        "MountLabel": "",
	        "ProcessLabel": "",
	        "AppArmorProfile": "unconfined",
	        "ExecIDs": null,
	        "HostConfig": {
	            "Binds": [
	                "/lib/modules:/lib/modules:ro",
	                "ingress-addon-legacy-235029:/var"
	            ],
	            "ContainerIDFile": "",
	            "LogConfig": {
	                "Type": "json-file",
	                "Config": {}
	            },
	            "NetworkMode": "ingress-addon-legacy-235029",
	            "PortBindings": {
	                "22/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "2376/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "32443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "5000/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "8443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ]
	            },
	            "RestartPolicy": {
	                "Name": "no",
	                "MaximumRetryCount": 0
	            },
	            "AutoRemove": false,
	            "VolumeDriver": "",
	            "VolumesFrom": null,
	            "ConsoleSize": [
	                0,
	                0
	            ],
	            "CapAdd": null,
	            "CapDrop": null,
	            "CgroupnsMode": "host",
	            "Dns": [],
	            "DnsOptions": [],
	            "DnsSearch": [],
	            "ExtraHosts": null,
	            "GroupAdd": null,
	            "IpcMode": "private",
	            "Cgroup": "",
	            "Links": null,
	            "OomScoreAdj": 0,
	            "PidMode": "",
	            "Privileged": true,
	            "PublishAllPorts": false,
	            "ReadonlyRootfs": false,
	            "SecurityOpt": [
	                "seccomp=unconfined",
	                "apparmor=unconfined",
	                "label=disable"
	            ],
	            "Tmpfs": {
	                "/run": "",
	                "/tmp": ""
	            },
	            "UTSMode": "",
	            "UsernsMode": "",
	            "ShmSize": 67108864,
	            "Runtime": "runc",
	            "Isolation": "",
	            "CpuShares": 0,
	            "Memory": 4294967296,
	            "NanoCpus": 2000000000,
	            "CgroupParent": "",
	            "BlkioWeight": 0,
	            "BlkioWeightDevice": [],
	            "BlkioDeviceReadBps": [],
	            "BlkioDeviceWriteBps": [],
	            "BlkioDeviceReadIOps": [],
	            "BlkioDeviceWriteIOps": [],
	            "CpuPeriod": 0,
	            "CpuQuota": 0,
	            "CpuRealtimePeriod": 0,
	            "CpuRealtimeRuntime": 0,
	            "CpusetCpus": "",
	            "CpusetMems": "",
	            "Devices": [],
	            "DeviceCgroupRules": null,
	            "DeviceRequests": null,
	            "MemoryReservation": 0,
	            "MemorySwap": 8589934592,
	            "MemorySwappiness": null,
	            "OomKillDisable": false,
	            "PidsLimit": null,
	            "Ulimits": null,
	            "CpuCount": 0,
	            "CpuPercent": 0,
	            "IOMaximumIOps": 0,
	            "IOMaximumBandwidth": 0,
	            "MaskedPaths": null,
	            "ReadonlyPaths": null
	        },
	        "GraphDriver": {
	            "Data": {
	                "LowerDir": "/var/lib/docker/overlay2/743fda40a6dd07c63aae7f34a2d2e0b6f60c91e0a69406293a5515d31ffaac47-init/diff:/var/lib/docker/overlay2/ecca48d9e9e3ae5c70dec79bae341547fcaf05f85377796ae2cd1184b3d7168f/diff",
	                "MergedDir": "/var/lib/docker/overlay2/743fda40a6dd07c63aae7f34a2d2e0b6f60c91e0a69406293a5515d31ffaac47/merged",
	                "UpperDir": "/var/lib/docker/overlay2/743fda40a6dd07c63aae7f34a2d2e0b6f60c91e0a69406293a5515d31ffaac47/diff",
	                "WorkDir": "/var/lib/docker/overlay2/743fda40a6dd07c63aae7f34a2d2e0b6f60c91e0a69406293a5515d31ffaac47/work"
	            },
	            "Name": "overlay2"
	        },
	        "Mounts": [
	            {
	                "Type": "bind",
	                "Source": "/lib/modules",
	                "Destination": "/lib/modules",
	                "Mode": "ro",
	                "RW": false,
	                "Propagation": "rprivate"
	            },
	            {
	                "Type": "volume",
	                "Name": "ingress-addon-legacy-235029",
	                "Source": "/var/lib/docker/volumes/ingress-addon-legacy-235029/_data",
	                "Destination": "/var",
	                "Driver": "local",
	                "Mode": "z",
	                "RW": true,
	                "Propagation": ""
	            }
	        ],
	        "Config": {
	            "Hostname": "ingress-addon-legacy-235029",
	            "Domainname": "",
	            "User": "",
	            "AttachStdin": false,
	            "AttachStdout": false,
	            "AttachStderr": false,
	            "ExposedPorts": {
	                "22/tcp": {},
	                "2376/tcp": {},
	                "32443/tcp": {},
	                "5000/tcp": {},
	                "8443/tcp": {}
	            },
	            "Tty": true,
	            "OpenStdin": false,
	            "StdinOnce": false,
	            "Env": [
	                "container=docker",
	                "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
	            ],
	            "Cmd": null,
	            "Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec",
	            "Volumes": null,
	            "WorkingDir": "/",
	            "Entrypoint": [
	                "/usr/local/bin/entrypoint",
	                "/sbin/init"
	            ],
	            "OnBuild": null,
	            "Labels": {
	                "created_by.minikube.sigs.k8s.io": "true",
	                "mode.minikube.sigs.k8s.io": "ingress-addon-legacy-235029",
	                "name.minikube.sigs.k8s.io": "ingress-addon-legacy-235029",
	                "role.minikube.sigs.k8s.io": ""
	            },
	            "StopSignal": "SIGRTMIN+3"
	        },
	        "NetworkSettings": {
	            "Bridge": "",
	            "SandboxID": "2f27f0d082a3b8784b0c4d62533b50cf49710e9b869b696ff47a71cc6e2ee688",
	            "HairpinMode": false,
	            "LinkLocalIPv6Address": "",
	            "LinkLocalIPv6PrefixLen": 0,
	            "Ports": {
	                "22/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34334"
	                    }
	                ],
	                "2376/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34333"
	                    }
	                ],
	                "32443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34330"
	                    }
	                ],
	                "5000/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34332"
	                    }
	                ],
	                "8443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34331"
	                    }
	                ]
	            },
	            "SandboxKey": "/var/run/docker/netns/2f27f0d082a3",
	            "SecondaryIPAddresses": null,
	            "SecondaryIPv6Addresses": null,
	            "EndpointID": "",
	            "Gateway": "",
	            "GlobalIPv6Address": "",
	            "GlobalIPv6PrefixLen": 0,
	            "IPAddress": "",
	            "IPPrefixLen": 0,
	            "IPv6Gateway": "",
	            "MacAddress": "",
	            "Networks": {
	                "ingress-addon-legacy-235029": {
	                    "IPAMConfig": {
	                        "IPv4Address": "192.168.49.2"
	                    },
	                    "Links": null,
	                    "Aliases": [
	                        "d0754b4eaea7",
	                        "ingress-addon-legacy-235029"
	                    ],
	                    "NetworkID": "6219c93885f18623fe4a6fcba0bc45fd6e661175073de5072697628f85420e65",
	                    "EndpointID": "ec8113e396256c762b2131b3b26cdd10abaadd3762a00ea996694523e44e767c",
	                    "Gateway": "192.168.49.1",
	                    "IPAddress": "192.168.49.2",
	                    "IPPrefixLen": 24,
	                    "IPv6Gateway": "",
	                    "GlobalIPv6Address": "",
	                    "GlobalIPv6PrefixLen": 0,
	                    "MacAddress": "02:42:c0:a8:31:02",
	                    "DriverOpts": null
	                }
	            }
	        }
	    }
	]

                                                
                                                
-- /stdout --
helpers_test.go:239: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p ingress-addon-legacy-235029 -n ingress-addon-legacy-235029
helpers_test.go:244: <<< TestIngressAddonLegacy/serial/ValidateIngressAddons FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======>  post-mortem[TestIngressAddonLegacy/serial/ValidateIngressAddons]: minikube logs <======
helpers_test.go:247: (dbg) Run:  out/minikube-linux-arm64 -p ingress-addon-legacy-235029 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-arm64 -p ingress-addon-legacy-235029 logs -n 25: (1.417622658s)
helpers_test.go:252: TestIngressAddonLegacy/serial/ValidateIngressAddons logs: 
-- stdout --
	* 
	* ==> Audit <==
	* |----------------|------------------------------------------------------------------------|-----------------------------|---------|---------|---------------------|---------------------|
	|    Command     |                                  Args                                  |           Profile           |  User   | Version |     Start Time      |      End Time       |
	|----------------|------------------------------------------------------------------------|-----------------------------|---------|---------|---------------------|---------------------|
	| mount          | -p functional-479614                                                   | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC |                     |
	|                | /tmp/TestFunctionalparallelMountCmdVerifyCleanup4213640551/001:/mount2 |                             |         |         |                     |                     |
	|                | --alsologtostderr -v=1                                                 |                             |         |         |                     |                     |
	| mount          | -p functional-479614                                                   | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC |                     |
	|                | /tmp/TestFunctionalparallelMountCmdVerifyCleanup4213640551/001:/mount1 |                             |         |         |                     |                     |
	|                | --alsologtostderr -v=1                                                 |                             |         |         |                     |                     |
	| ssh            | functional-479614 ssh findmnt                                          | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:03 UTC |
	|                | -T /mount1                                                             |                             |         |         |                     |                     |
	| mount          | -p functional-479614                                                   | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC |                     |
	|                | /tmp/TestFunctionalparallelMountCmdVerifyCleanup4213640551/001:/mount3 |                             |         |         |                     |                     |
	|                | --alsologtostderr -v=1                                                 |                             |         |         |                     |                     |
	| ssh            | functional-479614 ssh findmnt                                          | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:03 UTC |
	|                | -T /mount2                                                             |                             |         |         |                     |                     |
	| ssh            | functional-479614 ssh findmnt                                          | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:03 UTC |
	|                | -T /mount3                                                             |                             |         |         |                     |                     |
	| mount          | -p functional-479614                                                   | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC |                     |
	|                | --kill=true                                                            |                             |         |         |                     |                     |
	| update-context | functional-479614                                                      | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:03 UTC |
	|                | update-context                                                         |                             |         |         |                     |                     |
	|                | --alsologtostderr -v=2                                                 |                             |         |         |                     |                     |
	| update-context | functional-479614                                                      | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:03 UTC |
	|                | update-context                                                         |                             |         |         |                     |                     |
	|                | --alsologtostderr -v=2                                                 |                             |         |         |                     |                     |
	| update-context | functional-479614                                                      | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:03 UTC |
	|                | update-context                                                         |                             |         |         |                     |                     |
	|                | --alsologtostderr -v=2                                                 |                             |         |         |                     |                     |
	| image          | functional-479614                                                      | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:03 UTC |
	|                | image ls --format short                                                |                             |         |         |                     |                     |
	|                | --alsologtostderr                                                      |                             |         |         |                     |                     |
	| ssh            | functional-479614 ssh pgrep                                            | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC |                     |
	|                | buildkitd                                                              |                             |         |         |                     |                     |
	| image          | functional-479614 image build -t                                       | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:03 UTC |
	|                | localhost/my-image:functional-479614                                   |                             |         |         |                     |                     |
	|                | testdata/build --alsologtostderr                                       |                             |         |         |                     |                     |
	| image          | functional-479614                                                      | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:03 UTC |
	|                | image ls --format yaml                                                 |                             |         |         |                     |                     |
	|                | --alsologtostderr                                                      |                             |         |         |                     |                     |
	| image          | functional-479614                                                      | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:03 UTC |
	|                | image ls --format json                                                 |                             |         |         |                     |                     |
	|                | --alsologtostderr                                                      |                             |         |         |                     |                     |
	| image          | functional-479614                                                      | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:03 UTC |
	|                | image ls --format table                                                |                             |         |         |                     |                     |
	|                | --alsologtostderr                                                      |                             |         |         |                     |                     |
	| image          | functional-479614 image ls                                             | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:03 UTC |
	| delete         | -p functional-479614                                                   | functional-479614           | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:03 UTC |
	| start          | -p ingress-addon-legacy-235029                                         | ingress-addon-legacy-235029 | jenkins | v1.31.2 | 30 Aug 23 23:03 UTC | 30 Aug 23 23:05 UTC |
	|                | --kubernetes-version=v1.18.20                                          |                             |         |         |                     |                     |
	|                | --memory=4096 --wait=true                                              |                             |         |         |                     |                     |
	|                | --alsologtostderr                                                      |                             |         |         |                     |                     |
	|                | -v=5 --driver=docker                                                   |                             |         |         |                     |                     |
	|                | --container-runtime=containerd                                         |                             |         |         |                     |                     |
	| addons         | ingress-addon-legacy-235029                                            | ingress-addon-legacy-235029 | jenkins | v1.31.2 | 30 Aug 23 23:05 UTC | 30 Aug 23 23:05 UTC |
	|                | addons enable ingress                                                  |                             |         |         |                     |                     |
	|                | --alsologtostderr -v=5                                                 |                             |         |         |                     |                     |
	| addons         | ingress-addon-legacy-235029                                            | ingress-addon-legacy-235029 | jenkins | v1.31.2 | 30 Aug 23 23:05 UTC | 30 Aug 23 23:05 UTC |
	|                | addons enable ingress-dns                                              |                             |         |         |                     |                     |
	|                | --alsologtostderr -v=5                                                 |                             |         |         |                     |                     |
	| ssh            | ingress-addon-legacy-235029                                            | ingress-addon-legacy-235029 | jenkins | v1.31.2 | 30 Aug 23 23:05 UTC | 30 Aug 23 23:05 UTC |
	|                | ssh curl -s http://127.0.0.1/                                          |                             |         |         |                     |                     |
	|                | -H 'Host: nginx.example.com'                                           |                             |         |         |                     |                     |
	| ip             | ingress-addon-legacy-235029 ip                                         | ingress-addon-legacy-235029 | jenkins | v1.31.2 | 30 Aug 23 23:05 UTC | 30 Aug 23 23:05 UTC |
	| addons         | ingress-addon-legacy-235029                                            | ingress-addon-legacy-235029 | jenkins | v1.31.2 | 30 Aug 23 23:05 UTC | 30 Aug 23 23:06 UTC |
	|                | addons disable ingress-dns                                             |                             |         |         |                     |                     |
	|                | --alsologtostderr -v=1                                                 |                             |         |         |                     |                     |
	| addons         | ingress-addon-legacy-235029                                            | ingress-addon-legacy-235029 | jenkins | v1.31.2 | 30 Aug 23 23:06 UTC | 30 Aug 23 23:06 UTC |
	|                | addons disable ingress                                                 |                             |         |         |                     |                     |
	|                | --alsologtostderr -v=1                                                 |                             |         |         |                     |                     |
	|----------------|------------------------------------------------------------------------|-----------------------------|---------|---------|---------------------|---------------------|
	
	* 
	* ==> Last Start <==
	* Log file created at: 2023/08/30 23:03:38
	Running on machine: ip-172-31-31-251
	Binary: Built with gc go1.20.7 for linux/arm64
	Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
	I0830 23:03:38.869500 1258949 out.go:296] Setting OutFile to fd 1 ...
	I0830 23:03:38.869693 1258949 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:03:38.869705 1258949 out.go:309] Setting ErrFile to fd 2...
	I0830 23:03:38.869711 1258949 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:03:38.869992 1258949 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
	I0830 23:03:38.870577 1258949 out.go:303] Setting JSON to false
	I0830 23:03:38.871872 1258949 start.go:128] hostinfo: {"hostname":"ip-172-31-31-251","uptime":27953,"bootTime":1693408666,"procs":447,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1043-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"982e3628-3742-4b3e-bb63-ac1b07660ec7"}
	I0830 23:03:38.871954 1258949 start.go:138] virtualization:  
	I0830 23:03:38.874525 1258949 out.go:177] * [ingress-addon-legacy-235029] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	I0830 23:03:38.876717 1258949 out.go:177]   - MINIKUBE_LOCATION=17114
	I0830 23:03:38.876951 1258949 notify.go:220] Checking for updates...
	I0830 23:03:38.879929 1258949 out.go:177]   - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0830 23:03:38.881641 1258949 out.go:177]   - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	I0830 23:03:38.883475 1258949 out.go:177]   - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	I0830 23:03:38.885635 1258949 out.go:177]   - MINIKUBE_BIN=out/minikube-linux-arm64
	I0830 23:03:38.887567 1258949 out.go:177]   - MINIKUBE_FORCE_SYSTEMD=
	I0830 23:03:38.889653 1258949 driver.go:373] Setting default libvirt URI to qemu:///system
	I0830 23:03:38.914011 1258949 docker.go:121] docker version: linux-24.0.5:Docker Engine - Community
	I0830 23:03:38.914112 1258949 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 23:03:39.000353 1258949 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:24 OomKillDisable:true NGoroutines:36 SystemTime:2023-08-30 23:03:38.990195001 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 23:03:39.000478 1258949 docker.go:294] overlay module found
	I0830 23:03:39.002446 1258949 out.go:177] * Using the docker driver based on user configuration
	I0830 23:03:39.004177 1258949 start.go:298] selected driver: docker
	I0830 23:03:39.004209 1258949 start.go:902] validating driver "docker" against <nil>
	I0830 23:03:39.004223 1258949 start.go:913] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
	I0830 23:03:39.004967 1258949 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 23:03:39.086808 1258949 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:24 OomKillDisable:true NGoroutines:36 SystemTime:2023-08-30 23:03:39.071217693 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 23:03:39.086959 1258949 start_flags.go:305] no existing cluster config was found, will generate one from the flags 
	I0830 23:03:39.087178 1258949 start_flags.go:919] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
	I0830 23:03:39.090290 1258949 out.go:177] * Using Docker driver with root privileges
	I0830 23:03:39.092203 1258949 cni.go:84] Creating CNI manager for ""
	I0830 23:03:39.092240 1258949 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
	I0830 23:03:39.092265 1258949 start_flags.go:314] Found "CNI" CNI - setting NetworkPlugin=cni
	I0830 23:03:39.092282 1258949 start_flags.go:319] config:
	{Name:ingress-addon-legacy-235029 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:4096 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.18.20 ClusterName:ingress-addon-legacy-235029 Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.lo
cal ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0}
	I0830 23:03:39.094279 1258949 out.go:177] * Starting control plane node ingress-addon-legacy-235029 in cluster ingress-addon-legacy-235029
	I0830 23:03:39.096175 1258949 cache.go:122] Beginning downloading kic base image for docker with containerd
	I0830 23:03:39.097910 1258949 out.go:177] * Pulling base image ...
	I0830 23:03:39.099425 1258949 preload.go:132] Checking if preload exists for k8s version v1.18.20 and runtime containerd
	I0830 23:03:39.099460 1258949 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec in local docker daemon
	I0830 23:03:39.117857 1258949 image.go:83] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec in local docker daemon, skipping pull
	I0830 23:03:39.117883 1258949 cache.go:145] gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec exists in daemon, skipping load
	I0830 23:03:39.176384 1258949 preload.go:119] Found remote preload: https://storage.googleapis.com/minikube-preloaded-volume-tarballs/v18/v1.18.20/preloaded-images-k8s-v18-v1.18.20-containerd-overlay2-arm64.tar.lz4
	I0830 23:03:39.176414 1258949 cache.go:57] Caching tarball of preloaded images
	I0830 23:03:39.176624 1258949 preload.go:132] Checking if preload exists for k8s version v1.18.20 and runtime containerd
	I0830 23:03:39.178497 1258949 out.go:177] * Downloading Kubernetes v1.18.20 preload ...
	I0830 23:03:39.179982 1258949 preload.go:238] getting checksum for preloaded-images-k8s-v18-v1.18.20-containerd-overlay2-arm64.tar.lz4 ...
	I0830 23:03:39.297906 1258949 download.go:107] Downloading: https://storage.googleapis.com/minikube-preloaded-volume-tarballs/v18/v1.18.20/preloaded-images-k8s-v18-v1.18.20-containerd-overlay2-arm64.tar.lz4?checksum=md5:9e505be2989b8c051b1372c317471064 -> /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.18.20-containerd-overlay2-arm64.tar.lz4
	I0830 23:03:52.208041 1258949 preload.go:249] saving checksum for preloaded-images-k8s-v18-v1.18.20-containerd-overlay2-arm64.tar.lz4 ...
	I0830 23:03:52.208166 1258949 preload.go:256] verifying checksum of /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.18.20-containerd-overlay2-arm64.tar.lz4 ...
	I0830 23:03:53.349956 1258949 cache.go:60] Finished verifying existence of preloaded tar for  v1.18.20 on containerd
	I0830 23:03:53.350330 1258949 profile.go:148] Saving config to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/config.json ...
	I0830 23:03:53.350364 1258949 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/config.json: {Name:mk67aa2523d4e4b30094ae04c9054a058cea8232 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 23:03:53.350562 1258949 cache.go:195] Successfully downloaded all kic artifacts
	I0830 23:03:53.350610 1258949 start.go:365] acquiring machines lock for ingress-addon-legacy-235029: {Name:mk792f34440bee14346066532b197e32b10131fd Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
	I0830 23:03:53.350676 1258949 start.go:369] acquired machines lock for "ingress-addon-legacy-235029" in 50.396µs
	I0830 23:03:53.350700 1258949 start.go:93] Provisioning new machine with config: &{Name:ingress-addon-legacy-235029 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:4096 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.18.20 ClusterName:ingress-addon-legacy-235029 Namespace:defau
lt APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.18.20 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker Binar
yMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0} &{Name: IP: Port:8443 KubernetesVersion:v1.18.20 ContainerRuntime:containerd ControlPlane:true Worker:true}
	I0830 23:03:53.350781 1258949 start.go:125] createHost starting for "" (driver="docker")
	I0830 23:03:53.353427 1258949 out.go:204] * Creating docker container (CPUs=2, Memory=4096MB) ...
	I0830 23:03:53.353685 1258949 start.go:159] libmachine.API.Create for "ingress-addon-legacy-235029" (driver="docker")
	I0830 23:03:53.353723 1258949 client.go:168] LocalClient.Create starting
	I0830 23:03:53.353803 1258949 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem
	I0830 23:03:53.353846 1258949 main.go:141] libmachine: Decoding PEM data...
	I0830 23:03:53.353865 1258949 main.go:141] libmachine: Parsing certificate...
	I0830 23:03:53.353923 1258949 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem
	I0830 23:03:53.353945 1258949 main.go:141] libmachine: Decoding PEM data...
	I0830 23:03:53.353959 1258949 main.go:141] libmachine: Parsing certificate...
	I0830 23:03:53.354335 1258949 cli_runner.go:164] Run: docker network inspect ingress-addon-legacy-235029 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	W0830 23:03:53.373696 1258949 cli_runner.go:211] docker network inspect ingress-addon-legacy-235029 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
	I0830 23:03:53.373787 1258949 network_create.go:281] running [docker network inspect ingress-addon-legacy-235029] to gather additional debugging logs...
	I0830 23:03:53.373807 1258949 cli_runner.go:164] Run: docker network inspect ingress-addon-legacy-235029
	W0830 23:03:53.393733 1258949 cli_runner.go:211] docker network inspect ingress-addon-legacy-235029 returned with exit code 1
	I0830 23:03:53.393764 1258949 network_create.go:284] error running [docker network inspect ingress-addon-legacy-235029]: docker network inspect ingress-addon-legacy-235029: exit status 1
	stdout:
	[]
	
	stderr:
	Error response from daemon: network ingress-addon-legacy-235029 not found
	I0830 23:03:53.393780 1258949 network_create.go:286] output of [docker network inspect ingress-addon-legacy-235029]: -- stdout --
	[]
	
	-- /stdout --
	** stderr ** 
	Error response from daemon: network ingress-addon-legacy-235029 not found
	
	** /stderr **
	I0830 23:03:53.393845 1258949 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	I0830 23:03:53.411517 1258949 network.go:209] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x4000133f40}
	I0830 23:03:53.411559 1258949 network_create.go:123] attempt to create docker network ingress-addon-legacy-235029 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
	I0830 23:03:53.411620 1258949 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=ingress-addon-legacy-235029 ingress-addon-legacy-235029
	I0830 23:03:53.486662 1258949 network_create.go:107] docker network ingress-addon-legacy-235029 192.168.49.0/24 created
	I0830 23:03:53.486735 1258949 kic.go:117] calculated static IP "192.168.49.2" for the "ingress-addon-legacy-235029" container
	I0830 23:03:53.486827 1258949 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
	I0830 23:03:53.507089 1258949 cli_runner.go:164] Run: docker volume create ingress-addon-legacy-235029 --label name.minikube.sigs.k8s.io=ingress-addon-legacy-235029 --label created_by.minikube.sigs.k8s.io=true
	I0830 23:03:53.525979 1258949 oci.go:103] Successfully created a docker volume ingress-addon-legacy-235029
	I0830 23:03:53.526068 1258949 cli_runner.go:164] Run: docker run --rm --name ingress-addon-legacy-235029-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=ingress-addon-legacy-235029 --entrypoint /usr/bin/test -v ingress-addon-legacy-235029:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec -d /var/lib
	I0830 23:03:55.030676 1258949 cli_runner.go:217] Completed: docker run --rm --name ingress-addon-legacy-235029-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=ingress-addon-legacy-235029 --entrypoint /usr/bin/test -v ingress-addon-legacy-235029:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec -d /var/lib: (1.504557717s)
	I0830 23:03:55.030713 1258949 oci.go:107] Successfully prepared a docker volume ingress-addon-legacy-235029
	I0830 23:03:55.030743 1258949 preload.go:132] Checking if preload exists for k8s version v1.18.20 and runtime containerd
	I0830 23:03:55.030766 1258949 kic.go:190] Starting extracting preloaded images to volume ...
	I0830 23:03:55.030870 1258949 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.18.20-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v ingress-addon-legacy-235029:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec -I lz4 -xf /preloaded.tar -C /extractDir
	I0830 23:03:59.963775 1258949 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.18.20-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v ingress-addon-legacy-235029:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec -I lz4 -xf /preloaded.tar -C /extractDir: (4.93285783s)
	I0830 23:03:59.963810 1258949 kic.go:199] duration metric: took 4.933043 seconds to extract preloaded images to volume
	W0830 23:03:59.963981 1258949 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
	I0830 23:03:59.964089 1258949 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
	I0830 23:04:00.071062 1258949 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname ingress-addon-legacy-235029 --name ingress-addon-legacy-235029 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=ingress-addon-legacy-235029 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=ingress-addon-legacy-235029 --network ingress-addon-legacy-235029 --ip 192.168.49.2 --volume ingress-addon-legacy-235029:/var --security-opt apparmor=unconfined --memory=4096mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec
	I0830 23:04:00.708696 1258949 cli_runner.go:164] Run: docker container inspect ingress-addon-legacy-235029 --format={{.State.Running}}
	I0830 23:04:00.740909 1258949 cli_runner.go:164] Run: docker container inspect ingress-addon-legacy-235029 --format={{.State.Status}}
	I0830 23:04:00.777328 1258949 cli_runner.go:164] Run: docker exec ingress-addon-legacy-235029 stat /var/lib/dpkg/alternatives/iptables
	I0830 23:04:00.882982 1258949 oci.go:144] the created container "ingress-addon-legacy-235029" has a running status.
	I0830 23:04:00.883014 1258949 kic.go:221] Creating ssh key for kic: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/ingress-addon-legacy-235029/id_rsa...
	I0830 23:04:02.276589 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/ingress-addon-legacy-235029/id_rsa.pub -> /home/docker/.ssh/authorized_keys
	I0830 23:04:02.276644 1258949 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/ingress-addon-legacy-235029/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
	I0830 23:04:02.302511 1258949 cli_runner.go:164] Run: docker container inspect ingress-addon-legacy-235029 --format={{.State.Status}}
	I0830 23:04:02.321175 1258949 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
	I0830 23:04:02.321200 1258949 kic_runner.go:114] Args: [docker exec --privileged ingress-addon-legacy-235029 chown docker:docker /home/docker/.ssh/authorized_keys]
	I0830 23:04:02.392040 1258949 cli_runner.go:164] Run: docker container inspect ingress-addon-legacy-235029 --format={{.State.Status}}
	I0830 23:04:02.411420 1258949 machine.go:88] provisioning docker machine ...
	I0830 23:04:02.411456 1258949 ubuntu.go:169] provisioning hostname "ingress-addon-legacy-235029"
	I0830 23:04:02.411528 1258949 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" ingress-addon-legacy-235029
	I0830 23:04:02.429924 1258949 main.go:141] libmachine: Using SSH client type: native
	I0830 23:04:02.430408 1258949 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34334 <nil> <nil>}
	I0830 23:04:02.430427 1258949 main.go:141] libmachine: About to run SSH command:
	sudo hostname ingress-addon-legacy-235029 && echo "ingress-addon-legacy-235029" | sudo tee /etc/hostname
	I0830 23:04:02.598454 1258949 main.go:141] libmachine: SSH cmd err, output: <nil>: ingress-addon-legacy-235029
	
	I0830 23:04:02.598540 1258949 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" ingress-addon-legacy-235029
	I0830 23:04:02.619581 1258949 main.go:141] libmachine: Using SSH client type: native
	I0830 23:04:02.620064 1258949 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34334 <nil> <nil>}
	I0830 23:04:02.620092 1258949 main.go:141] libmachine: About to run SSH command:
	
			if ! grep -xq '.*\singress-addon-legacy-235029' /etc/hosts; then
				if grep -xq '127.0.1.1\s.*' /etc/hosts; then
					sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 ingress-addon-legacy-235029/g' /etc/hosts;
				else 
					echo '127.0.1.1 ingress-addon-legacy-235029' | sudo tee -a /etc/hosts; 
				fi
			fi
	I0830 23:04:02.767400 1258949 main.go:141] libmachine: SSH cmd err, output: <nil>: 
	I0830 23:04:02.767427 1258949 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/17114-1219981/.minikube CaCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/17114-1219981/.minikube}
	I0830 23:04:02.767448 1258949 ubuntu.go:177] setting up certificates
	I0830 23:04:02.767458 1258949 provision.go:83] configureAuth start
	I0830 23:04:02.767521 1258949 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" ingress-addon-legacy-235029
	I0830 23:04:02.788376 1258949 provision.go:138] copyHostCerts
	I0830 23:04:02.788426 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem -> /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem
	I0830 23:04:02.788459 1258949 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem, removing ...
	I0830 23:04:02.788469 1258949 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem
	I0830 23:04:02.788570 1258949 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem (1082 bytes)
	I0830 23:04:02.788700 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem -> /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem
	I0830 23:04:02.788719 1258949 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem, removing ...
	I0830 23:04:02.788723 1258949 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem
	I0830 23:04:02.788750 1258949 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem (1123 bytes)
	I0830 23:04:02.788806 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem -> /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem
	I0830 23:04:02.788824 1258949 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem, removing ...
	I0830 23:04:02.788828 1258949 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem
	I0830 23:04:02.788870 1258949 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem (1679 bytes)
	I0830 23:04:02.788929 1258949 provision.go:112] generating server cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem org=jenkins.ingress-addon-legacy-235029 san=[192.168.49.2 127.0.0.1 localhost 127.0.0.1 minikube ingress-addon-legacy-235029]
	I0830 23:04:03.296250 1258949 provision.go:172] copyRemoteCerts
	I0830 23:04:03.296328 1258949 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
	I0830 23:04:03.296382 1258949 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" ingress-addon-legacy-235029
	I0830 23:04:03.314981 1258949 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34334 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/ingress-addon-legacy-235029/id_rsa Username:docker}
	I0830 23:04:03.420903 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem -> /etc/docker/ca.pem
	I0830 23:04:03.420985 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
	I0830 23:04:03.454342 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem -> /etc/docker/server.pem
	I0830 23:04:03.454403 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem --> /etc/docker/server.pem (1257 bytes)
	I0830 23:04:03.484806 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem -> /etc/docker/server-key.pem
	I0830 23:04:03.484872 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
	I0830 23:04:03.514716 1258949 provision.go:86] duration metric: configureAuth took 747.242054ms
	I0830 23:04:03.514782 1258949 ubuntu.go:193] setting minikube options for container-runtime
	I0830 23:04:03.515022 1258949 config.go:182] Loaded profile config "ingress-addon-legacy-235029": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.18.20
	I0830 23:04:03.515037 1258949 machine.go:91] provisioned docker machine in 1.103594171s
	I0830 23:04:03.515044 1258949 client.go:171] LocalClient.Create took 10.161314375s
	I0830 23:04:03.515059 1258949 start.go:167] duration metric: libmachine.API.Create for "ingress-addon-legacy-235029" took 10.161374804s
	I0830 23:04:03.515066 1258949 start.go:300] post-start starting for "ingress-addon-legacy-235029" (driver="docker")
	I0830 23:04:03.515080 1258949 start.go:329] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
	I0830 23:04:03.515147 1258949 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
	I0830 23:04:03.515202 1258949 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" ingress-addon-legacy-235029
	I0830 23:04:03.534171 1258949 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34334 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/ingress-addon-legacy-235029/id_rsa Username:docker}
	I0830 23:04:03.640621 1258949 ssh_runner.go:195] Run: cat /etc/os-release
	I0830 23:04:03.644942 1258949 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
	I0830 23:04:03.644983 1258949 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
	I0830 23:04:03.644997 1258949 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
	I0830 23:04:03.645004 1258949 info.go:137] Remote host: Ubuntu 22.04.3 LTS
	I0830 23:04:03.645018 1258949 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/addons for local assets ...
	I0830 23:04:03.645082 1258949 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/files for local assets ...
	I0830 23:04:03.645216 1258949 filesync.go:149] local asset: /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem -> 12253752.pem in /etc/ssl/certs
	I0830 23:04:03.645231 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem -> /etc/ssl/certs/12253752.pem
	I0830 23:04:03.645346 1258949 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
	I0830 23:04:03.656591 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem --> /etc/ssl/certs/12253752.pem (1708 bytes)
	I0830 23:04:03.686592 1258949 start.go:303] post-start completed in 171.505509ms
	I0830 23:04:03.687049 1258949 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" ingress-addon-legacy-235029
	I0830 23:04:03.705753 1258949 profile.go:148] Saving config to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/config.json ...
	I0830 23:04:03.706031 1258949 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0830 23:04:03.706090 1258949 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" ingress-addon-legacy-235029
	I0830 23:04:03.724234 1258949 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34334 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/ingress-addon-legacy-235029/id_rsa Username:docker}
	I0830 23:04:03.823796 1258949 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
	I0830 23:04:03.830012 1258949 start.go:128] duration metric: createHost completed in 10.479214368s
	I0830 23:04:03.830033 1258949 start.go:83] releasing machines lock for "ingress-addon-legacy-235029", held for 10.479346499s
	I0830 23:04:03.830106 1258949 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" ingress-addon-legacy-235029
	I0830 23:04:03.848627 1258949 ssh_runner.go:195] Run: cat /version.json
	I0830 23:04:03.848687 1258949 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" ingress-addon-legacy-235029
	I0830 23:04:03.848947 1258949 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
	I0830 23:04:03.849031 1258949 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" ingress-addon-legacy-235029
	I0830 23:04:03.872324 1258949 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34334 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/ingress-addon-legacy-235029/id_rsa Username:docker}
	I0830 23:04:03.887302 1258949 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34334 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/ingress-addon-legacy-235029/id_rsa Username:docker}
	I0830 23:04:03.977646 1258949 ssh_runner.go:195] Run: systemctl --version
	I0830 23:04:04.115619 1258949 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
	I0830 23:04:04.121386 1258949 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
	I0830 23:04:04.151902 1258949 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
	I0830 23:04:04.151980 1258949 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%!p(MISSING), " -exec sh -c "sudo mv {} {}.mk_disabled" ;
	I0830 23:04:04.186461 1258949 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
	I0830 23:04:04.186485 1258949 start.go:466] detecting cgroup driver to use...
	I0830 23:04:04.186518 1258949 detect.go:196] detected "cgroupfs" cgroup driver on host os
	I0830 23:04:04.186571 1258949 ssh_runner.go:195] Run: sudo systemctl stop -f crio
	I0830 23:04:04.201397 1258949 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
	I0830 23:04:04.214995 1258949 docker.go:196] disabling cri-docker service (if available) ...
	I0830 23:04:04.215092 1258949 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
	I0830 23:04:04.231420 1258949 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
	I0830 23:04:04.248416 1258949 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
	I0830 23:04:04.352167 1258949 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
	I0830 23:04:04.462520 1258949 docker.go:212] disabling docker service ...
	I0830 23:04:04.462611 1258949 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
	I0830 23:04:04.484512 1258949 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
	I0830 23:04:04.499235 1258949 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
	I0830 23:04:04.600799 1258949 ssh_runner.go:195] Run: sudo systemctl mask docker.service
	I0830 23:04:04.705228 1258949 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
	I0830 23:04:04.719035 1258949 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///run/containerd/containerd.sock
	" | sudo tee /etc/crictl.yaml"
	I0830 23:04:04.740674 1258949 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.2"|' /etc/containerd/config.toml"
	I0830 23:04:04.753563 1258949 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
	I0830 23:04:04.766560 1258949 containerd.go:145] configuring containerd to use "cgroupfs" as cgroup driver...
	I0830 23:04:04.766662 1258949 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
	I0830 23:04:04.779981 1258949 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
	I0830 23:04:04.793017 1258949 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
	I0830 23:04:04.806433 1258949 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
	I0830 23:04:04.819586 1258949 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
	I0830 23:04:04.831548 1258949 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
	I0830 23:04:04.844360 1258949 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
	I0830 23:04:04.854619 1258949 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
	I0830 23:04:04.865376 1258949 ssh_runner.go:195] Run: sudo systemctl daemon-reload
	I0830 23:04:04.966389 1258949 ssh_runner.go:195] Run: sudo systemctl restart containerd
	I0830 23:04:05.116874 1258949 start.go:513] Will wait 60s for socket path /run/containerd/containerd.sock
	I0830 23:04:05.116993 1258949 ssh_runner.go:195] Run: stat /run/containerd/containerd.sock
	I0830 23:04:05.122187 1258949 start.go:534] Will wait 60s for crictl version
	I0830 23:04:05.122302 1258949 ssh_runner.go:195] Run: which crictl
	I0830 23:04:05.127235 1258949 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
	I0830 23:04:05.173482 1258949 start.go:550] Version:  0.1.0
	RuntimeName:  containerd
	RuntimeVersion:  1.6.22
	RuntimeApiVersion:  v1
	I0830 23:04:05.173554 1258949 ssh_runner.go:195] Run: containerd --version
	I0830 23:04:05.201893 1258949 ssh_runner.go:195] Run: containerd --version
	I0830 23:04:05.237743 1258949 out.go:177] * Preparing Kubernetes v1.18.20 on containerd 1.6.22 ...
	I0830 23:04:05.239768 1258949 cli_runner.go:164] Run: docker network inspect ingress-addon-legacy-235029 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	I0830 23:04:05.257162 1258949 ssh_runner.go:195] Run: grep 192.168.49.1	host.minikube.internal$ /etc/hosts
	I0830 23:04:05.261797 1258949 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1	host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
	I0830 23:04:05.275634 1258949 preload.go:132] Checking if preload exists for k8s version v1.18.20 and runtime containerd
	I0830 23:04:05.275718 1258949 ssh_runner.go:195] Run: sudo crictl images --output json
	I0830 23:04:05.314773 1258949 containerd.go:600] couldn't find preloaded image for "registry.k8s.io/kube-apiserver:v1.18.20". assuming images are not preloaded.
	I0830 23:04:05.314847 1258949 ssh_runner.go:195] Run: which lz4
	I0830 23:04:05.319352 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.18.20-containerd-overlay2-arm64.tar.lz4 -> /preloaded.tar.lz4
	I0830 23:04:05.319461 1258949 ssh_runner.go:195] Run: stat -c "%!s(MISSING) %!y(MISSING)" /preloaded.tar.lz4
	I0830 23:04:05.324177 1258949 ssh_runner.go:352] existence check for /preloaded.tar.lz4: stat -c "%!s(MISSING) %!y(MISSING)" /preloaded.tar.lz4: Process exited with status 1
	stdout:
	
	stderr:
	stat: cannot statx '/preloaded.tar.lz4': No such file or directory
	I0830 23:04:05.324219 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.18.20-containerd-overlay2-arm64.tar.lz4 --> /preloaded.tar.lz4 (489149349 bytes)
	I0830 23:04:07.332167 1258949 containerd.go:547] Took 2.012725 seconds to copy over tarball
	I0830 23:04:07.332322 1258949 ssh_runner.go:195] Run: sudo tar -I lz4 -C /var -xf /preloaded.tar.lz4
	I0830 23:04:10.089757 1258949 ssh_runner.go:235] Completed: sudo tar -I lz4 -C /var -xf /preloaded.tar.lz4: (2.757397491s)
	I0830 23:04:10.089786 1258949 containerd.go:554] Took 2.757574 seconds to extract the tarball
	I0830 23:04:10.089797 1258949 ssh_runner.go:146] rm: /preloaded.tar.lz4
	I0830 23:04:10.175934 1258949 ssh_runner.go:195] Run: sudo systemctl daemon-reload
	I0830 23:04:10.280788 1258949 ssh_runner.go:195] Run: sudo systemctl restart containerd
	I0830 23:04:10.400226 1258949 ssh_runner.go:195] Run: sudo crictl images --output json
	I0830 23:04:10.446007 1258949 containerd.go:600] couldn't find preloaded image for "registry.k8s.io/kube-apiserver:v1.18.20". assuming images are not preloaded.
	I0830 23:04:10.446039 1258949 cache_images.go:88] LoadImages start: [registry.k8s.io/kube-apiserver:v1.18.20 registry.k8s.io/kube-controller-manager:v1.18.20 registry.k8s.io/kube-scheduler:v1.18.20 registry.k8s.io/kube-proxy:v1.18.20 registry.k8s.io/pause:3.2 registry.k8s.io/etcd:3.4.3-0 registry.k8s.io/coredns:1.6.7 gcr.io/k8s-minikube/storage-provisioner:v5]
	I0830 23:04:10.446092 1258949 image.go:134] retrieving image: gcr.io/k8s-minikube/storage-provisioner:v5
	I0830 23:04:10.446292 1258949 image.go:134] retrieving image: registry.k8s.io/kube-apiserver:v1.18.20
	I0830 23:04:10.446370 1258949 image.go:134] retrieving image: registry.k8s.io/kube-controller-manager:v1.18.20
	I0830 23:04:10.446445 1258949 image.go:134] retrieving image: registry.k8s.io/kube-scheduler:v1.18.20
	I0830 23:04:10.446534 1258949 image.go:134] retrieving image: registry.k8s.io/kube-proxy:v1.18.20
	I0830 23:04:10.446640 1258949 image.go:134] retrieving image: registry.k8s.io/pause:3.2
	I0830 23:04:10.446728 1258949 image.go:134] retrieving image: registry.k8s.io/etcd:3.4.3-0
	I0830 23:04:10.446799 1258949 image.go:134] retrieving image: registry.k8s.io/coredns:1.6.7
	I0830 23:04:10.447557 1258949 image.go:177] daemon lookup for registry.k8s.io/kube-proxy:v1.18.20: Error response from daemon: No such image: registry.k8s.io/kube-proxy:v1.18.20
	I0830 23:04:10.448138 1258949 image.go:177] daemon lookup for registry.k8s.io/etcd:3.4.3-0: Error response from daemon: No such image: registry.k8s.io/etcd:3.4.3-0
	I0830 23:04:10.448317 1258949 image.go:177] daemon lookup for registry.k8s.io/pause:3.2: Error response from daemon: No such image: registry.k8s.io/pause:3.2
	I0830 23:04:10.448466 1258949 image.go:177] daemon lookup for gcr.io/k8s-minikube/storage-provisioner:v5: Error response from daemon: No such image: gcr.io/k8s-minikube/storage-provisioner:v5
	I0830 23:04:10.448765 1258949 image.go:177] daemon lookup for registry.k8s.io/kube-apiserver:v1.18.20: Error response from daemon: No such image: registry.k8s.io/kube-apiserver:v1.18.20
	I0830 23:04:10.448923 1258949 image.go:177] daemon lookup for registry.k8s.io/kube-scheduler:v1.18.20: Error response from daemon: No such image: registry.k8s.io/kube-scheduler:v1.18.20
	I0830 23:04:10.449062 1258949 image.go:177] daemon lookup for registry.k8s.io/kube-controller-manager:v1.18.20: Error response from daemon: No such image: registry.k8s.io/kube-controller-manager:v1.18.20
	I0830 23:04:10.449883 1258949 image.go:177] daemon lookup for registry.k8s.io/coredns:1.6.7: Error response from daemon: No such image: registry.k8s.io/coredns:1.6.7
	W0830 23:04:10.819360 1258949 image.go:265] image registry.k8s.io/coredns:1.6.7 arch mismatch: want arm64 got amd64. fixing
	I0830 23:04:10.819504 1258949 ssh_runner.go:195] Run: /bin/bash -c "sudo ctr -n=k8s.io images check | grep registry.k8s.io/coredns:1.6.7"
	W0830 23:04:10.876944 1258949 image.go:265] image registry.k8s.io/kube-scheduler:v1.18.20 arch mismatch: want arm64 got amd64. fixing
	I0830 23:04:10.877071 1258949 ssh_runner.go:195] Run: /bin/bash -c "sudo ctr -n=k8s.io images check | grep registry.k8s.io/kube-scheduler:v1.18.20"
	I0830 23:04:10.887237 1258949 ssh_runner.go:195] Run: /bin/bash -c "sudo ctr -n=k8s.io images check | grep registry.k8s.io/pause:3.2"
	W0830 23:04:10.895579 1258949 image.go:265] image registry.k8s.io/kube-proxy:v1.18.20 arch mismatch: want arm64 got amd64. fixing
	I0830 23:04:10.895806 1258949 ssh_runner.go:195] Run: /bin/bash -c "sudo ctr -n=k8s.io images check | grep registry.k8s.io/kube-proxy:v1.18.20"
	W0830 23:04:10.913994 1258949 image.go:265] image registry.k8s.io/etcd:3.4.3-0 arch mismatch: want arm64 got amd64. fixing
	I0830 23:04:10.914243 1258949 ssh_runner.go:195] Run: /bin/bash -c "sudo ctr -n=k8s.io images check | grep registry.k8s.io/etcd:3.4.3-0"
	W0830 23:04:10.942052 1258949 image.go:265] image registry.k8s.io/kube-apiserver:v1.18.20 arch mismatch: want arm64 got amd64. fixing
	I0830 23:04:10.942180 1258949 ssh_runner.go:195] Run: /bin/bash -c "sudo ctr -n=k8s.io images check | grep registry.k8s.io/kube-apiserver:v1.18.20"
	W0830 23:04:10.950130 1258949 image.go:265] image registry.k8s.io/kube-controller-manager:v1.18.20 arch mismatch: want arm64 got amd64. fixing
	I0830 23:04:10.950257 1258949 ssh_runner.go:195] Run: /bin/bash -c "sudo ctr -n=k8s.io images check | grep registry.k8s.io/kube-controller-manager:v1.18.20"
	W0830 23:04:11.086558 1258949 image.go:265] image gcr.io/k8s-minikube/storage-provisioner:v5 arch mismatch: want arm64 got amd64. fixing
	I0830 23:04:11.086743 1258949 ssh_runner.go:195] Run: /bin/bash -c "sudo ctr -n=k8s.io images check | grep gcr.io/k8s-minikube/storage-provisioner:v5"
	I0830 23:04:11.200904 1258949 cache_images.go:116] "registry.k8s.io/kube-scheduler:v1.18.20" needs transfer: "registry.k8s.io/kube-scheduler:v1.18.20" does not exist at hash "177548d745cb87f773d02f41d453af2f2a1479dbe3c32e749cf6d8145c005e79" in container runtime
	I0830 23:04:11.200998 1258949 cri.go:218] Removing image: registry.k8s.io/kube-scheduler:v1.18.20
	I0830 23:04:11.201073 1258949 ssh_runner.go:195] Run: which crictl
	I0830 23:04:11.201436 1258949 cache_images.go:116] "registry.k8s.io/coredns:1.6.7" needs transfer: "registry.k8s.io/coredns:1.6.7" does not exist at hash "ff3af22d8878afc6985d3fec3e066d00ef431aa166c3a01ac58f1990adc92a2c" in container runtime
	I0830 23:04:11.201496 1258949 cri.go:218] Removing image: registry.k8s.io/coredns:1.6.7
	I0830 23:04:11.201559 1258949 ssh_runner.go:195] Run: which crictl
	I0830 23:04:11.583501 1258949 cache_images.go:116] "registry.k8s.io/pause:3.2" needs transfer: "registry.k8s.io/pause:3.2" does not exist at hash "2a060e2e7101d419352bf82c613158587400be743482d9a537ec4a9d1b4eb93c" in container runtime
	I0830 23:04:11.583595 1258949 cri.go:218] Removing image: registry.k8s.io/pause:3.2
	I0830 23:04:11.583672 1258949 ssh_runner.go:195] Run: which crictl
	I0830 23:04:11.583779 1258949 cache_images.go:116] "registry.k8s.io/kube-proxy:v1.18.20" needs transfer: "registry.k8s.io/kube-proxy:v1.18.20" does not exist at hash "b11cdc97ac6ac4ef2b3b0662edbe16597084b17cbc8e3d61fcaf4ef827a7ed18" in container runtime
	I0830 23:04:11.583827 1258949 cri.go:218] Removing image: registry.k8s.io/kube-proxy:v1.18.20
	I0830 23:04:11.583884 1258949 ssh_runner.go:195] Run: which crictl
	I0830 23:04:11.658904 1258949 cache_images.go:116] "registry.k8s.io/etcd:3.4.3-0" needs transfer: "registry.k8s.io/etcd:3.4.3-0" does not exist at hash "29dd247b2572efbe28fcaea3fef1c5d72593da59f7350e3f6d2e6618983f9c03" in container runtime
	I0830 23:04:11.659008 1258949 cri.go:218] Removing image: registry.k8s.io/etcd:3.4.3-0
	I0830 23:04:11.659089 1258949 ssh_runner.go:195] Run: which crictl
	I0830 23:04:11.659203 1258949 cache_images.go:116] "registry.k8s.io/kube-apiserver:v1.18.20" needs transfer: "registry.k8s.io/kube-apiserver:v1.18.20" does not exist at hash "d353007847ec85700463981309a5846c8d9c93fbcd1323104266212926d68257" in container runtime
	I0830 23:04:11.659250 1258949 cri.go:218] Removing image: registry.k8s.io/kube-apiserver:v1.18.20
	I0830 23:04:11.659288 1258949 ssh_runner.go:195] Run: which crictl
	I0830 23:04:11.789229 1258949 cache_images.go:116] "registry.k8s.io/kube-controller-manager:v1.18.20" needs transfer: "registry.k8s.io/kube-controller-manager:v1.18.20" does not exist at hash "297c79afbdb81ceb4cf857e0c54a0de7b6ce7ebe01e6cab68fc8baf342be3ea7" in container runtime
	I0830 23:04:11.789275 1258949 cri.go:218] Removing image: registry.k8s.io/kube-controller-manager:v1.18.20
	I0830 23:04:11.789327 1258949 ssh_runner.go:195] Run: which crictl
	I0830 23:04:11.789426 1258949 cache_images.go:116] "gcr.io/k8s-minikube/storage-provisioner:v5" needs transfer: "gcr.io/k8s-minikube/storage-provisioner:v5" does not exist at hash "66749159455b3f08c8318fe0233122f54d0f5889f9c5fdfb73c3fd9d99895b51" in container runtime
	I0830 23:04:11.789448 1258949 cri.go:218] Removing image: gcr.io/k8s-minikube/storage-provisioner:v5
	I0830 23:04:11.789477 1258949 ssh_runner.go:195] Run: which crictl
	I0830 23:04:11.789544 1258949 ssh_runner.go:195] Run: sudo /usr/bin/crictl rmi registry.k8s.io/kube-scheduler:v1.18.20
	I0830 23:04:11.789604 1258949 ssh_runner.go:195] Run: sudo /usr/bin/crictl rmi registry.k8s.io/coredns:1.6.7
	I0830 23:04:11.789671 1258949 ssh_runner.go:195] Run: sudo /usr/bin/crictl rmi registry.k8s.io/kube-proxy:v1.18.20
	I0830 23:04:11.789725 1258949 ssh_runner.go:195] Run: sudo /usr/bin/crictl rmi registry.k8s.io/pause:3.2
	I0830 23:04:11.789794 1258949 ssh_runner.go:195] Run: sudo /usr/bin/crictl rmi registry.k8s.io/kube-apiserver:v1.18.20
	I0830 23:04:11.789859 1258949 ssh_runner.go:195] Run: sudo /usr/bin/crictl rmi registry.k8s.io/etcd:3.4.3-0
	I0830 23:04:11.955470 1258949 cache_images.go:286] Loading image from: /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/images/arm64/registry.k8s.io/etcd_3.4.3-0
	I0830 23:04:11.955550 1258949 ssh_runner.go:195] Run: sudo /usr/bin/crictl rmi registry.k8s.io/kube-controller-manager:v1.18.20
	I0830 23:04:11.955625 1258949 ssh_runner.go:195] Run: sudo /usr/bin/crictl rmi gcr.io/k8s-minikube/storage-provisioner:v5
	I0830 23:04:11.955685 1258949 cache_images.go:286] Loading image from: /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/images/arm64/registry.k8s.io/kube-scheduler_v1.18.20
	I0830 23:04:11.955722 1258949 cache_images.go:286] Loading image from: /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/images/arm64/registry.k8s.io/coredns_1.6.7
	I0830 23:04:11.955759 1258949 cache_images.go:286] Loading image from: /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/images/arm64/registry.k8s.io/kube-proxy_v1.18.20
	I0830 23:04:11.955803 1258949 cache_images.go:286] Loading image from: /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/images/arm64/registry.k8s.io/pause_3.2
	I0830 23:04:11.955836 1258949 cache_images.go:286] Loading image from: /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/images/arm64/registry.k8s.io/kube-apiserver_v1.18.20
	I0830 23:04:12.019366 1258949 cache_images.go:286] Loading image from: /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/images/arm64/gcr.io/k8s-minikube/storage-provisioner_v5
	I0830 23:04:12.019490 1258949 cache_images.go:286] Loading image from: /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/images/arm64/registry.k8s.io/kube-controller-manager_v1.18.20
	I0830 23:04:12.019521 1258949 cache_images.go:92] LoadImages completed in 1.57346966s
	W0830 23:04:12.019591 1258949 out.go:239] X Unable to load cached images: loading cached images: stat /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/images/arm64/registry.k8s.io/etcd_3.4.3-0: no such file or directory
	I0830 23:04:12.019645 1258949 ssh_runner.go:195] Run: sudo crictl info
	I0830 23:04:12.065905 1258949 cni.go:84] Creating CNI manager for ""
	I0830 23:04:12.065929 1258949 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
	I0830 23:04:12.065963 1258949 kubeadm.go:87] Using pod CIDR: 10.244.0.0/16
	I0830 23:04:12.065992 1258949 kubeadm.go:176] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.18.20 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:ingress-addon-legacy-235029 NodeName:ingress-addon-legacy-235029 DNSDomain:cluster.local CRISocket:/run/containerd/containerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/ce
rts/ca.crt StaticPodPath:/etc/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:false}
	I0830 23:04:12.066140 1258949 kubeadm.go:181] kubeadm config:
	apiVersion: kubeadm.k8s.io/v1beta2
	kind: InitConfiguration
	localAPIEndpoint:
	  advertiseAddress: 192.168.49.2
	  bindPort: 8443
	bootstrapTokens:
	  - groups:
	      - system:bootstrappers:kubeadm:default-node-token
	    ttl: 24h0m0s
	    usages:
	      - signing
	      - authentication
	nodeRegistration:
	  criSocket: /run/containerd/containerd.sock
	  name: "ingress-addon-legacy-235029"
	  kubeletExtraArgs:
	    node-ip: 192.168.49.2
	  taints: []
	---
	apiVersion: kubeadm.k8s.io/v1beta2
	kind: ClusterConfiguration
	apiServer:
	  certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
	  extraArgs:
	    enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
	controllerManager:
	  extraArgs:
	    allocate-node-cidrs: "true"
	    leader-elect: "false"
	scheduler:
	  extraArgs:
	    leader-elect: "false"
	certificatesDir: /var/lib/minikube/certs
	clusterName: mk
	controlPlaneEndpoint: control-plane.minikube.internal:8443
	dns:
	  type: CoreDNS
	etcd:
	  local:
	    dataDir: /var/lib/minikube/etcd
	    extraArgs:
	      proxy-refresh-interval: "70000"
	kubernetesVersion: v1.18.20
	networking:
	  dnsDomain: cluster.local
	  podSubnet: "10.244.0.0/16"
	  serviceSubnet: 10.96.0.0/12
	---
	apiVersion: kubelet.config.k8s.io/v1beta1
	kind: KubeletConfiguration
	authentication:
	  x509:
	    clientCAFile: /var/lib/minikube/certs/ca.crt
	cgroupDriver: cgroupfs
	hairpinMode: hairpin-veth
	runtimeRequestTimeout: 15m
	clusterDomain: "cluster.local"
	# disable disk resource management by default
	imageGCHighThresholdPercent: 100
	evictionHard:
	  nodefs.available: "0%!"(MISSING)
	  nodefs.inodesFree: "0%!"(MISSING)
	  imagefs.available: "0%!"(MISSING)
	failSwapOn: false
	staticPodPath: /etc/kubernetes/manifests
	---
	apiVersion: kubeproxy.config.k8s.io/v1alpha1
	kind: KubeProxyConfiguration
	clusterCIDR: "10.244.0.0/16"
	metricsBindAddress: 0.0.0.0:10249
	conntrack:
	  maxPerCore: 0
	# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
	  tcpEstablishedTimeout: 0s
	# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
	  tcpCloseWaitTimeout: 0s
	
	I0830 23:04:12.066219 1258949 kubeadm.go:976] kubelet [Unit]
	Wants=containerd.service
	
	[Service]
	ExecStart=
	ExecStart=/var/lib/minikube/binaries/v1.18.20/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --container-runtime=remote --container-runtime-endpoint=unix:///run/containerd/containerd.sock --hostname-override=ingress-addon-legacy-235029 --kubeconfig=/etc/kubernetes/kubelet.conf --network-plugin=cni --node-ip=192.168.49.2
	
	[Install]
	 config:
	{KubernetesVersion:v1.18.20 ClusterName:ingress-addon-legacy-235029 Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:}
	I0830 23:04:12.066295 1258949 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.18.20
	I0830 23:04:12.077857 1258949 binaries.go:44] Found k8s binaries, skipping transfer
	I0830 23:04:12.078000 1258949 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
	I0830 23:04:12.090003 1258949 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (448 bytes)
	I0830 23:04:12.113592 1258949 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (353 bytes)
	I0830 23:04:12.136426 1258949 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2131 bytes)
	I0830 23:04:12.158077 1258949 ssh_runner.go:195] Run: grep 192.168.49.2	control-plane.minikube.internal$ /etc/hosts
	I0830 23:04:12.162524 1258949 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2	control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
	I0830 23:04:12.175841 1258949 certs.go:56] Setting up /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029 for IP: 192.168.49.2
	I0830 23:04:12.175922 1258949 certs.go:190] acquiring lock for shared ca certs: {Name:mk74152f2fceba9f1b7e69c343f47396a1a7db69 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 23:04:12.176077 1258949 certs.go:199] skipping minikubeCA CA generation: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.key
	I0830 23:04:12.176130 1258949 certs.go:199] skipping proxyClientCA CA generation: /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.key
	I0830 23:04:12.176181 1258949 certs.go:319] generating minikube-user signed cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.key
	I0830 23:04:12.176196 1258949 crypto.go:68] Generating cert /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt with IP's: []
	I0830 23:04:12.713192 1258949 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt ...
	I0830 23:04:12.713266 1258949 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: {Name:mka5687b8ea4dfe64652efe87a32477c30d18125 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 23:04:12.713477 1258949 crypto.go:164] Writing key to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.key ...
	I0830 23:04:12.713491 1258949 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.key: {Name:mk7e747bee050385c594810f963395b4c63897f7 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 23:04:12.713577 1258949 certs.go:319] generating minikube signed cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.key.dd3b5fb2
	I0830 23:04:12.713594 1258949 crypto.go:68] Generating cert /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.crt.dd3b5fb2 with IP's: [192.168.49.2 10.96.0.1 127.0.0.1 10.0.0.1]
	I0830 23:04:13.490337 1258949 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.crt.dd3b5fb2 ...
	I0830 23:04:13.490371 1258949 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.crt.dd3b5fb2: {Name:mke2fa59f5db55c8734774e14eab198a6d58d582 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 23:04:13.490565 1258949 crypto.go:164] Writing key to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.key.dd3b5fb2 ...
	I0830 23:04:13.490580 1258949 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.key.dd3b5fb2: {Name:mkaf25530c7f2f574ab40c31d789b2d328291f14 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 23:04:13.490663 1258949 certs.go:337] copying /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.crt.dd3b5fb2 -> /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.crt
	I0830 23:04:13.490741 1258949 certs.go:341] copying /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.key.dd3b5fb2 -> /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.key
	I0830 23:04:13.490794 1258949 certs.go:319] generating aggregator signed cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/proxy-client.key
	I0830 23:04:13.490813 1258949 crypto.go:68] Generating cert /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/proxy-client.crt with IP's: []
	I0830 23:04:13.802977 1258949 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/proxy-client.crt ...
	I0830 23:04:13.803010 1258949 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/proxy-client.crt: {Name:mk395c08ae3d3b51e8bcd3cc69c845ce839b9def Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 23:04:13.803195 1258949 crypto.go:164] Writing key to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/proxy-client.key ...
	I0830 23:04:13.803209 1258949 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/proxy-client.key: {Name:mkde16df416cdb5e99421ac1690cfc65150bc66f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 23:04:13.803299 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.crt -> /var/lib/minikube/certs/apiserver.crt
	I0830 23:04:13.803316 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.key -> /var/lib/minikube/certs/apiserver.key
	I0830 23:04:13.803330 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/proxy-client.crt -> /var/lib/minikube/certs/proxy-client.crt
	I0830 23:04:13.803343 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/proxy-client.key -> /var/lib/minikube/certs/proxy-client.key
	I0830 23:04:13.803357 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt -> /var/lib/minikube/certs/ca.crt
	I0830 23:04:13.803378 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.key -> /var/lib/minikube/certs/ca.key
	I0830 23:04:13.803395 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.crt -> /var/lib/minikube/certs/proxy-client-ca.crt
	I0830 23:04:13.803408 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.key -> /var/lib/minikube/certs/proxy-client-ca.key
	I0830 23:04:13.803465 1258949 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/1225375.pem (1338 bytes)
	W0830 23:04:13.803506 1258949 certs.go:433] ignoring /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/1225375_empty.pem, impossibly tiny 0 bytes
	I0830 23:04:13.803519 1258949 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem (1679 bytes)
	I0830 23:04:13.803547 1258949 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem (1082 bytes)
	I0830 23:04:13.803574 1258949 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem (1123 bytes)
	I0830 23:04:13.803601 1258949 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem (1679 bytes)
	I0830 23:04:13.803650 1258949 certs.go:437] found cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem (1708 bytes)
	I0830 23:04:13.803678 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem -> /usr/share/ca-certificates/12253752.pem
	I0830 23:04:13.803695 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt -> /usr/share/ca-certificates/minikubeCA.pem
	I0830 23:04:13.803706 1258949 vm_assets.go:163] NewFileAsset: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/1225375.pem -> /usr/share/ca-certificates/1225375.pem
	I0830 23:04:13.804275 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1399 bytes)
	I0830 23:04:13.834657 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
	I0830 23:04:13.863786 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
	I0830 23:04:13.892490 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
	I0830 23:04:13.920991 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
	I0830 23:04:13.949675 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
	I0830 23:04:13.979529 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
	I0830 23:04:14.008996 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
	I0830 23:04:14.042064 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem --> /usr/share/ca-certificates/12253752.pem (1708 bytes)
	I0830 23:04:14.072206 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
	I0830 23:04:14.101604 1258949 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/1225375.pem --> /usr/share/ca-certificates/1225375.pem (1338 bytes)
	I0830 23:04:14.130788 1258949 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
	I0830 23:04:14.152852 1258949 ssh_runner.go:195] Run: openssl version
	I0830 23:04:14.160003 1258949 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/12253752.pem && ln -fs /usr/share/ca-certificates/12253752.pem /etc/ssl/certs/12253752.pem"
	I0830 23:04:14.172064 1258949 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/12253752.pem
	I0830 23:04:14.176855 1258949 certs.go:480] hashing: -rw-r--r-- 1 root root 1708 Aug 30 23:00 /usr/share/ca-certificates/12253752.pem
	I0830 23:04:14.176926 1258949 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/12253752.pem
	I0830 23:04:14.185699 1258949 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/12253752.pem /etc/ssl/certs/3ec20f2e.0"
	I0830 23:04:14.197862 1258949 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
	I0830 23:04:14.209592 1258949 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
	I0830 23:04:14.214065 1258949 certs.go:480] hashing: -rw-r--r-- 1 root root 1111 Aug 30 22:55 /usr/share/ca-certificates/minikubeCA.pem
	I0830 23:04:14.214126 1258949 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
	I0830 23:04:14.223002 1258949 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
	I0830 23:04:14.235040 1258949 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/1225375.pem && ln -fs /usr/share/ca-certificates/1225375.pem /etc/ssl/certs/1225375.pem"
	I0830 23:04:14.247055 1258949 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/1225375.pem
	I0830 23:04:14.251820 1258949 certs.go:480] hashing: -rw-r--r-- 1 root root 1338 Aug 30 23:00 /usr/share/ca-certificates/1225375.pem
	I0830 23:04:14.251885 1258949 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/1225375.pem
	I0830 23:04:14.260863 1258949 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/1225375.pem /etc/ssl/certs/51391683.0"
	I0830 23:04:14.273205 1258949 ssh_runner.go:195] Run: ls /var/lib/minikube/certs/etcd
	I0830 23:04:14.277852 1258949 certs.go:353] certs directory doesn't exist, likely first start: ls /var/lib/minikube/certs/etcd: Process exited with status 2
	stdout:
	
	stderr:
	ls: cannot access '/var/lib/minikube/certs/etcd': No such file or directory
	I0830 23:04:14.277951 1258949 kubeadm.go:404] StartCluster: {Name:ingress-addon-legacy-235029 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:4096 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.18.20 ClusterName:ingress-addon-legacy-235029 Namespace:default APIServerName:minik
ubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.18.20 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: D
isableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0}
	I0830 23:04:14.278046 1258949 cri.go:54] listing CRI containers in root /run/containerd/runc/k8s.io: {State:paused Name: Namespaces:[kube-system]}
	I0830 23:04:14.278105 1258949 ssh_runner.go:195] Run: sudo -s eval "crictl ps -a --quiet --label io.kubernetes.pod.namespace=kube-system"
	I0830 23:04:14.320496 1258949 cri.go:89] found id: ""
	I0830 23:04:14.320601 1258949 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
	I0830 23:04:14.332589 1258949 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
	I0830 23:04:14.343649 1258949 kubeadm.go:226] ignoring SystemVerification for kubeadm because of docker driver
	I0830 23:04:14.343741 1258949 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
	I0830 23:04:14.356780 1258949 kubeadm.go:152] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
	stdout:
	
	stderr:
	ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
	ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
	ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
	ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
	I0830 23:04:14.356867 1258949 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.18.20:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml  --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
	I0830 23:04:14.411637 1258949 kubeadm.go:322] [init] Using Kubernetes version: v1.18.20
	I0830 23:04:14.414645 1258949 kubeadm.go:322] [preflight] Running pre-flight checks
	I0830 23:04:14.463316 1258949 kubeadm.go:322] [preflight] The system verification failed. Printing the output from the verification:
	I0830 23:04:14.463470 1258949 kubeadm.go:322] KERNEL_VERSION: 5.15.0-1043-aws
	I0830 23:04:14.463536 1258949 kubeadm.go:322] OS: Linux
	I0830 23:04:14.463610 1258949 kubeadm.go:322] CGROUPS_CPU: enabled
	I0830 23:04:14.463686 1258949 kubeadm.go:322] CGROUPS_CPUACCT: enabled
	I0830 23:04:14.463759 1258949 kubeadm.go:322] CGROUPS_CPUSET: enabled
	I0830 23:04:14.463838 1258949 kubeadm.go:322] CGROUPS_DEVICES: enabled
	I0830 23:04:14.463917 1258949 kubeadm.go:322] CGROUPS_FREEZER: enabled
	I0830 23:04:14.463999 1258949 kubeadm.go:322] CGROUPS_MEMORY: enabled
	I0830 23:04:14.559427 1258949 kubeadm.go:322] [preflight] Pulling images required for setting up a Kubernetes cluster
	I0830 23:04:14.559601 1258949 kubeadm.go:322] [preflight] This might take a minute or two, depending on the speed of your internet connection
	I0830 23:04:14.559729 1258949 kubeadm.go:322] [preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
	I0830 23:04:14.799305 1258949 kubeadm.go:322] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
	I0830 23:04:14.800899 1258949 kubeadm.go:322] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
	I0830 23:04:14.801045 1258949 kubeadm.go:322] [kubelet-start] Starting the kubelet
	I0830 23:04:14.902943 1258949 kubeadm.go:322] [certs] Using certificateDir folder "/var/lib/minikube/certs"
	I0830 23:04:14.906126 1258949 out.go:204]   - Generating certificates and keys ...
	I0830 23:04:14.906360 1258949 kubeadm.go:322] [certs] Using existing ca certificate authority
	I0830 23:04:14.906444 1258949 kubeadm.go:322] [certs] Using existing apiserver certificate and key on disk
	I0830 23:04:15.452893 1258949 kubeadm.go:322] [certs] Generating "apiserver-kubelet-client" certificate and key
	I0830 23:04:15.803277 1258949 kubeadm.go:322] [certs] Generating "front-proxy-ca" certificate and key
	I0830 23:04:15.974353 1258949 kubeadm.go:322] [certs] Generating "front-proxy-client" certificate and key
	I0830 23:04:16.805364 1258949 kubeadm.go:322] [certs] Generating "etcd/ca" certificate and key
	I0830 23:04:17.005216 1258949 kubeadm.go:322] [certs] Generating "etcd/server" certificate and key
	I0830 23:04:17.005506 1258949 kubeadm.go:322] [certs] etcd/server serving cert is signed for DNS names [ingress-addon-legacy-235029 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
	I0830 23:04:17.747530 1258949 kubeadm.go:322] [certs] Generating "etcd/peer" certificate and key
	I0830 23:04:17.747740 1258949 kubeadm.go:322] [certs] etcd/peer serving cert is signed for DNS names [ingress-addon-legacy-235029 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
	I0830 23:04:17.982015 1258949 kubeadm.go:322] [certs] Generating "etcd/healthcheck-client" certificate and key
	I0830 23:04:18.442098 1258949 kubeadm.go:322] [certs] Generating "apiserver-etcd-client" certificate and key
	I0830 23:04:18.775064 1258949 kubeadm.go:322] [certs] Generating "sa" key and public key
	I0830 23:04:18.775378 1258949 kubeadm.go:322] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
	I0830 23:04:19.372963 1258949 kubeadm.go:322] [kubeconfig] Writing "admin.conf" kubeconfig file
	I0830 23:04:20.304186 1258949 kubeadm.go:322] [kubeconfig] Writing "kubelet.conf" kubeconfig file
	I0830 23:04:20.589781 1258949 kubeadm.go:322] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
	I0830 23:04:20.919438 1258949 kubeadm.go:322] [kubeconfig] Writing "scheduler.conf" kubeconfig file
	I0830 23:04:20.920148 1258949 kubeadm.go:322] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
	I0830 23:04:20.922410 1258949 out.go:204]   - Booting up control plane ...
	I0830 23:04:20.922506 1258949 kubeadm.go:322] [control-plane] Creating static Pod manifest for "kube-apiserver"
	I0830 23:04:20.932749 1258949 kubeadm.go:322] [control-plane] Creating static Pod manifest for "kube-controller-manager"
	I0830 23:04:20.937854 1258949 kubeadm.go:322] [control-plane] Creating static Pod manifest for "kube-scheduler"
	I0830 23:04:20.938972 1258949 kubeadm.go:322] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
	I0830 23:04:20.942395 1258949 kubeadm.go:322] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
	I0830 23:04:34.451230 1258949 kubeadm.go:322] [apiclient] All control plane components are healthy after 13.502622 seconds
	I0830 23:04:34.451363 1258949 kubeadm.go:322] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
	I0830 23:04:34.467646 1258949 kubeadm.go:322] [kubelet] Creating a ConfigMap "kubelet-config-1.18" in namespace kube-system with the configuration for the kubelets in the cluster
	I0830 23:04:34.994353 1258949 kubeadm.go:322] [upload-certs] Skipping phase. Please see --upload-certs
	I0830 23:04:34.994495 1258949 kubeadm.go:322] [mark-control-plane] Marking the node ingress-addon-legacy-235029 as control-plane by adding the label "node-role.kubernetes.io/master=''"
	I0830 23:04:35.504463 1258949 kubeadm.go:322] [bootstrap-token] Using token: konhvr.q5ajvf7af58wkebs
	I0830 23:04:35.506856 1258949 out.go:204]   - Configuring RBAC rules ...
	I0830 23:04:35.507024 1258949 kubeadm.go:322] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
	I0830 23:04:35.512636 1258949 kubeadm.go:322] [bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to get nodes
	I0830 23:04:35.521774 1258949 kubeadm.go:322] [bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
	I0830 23:04:35.524960 1258949 kubeadm.go:322] [bootstrap-token] configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
	I0830 23:04:35.531640 1258949 kubeadm.go:322] [bootstrap-token] configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
	I0830 23:04:35.537741 1258949 kubeadm.go:322] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
	I0830 23:04:35.553580 1258949 kubeadm.go:322] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
	I0830 23:04:35.824153 1258949 kubeadm.go:322] [addons] Applied essential addon: CoreDNS
	I0830 23:04:35.993504 1258949 kubeadm.go:322] [addons] Applied essential addon: kube-proxy
	I0830 23:04:35.993526 1258949 kubeadm.go:322] 
	I0830 23:04:35.993584 1258949 kubeadm.go:322] Your Kubernetes control-plane has initialized successfully!
	I0830 23:04:35.993593 1258949 kubeadm.go:322] 
	I0830 23:04:35.993671 1258949 kubeadm.go:322] To start using your cluster, you need to run the following as a regular user:
	I0830 23:04:35.993680 1258949 kubeadm.go:322] 
	I0830 23:04:35.993704 1258949 kubeadm.go:322]   mkdir -p $HOME/.kube
	I0830 23:04:35.993773 1258949 kubeadm.go:322]   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
	I0830 23:04:35.993832 1258949 kubeadm.go:322]   sudo chown $(id -u):$(id -g) $HOME/.kube/config
	I0830 23:04:35.993840 1258949 kubeadm.go:322] 
	I0830 23:04:35.993890 1258949 kubeadm.go:322] You should now deploy a pod network to the cluster.
	I0830 23:04:35.993964 1258949 kubeadm.go:322] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
	I0830 23:04:35.994032 1258949 kubeadm.go:322]   https://kubernetes.io/docs/concepts/cluster-administration/addons/
	I0830 23:04:35.994039 1258949 kubeadm.go:322] 
	I0830 23:04:35.994118 1258949 kubeadm.go:322] You can now join any number of control-plane nodes by copying certificate authorities
	I0830 23:04:35.994193 1258949 kubeadm.go:322] and service account keys on each node and then running the following as root:
	I0830 23:04:35.994200 1258949 kubeadm.go:322] 
	I0830 23:04:35.994278 1258949 kubeadm.go:322]   kubeadm join control-plane.minikube.internal:8443 --token konhvr.q5ajvf7af58wkebs \
	I0830 23:04:35.994382 1258949 kubeadm.go:322]     --discovery-token-ca-cert-hash sha256:0d761a60bc6e6ab59708493c3270ff221f9632510a1573cc948daf3f99318f4a \
	I0830 23:04:35.994407 1258949 kubeadm.go:322]     --control-plane 
	I0830 23:04:35.994415 1258949 kubeadm.go:322] 
	I0830 23:04:35.994495 1258949 kubeadm.go:322] Then you can join any number of worker nodes by running the following on each as root:
	I0830 23:04:35.994503 1258949 kubeadm.go:322] 
	I0830 23:04:35.994581 1258949 kubeadm.go:322] kubeadm join control-plane.minikube.internal:8443 --token konhvr.q5ajvf7af58wkebs \
	I0830 23:04:35.994752 1258949 kubeadm.go:322]     --discovery-token-ca-cert-hash sha256:0d761a60bc6e6ab59708493c3270ff221f9632510a1573cc948daf3f99318f4a 
	I0830 23:04:35.997416 1258949 kubeadm.go:322] W0830 23:04:14.411023    1120 configset.go:202] WARNING: kubeadm cannot validate component configs for API groups [kubelet.config.k8s.io kubeproxy.config.k8s.io]
	I0830 23:04:35.997626 1258949 kubeadm.go:322] 	[WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1043-aws\n", err: exit status 1
	I0830 23:04:35.997725 1258949 kubeadm.go:322] 	[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
	I0830 23:04:35.997844 1258949 kubeadm.go:322] W0830 23:04:20.932507    1120 manifests.go:225] the default kube-apiserver authorization-mode is "Node,RBAC"; using "Node,RBAC"
	I0830 23:04:35.997960 1258949 kubeadm.go:322] W0830 23:04:20.937665    1120 manifests.go:225] the default kube-apiserver authorization-mode is "Node,RBAC"; using "Node,RBAC"
	I0830 23:04:35.997973 1258949 cni.go:84] Creating CNI manager for ""
	I0830 23:04:35.997979 1258949 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
	I0830 23:04:35.999835 1258949 out.go:177] * Configuring CNI (Container Networking Interface) ...
	I0830 23:04:36.001686 1258949 ssh_runner.go:195] Run: stat /opt/cni/bin/portmap
	I0830 23:04:36.006649 1258949 cni.go:182] applying CNI manifest using /var/lib/minikube/binaries/v1.18.20/kubectl ...
	I0830 23:04:36.006666 1258949 ssh_runner.go:362] scp memory --> /var/tmp/minikube/cni.yaml (2438 bytes)
	I0830 23:04:36.034615 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml
	I0830 23:04:36.509904 1258949 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
	I0830 23:04:36.510047 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:36.510136 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl label nodes minikube.k8s.io/version=v1.31.2 minikube.k8s.io/commit=dcfed3f069eb419c2ffae8f904d3fba5b9405fc5 minikube.k8s.io/name=ingress-addon-legacy-235029 minikube.k8s.io/updated_at=2023_08_30T23_04_36_0700 minikube.k8s.io/primary=true --all --overwrite --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:36.671426 1258949 ops.go:34] apiserver oom_adj: -16
	I0830 23:04:36.674983 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:36.792537 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:37.391173 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:37.890555 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:38.390805 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:38.891003 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:39.391403 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:39.890989 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:40.391536 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:40.890632 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:41.390862 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:41.890794 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:42.390957 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:42.891176 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:43.390546 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:43.891263 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:44.390571 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:44.891072 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:45.390724 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:45.891486 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:46.391137 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:46.890806 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:47.391012 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:47.891475 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:48.390925 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:48.890809 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:49.390852 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:49.890920 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:50.390818 1258949 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.18.20/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0830 23:04:50.545225 1258949 kubeadm.go:1081] duration metric: took 14.035222888s to wait for elevateKubeSystemPrivileges.
	I0830 23:04:50.545263 1258949 kubeadm.go:406] StartCluster complete in 36.267315646s
	I0830 23:04:50.545279 1258949 settings.go:142] acquiring lock: {Name:mk7f59b3a5ac74e6581a6632908a77e182a8ae89 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 23:04:50.545349 1258949 settings.go:150] Updating kubeconfig:  /home/jenkins/minikube-integration/17114-1219981/kubeconfig
	I0830 23:04:50.546051 1258949 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/kubeconfig: {Name:mkdc9959048bf1aa0b8afb8aa74417552569b184 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 23:04:50.546277 1258949 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.18.20/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
	I0830 23:04:50.546671 1258949 config.go:182] Loaded profile config "ingress-addon-legacy-235029": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.18.20
	I0830 23:04:50.546800 1258949 addons.go:499] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false helm-tiller:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false volumesnapshots:false]
	I0830 23:04:50.546878 1258949 addons.go:69] Setting storage-provisioner=true in profile "ingress-addon-legacy-235029"
	I0830 23:04:50.546892 1258949 addons.go:231] Setting addon storage-provisioner=true in "ingress-addon-legacy-235029"
	I0830 23:04:50.546948 1258949 host.go:66] Checking if "ingress-addon-legacy-235029" exists ...
	I0830 23:04:50.546900 1258949 kapi.go:59] client config for ingress-addon-legacy-235029: &rest.Config{Host:"https://192.168.49.2:8443", APIPath:"", ContentConfig:rest.ContentConfig{AcceptContentTypes:"", ContentType:"", GroupVersion:(*schema.GroupVersion)(nil), NegotiatedSerializer:runtime.NegotiatedSerializer(nil)}, Username:"", Password:"", BearerToken:"", BearerTokenFile:"", Impersonate:rest.ImpersonationConfig{UserName:"", UID:"", Groups:[]string(nil), Extra:map[string][]string(nil)}, AuthProvider:<nil>, AuthConfigPersister:rest.AuthProviderConfigPersister(nil), ExecProvider:<nil>, TLSClientConfig:rest.sanitizedTLSClientConfig{Insecure:false, ServerName:"", CertFile:"/home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt", KeyFile:"/home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.key", CAFile:"/home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt", CertData:[]uint8(nil), KeyData:[
]uint8(nil), CAData:[]uint8(nil), NextProtos:[]string(nil)}, UserAgent:"", DisableCompression:false, Transport:http.RoundTripper(nil), WrapTransport:(transport.WrapperFunc)(0x1723960), QPS:0, Burst:0, RateLimiter:flowcontrol.RateLimiter(nil), WarningHandler:rest.WarningHandler(nil), Timeout:0, Dial:(func(context.Context, string, string) (net.Conn, error))(nil), Proxy:(func(*http.Request) (*url.URL, error))(nil)}
	I0830 23:04:50.547413 1258949 cli_runner.go:164] Run: docker container inspect ingress-addon-legacy-235029 --format={{.State.Status}}
	I0830 23:04:50.547880 1258949 addons.go:69] Setting default-storageclass=true in profile "ingress-addon-legacy-235029"
	I0830 23:04:50.547901 1258949 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "ingress-addon-legacy-235029"
	I0830 23:04:50.548167 1258949 cert_rotation.go:137] Starting client certificate rotation controller
	I0830 23:04:50.548181 1258949 cli_runner.go:164] Run: docker container inspect ingress-addon-legacy-235029 --format={{.State.Status}}
	I0830 23:04:50.597201 1258949 out.go:177]   - Using image gcr.io/k8s-minikube/storage-provisioner:v5
	I0830 23:04:50.599260 1258949 addons.go:423] installing /etc/kubernetes/addons/storage-provisioner.yaml
	I0830 23:04:50.599283 1258949 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
	I0830 23:04:50.599362 1258949 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" ingress-addon-legacy-235029
	I0830 23:04:50.605971 1258949 kapi.go:59] client config for ingress-addon-legacy-235029: &rest.Config{Host:"https://192.168.49.2:8443", APIPath:"", ContentConfig:rest.ContentConfig{AcceptContentTypes:"", ContentType:"", GroupVersion:(*schema.GroupVersion)(nil), NegotiatedSerializer:runtime.NegotiatedSerializer(nil)}, Username:"", Password:"", BearerToken:"", BearerTokenFile:"", Impersonate:rest.ImpersonationConfig{UserName:"", UID:"", Groups:[]string(nil), Extra:map[string][]string(nil)}, AuthProvider:<nil>, AuthConfigPersister:rest.AuthProviderConfigPersister(nil), ExecProvider:<nil>, TLSClientConfig:rest.sanitizedTLSClientConfig{Insecure:false, ServerName:"", CertFile:"/home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt", KeyFile:"/home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.key", CAFile:"/home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt", CertData:[]uint8(nil), KeyData:[
]uint8(nil), CAData:[]uint8(nil), NextProtos:[]string(nil)}, UserAgent:"", DisableCompression:false, Transport:http.RoundTripper(nil), WrapTransport:(transport.WrapperFunc)(0x1723960), QPS:0, Burst:0, RateLimiter:flowcontrol.RateLimiter(nil), WarningHandler:rest.WarningHandler(nil), Timeout:0, Dial:(func(context.Context, string, string) (net.Conn, error))(nil), Proxy:(func(*http.Request) (*url.URL, error))(nil)}
	I0830 23:04:50.606574 1258949 kapi.go:248] "coredns" deployment in "kube-system" namespace and "ingress-addon-legacy-235029" context rescaled to 1 replicas
	I0830 23:04:50.606765 1258949 start.go:223] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.18.20 ContainerRuntime:containerd ControlPlane:true Worker:true}
	I0830 23:04:50.608940 1258949 out.go:177] * Verifying Kubernetes components...
	I0830 23:04:50.611043 1258949 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0830 23:04:50.627909 1258949 addons.go:231] Setting addon default-storageclass=true in "ingress-addon-legacy-235029"
	I0830 23:04:50.627954 1258949 host.go:66] Checking if "ingress-addon-legacy-235029" exists ...
	I0830 23:04:50.628406 1258949 cli_runner.go:164] Run: docker container inspect ingress-addon-legacy-235029 --format={{.State.Status}}
	I0830 23:04:50.638905 1258949 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34334 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/ingress-addon-legacy-235029/id_rsa Username:docker}
	I0830 23:04:50.673717 1258949 addons.go:423] installing /etc/kubernetes/addons/storageclass.yaml
	I0830 23:04:50.673738 1258949 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
	I0830 23:04:50.673808 1258949 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" ingress-addon-legacy-235029
	I0830 23:04:50.706423 1258949 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34334 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/ingress-addon-legacy-235029/id_rsa Username:docker}
	I0830 23:04:50.752938 1258949 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.18.20/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^        forward . \/etc\/resolv.conf.*/i \        hosts {\n           192.168.49.1 host.minikube.internal\n           fallthrough\n        }' -e '/^        errors *$/i \        log' | sudo /var/lib/minikube/binaries/v1.18.20/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
	I0830 23:04:50.753633 1258949 kapi.go:59] client config for ingress-addon-legacy-235029: &rest.Config{Host:"https://192.168.49.2:8443", APIPath:"", ContentConfig:rest.ContentConfig{AcceptContentTypes:"", ContentType:"", GroupVersion:(*schema.GroupVersion)(nil), NegotiatedSerializer:runtime.NegotiatedSerializer(nil)}, Username:"", Password:"", BearerToken:"", BearerTokenFile:"", Impersonate:rest.ImpersonationConfig{UserName:"", UID:"", Groups:[]string(nil), Extra:map[string][]string(nil)}, AuthProvider:<nil>, AuthConfigPersister:rest.AuthProviderConfigPersister(nil), ExecProvider:<nil>, TLSClientConfig:rest.sanitizedTLSClientConfig{Insecure:false, ServerName:"", CertFile:"/home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt", KeyFile:"/home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.key", CAFile:"/home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt", CertData:[]uint8(nil), KeyData:[
]uint8(nil), CAData:[]uint8(nil), NextProtos:[]string(nil)}, UserAgent:"", DisableCompression:false, Transport:http.RoundTripper(nil), WrapTransport:(transport.WrapperFunc)(0x1723960), QPS:0, Burst:0, RateLimiter:flowcontrol.RateLimiter(nil), WarningHandler:rest.WarningHandler(nil), Timeout:0, Dial:(func(context.Context, string, string) (net.Conn, error))(nil), Proxy:(func(*http.Request) (*url.URL, error))(nil)}
	I0830 23:04:50.753895 1258949 node_ready.go:35] waiting up to 6m0s for node "ingress-addon-legacy-235029" to be "Ready" ...
	I0830 23:04:50.757107 1258949 node_ready.go:49] node "ingress-addon-legacy-235029" has status "Ready":"True"
	I0830 23:04:50.757152 1258949 node_ready.go:38] duration metric: took 3.241404ms waiting for node "ingress-addon-legacy-235029" to be "Ready" ...
	I0830 23:04:50.757163 1258949 pod_ready.go:35] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
	I0830 23:04:50.763713 1258949 pod_ready.go:78] waiting up to 6m0s for pod "etcd-ingress-addon-legacy-235029" in "kube-system" namespace to be "Ready" ...
	I0830 23:04:50.778243 1258949 pod_ready.go:92] pod "etcd-ingress-addon-legacy-235029" in "kube-system" namespace has status "Ready":"True"
	I0830 23:04:50.778274 1258949 pod_ready.go:81] duration metric: took 14.527276ms waiting for pod "etcd-ingress-addon-legacy-235029" in "kube-system" namespace to be "Ready" ...
	I0830 23:04:50.778287 1258949 pod_ready.go:78] waiting up to 6m0s for pod "kube-apiserver-ingress-addon-legacy-235029" in "kube-system" namespace to be "Ready" ...
	I0830 23:04:50.796265 1258949 pod_ready.go:92] pod "kube-apiserver-ingress-addon-legacy-235029" in "kube-system" namespace has status "Ready":"True"
	I0830 23:04:50.796293 1258949 pod_ready.go:81] duration metric: took 17.974798ms waiting for pod "kube-apiserver-ingress-addon-legacy-235029" in "kube-system" namespace to be "Ready" ...
	I0830 23:04:50.796329 1258949 pod_ready.go:78] waiting up to 6m0s for pod "kube-controller-manager-ingress-addon-legacy-235029" in "kube-system" namespace to be "Ready" ...
	I0830 23:04:50.855984 1258949 pod_ready.go:92] pod "kube-controller-manager-ingress-addon-legacy-235029" in "kube-system" namespace has status "Ready":"True"
	I0830 23:04:50.856014 1258949 pod_ready.go:81] duration metric: took 59.670224ms waiting for pod "kube-controller-manager-ingress-addon-legacy-235029" in "kube-system" namespace to be "Ready" ...
	I0830 23:04:50.856025 1258949 pod_ready.go:78] waiting up to 6m0s for pod "kube-scheduler-ingress-addon-legacy-235029" in "kube-system" namespace to be "Ready" ...
	I0830 23:04:50.888025 1258949 pod_ready.go:92] pod "kube-scheduler-ingress-addon-legacy-235029" in "kube-system" namespace has status "Ready":"True"
	I0830 23:04:50.888055 1258949 pod_ready.go:81] duration metric: took 32.019681ms waiting for pod "kube-scheduler-ingress-addon-legacy-235029" in "kube-system" namespace to be "Ready" ...
	I0830 23:04:50.888065 1258949 pod_ready.go:38] duration metric: took 130.867606ms for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
	I0830 23:04:50.888113 1258949 api_server.go:52] waiting for apiserver process to appear ...
	I0830 23:04:50.888203 1258949 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:04:50.957195 1258949 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.18.20/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
	I0830 23:04:50.967950 1258949 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.18.20/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
	I0830 23:04:51.307090 1258949 start.go:901] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
	I0830 23:04:51.307171 1258949 api_server.go:72] duration metric: took 700.218849ms to wait for apiserver process to appear ...
	I0830 23:04:51.307192 1258949 api_server.go:88] waiting for apiserver healthz status ...
	I0830 23:04:51.307209 1258949 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
	I0830 23:04:51.317039 1258949 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
	ok
	I0830 23:04:51.318385 1258949 api_server.go:141] control plane version: v1.18.20
	I0830 23:04:51.318418 1258949 api_server.go:131] duration metric: took 11.219063ms to wait for apiserver health ...
	I0830 23:04:51.318442 1258949 system_pods.go:43] waiting for kube-system pods to appear ...
	I0830 23:04:51.326150 1258949 system_pods.go:59] 7 kube-system pods found
	I0830 23:04:51.326198 1258949 system_pods.go:61] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending: PodScheduled:Unschedulable (0/1 nodes are available: 1 node(s) had taint {node.kubernetes.io/not-ready: }, that the pod didn't tolerate.)
	I0830 23:04:51.326207 1258949 system_pods.go:61] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:04:51.326233 1258949 system_pods.go:61] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Pending / Ready:ContainersNotReady (containers with unready status: [kindnet-cni]) / ContainersReady:ContainersNotReady (containers with unready status: [kindnet-cni])
	I0830 23:04:51.326249 1258949 system_pods.go:61] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:04:51.326255 1258949 system_pods.go:61] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:04:51.326281 1258949 system_pods.go:61] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Pending / Ready:ContainersNotReady (containers with unready status: [kube-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-proxy])
	I0830 23:04:51.326293 1258949 system_pods.go:61] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:04:51.326300 1258949 system_pods.go:74] duration metric: took 7.844941ms to wait for pod list to return data ...
	I0830 23:04:51.326323 1258949 default_sa.go:34] waiting for default service account to be created ...
	I0830 23:04:51.329458 1258949 default_sa.go:45] found service account: "default"
	I0830 23:04:51.329485 1258949 default_sa.go:55] duration metric: took 3.155026ms for default service account to be created ...
	I0830 23:04:51.329495 1258949 system_pods.go:116] waiting for k8s-apps to be running ...
	I0830 23:04:51.361715 1258949 system_pods.go:86] 7 kube-system pods found
	I0830 23:04:51.361759 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending: PodScheduled:Unschedulable (0/1 nodes are available: 1 node(s) had taint {node.kubernetes.io/not-ready: }, that the pod didn't tolerate.)
	I0830 23:04:51.361771 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:04:51.361801 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Pending / Ready:ContainersNotReady (containers with unready status: [kindnet-cni]) / ContainersReady:ContainersNotReady (containers with unready status: [kindnet-cni])
	I0830 23:04:51.361814 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:04:51.361835 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:04:51.361848 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Pending / Ready:ContainersNotReady (containers with unready status: [kube-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-proxy])
	I0830 23:04:51.361863 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:04:51.361885 1258949 retry.go:31] will retry after 243.882444ms: missing components: kube-dns, kube-proxy
	I0830 23:04:51.481795 1258949 out.go:177] * Enabled addons: storage-provisioner, default-storageclass
	I0830 23:04:51.483497 1258949 addons.go:502] enable addons completed in 936.677751ms: enabled=[storage-provisioner default-storageclass]
	I0830 23:04:51.613673 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:04:51.613706 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending: PodScheduled:Unschedulable (0/1 nodes are available: 1 node(s) had taint {node.kubernetes.io/not-ready: }, that the pod didn't tolerate.)
	I0830 23:04:51.613715 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:04:51.613723 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Pending / Ready:ContainersNotReady (containers with unready status: [kindnet-cni]) / ContainersReady:ContainersNotReady (containers with unready status: [kindnet-cni])
	I0830 23:04:51.613729 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:04:51.613735 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:04:51.613742 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Pending / Ready:ContainersNotReady (containers with unready status: [kube-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-proxy])
	I0830 23:04:51.613754 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:04:51.613763 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Pending / Ready:ContainersNotReady (containers with unready status: [storage-provisioner]) / ContainersReady:ContainersNotReady (containers with unready status: [storage-provisioner])
	I0830 23:04:51.613778 1258949 retry.go:31] will retry after 360.068023ms: missing components: kube-dns, kube-proxy
	I0830 23:04:51.980393 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:04:51.980487 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending / Ready:ContainersNotReady (containers with unready status: [coredns]) / ContainersReady:ContainersNotReady (containers with unready status: [coredns])
	I0830 23:04:51.980504 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:04:51.980513 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Pending / Ready:ContainersNotReady (containers with unready status: [kindnet-cni]) / ContainersReady:ContainersNotReady (containers with unready status: [kindnet-cni])
	I0830 23:04:51.980518 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:04:51.980536 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:04:51.980549 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Pending / Ready:ContainersNotReady (containers with unready status: [kube-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-proxy])
	I0830 23:04:51.980555 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:04:51.980561 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Pending / Ready:ContainersNotReady (containers with unready status: [storage-provisioner]) / ContainersReady:ContainersNotReady (containers with unready status: [storage-provisioner])
	I0830 23:04:51.980578 1258949 retry.go:31] will retry after 477.23931ms: missing components: kube-dns, kube-proxy
	I0830 23:04:52.465419 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:04:52.465452 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending / Ready:ContainersNotReady (containers with unready status: [coredns]) / ContainersReady:ContainersNotReady (containers with unready status: [coredns])
	I0830 23:04:52.465460 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:04:52.465468 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Pending / Ready:ContainersNotReady (containers with unready status: [kindnet-cni]) / ContainersReady:ContainersNotReady (containers with unready status: [kindnet-cni])
	I0830 23:04:52.465474 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:04:52.465480 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:04:52.465486 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Pending / Ready:ContainersNotReady (containers with unready status: [kube-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-proxy])
	I0830 23:04:52.465494 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:04:52.465501 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Pending / Ready:ContainersNotReady (containers with unready status: [storage-provisioner]) / ContainersReady:ContainersNotReady (containers with unready status: [storage-provisioner])
	I0830 23:04:52.465522 1258949 retry.go:31] will retry after 598.979769ms: missing components: kube-dns, kube-proxy
	I0830 23:04:53.072191 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:04:53.072228 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending / Ready:ContainersNotReady (containers with unready status: [coredns]) / ContainersReady:ContainersNotReady (containers with unready status: [coredns])
	I0830 23:04:53.072243 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:04:53.072254 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Pending / Ready:ContainersNotReady (containers with unready status: [kindnet-cni]) / ContainersReady:ContainersNotReady (containers with unready status: [kindnet-cni])
	I0830 23:04:53.072260 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:04:53.072266 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:04:53.072273 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Pending / Ready:ContainersNotReady (containers with unready status: [kube-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-proxy])
	I0830 23:04:53.072286 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:04:53.072296 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Pending / Ready:ContainersNotReady (containers with unready status: [storage-provisioner]) / ContainersReady:ContainersNotReady (containers with unready status: [storage-provisioner])
	I0830 23:04:53.072311 1258949 retry.go:31] will retry after 736.106853ms: missing components: kube-dns, kube-proxy
	I0830 23:04:53.815718 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:04:53.815756 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending / Ready:ContainersNotReady (containers with unready status: [coredns]) / ContainersReady:ContainersNotReady (containers with unready status: [coredns])
	I0830 23:04:53.815764 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:04:53.815771 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Running
	I0830 23:04:53.815780 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:04:53.815787 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:04:53.815796 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Running
	I0830 23:04:53.815801 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:04:53.815807 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Pending / Ready:ContainersNotReady (containers with unready status: [storage-provisioner]) / ContainersReady:ContainersNotReady (containers with unready status: [storage-provisioner])
	I0830 23:04:53.815822 1258949 retry.go:31] will retry after 822.574733ms: missing components: kube-dns
	I0830 23:04:54.645420 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:04:54.645457 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending / Ready:ContainersNotReady (containers with unready status: [coredns]) / ContainersReady:ContainersNotReady (containers with unready status: [coredns])
	I0830 23:04:54.645466 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:04:54.645473 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Running
	I0830 23:04:54.645478 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:04:54.645506 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:04:54.645516 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Running
	I0830 23:04:54.645522 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:04:54.645532 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Pending / Ready:ContainersNotReady (containers with unready status: [storage-provisioner]) / ContainersReady:ContainersNotReady (containers with unready status: [storage-provisioner])
	I0830 23:04:54.645546 1258949 retry.go:31] will retry after 794.2633ms: missing components: kube-dns
	I0830 23:04:55.446445 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:04:55.446484 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending / Ready:ContainersNotReady (containers with unready status: [coredns]) / ContainersReady:ContainersNotReady (containers with unready status: [coredns])
	I0830 23:04:55.446493 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:04:55.446499 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Running
	I0830 23:04:55.446504 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:04:55.446509 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:04:55.446514 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Running
	I0830 23:04:55.446522 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:04:55.446528 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Running
	I0830 23:04:55.446543 1258949 retry.go:31] will retry after 990.493001ms: missing components: kube-dns
	I0830 23:04:56.443451 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:04:56.443490 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending / Ready:ContainersNotReady (containers with unready status: [coredns]) / ContainersReady:ContainersNotReady (containers with unready status: [coredns])
	I0830 23:04:56.443499 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:04:56.443505 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Running
	I0830 23:04:56.443511 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:04:56.443517 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:04:56.443522 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Running
	I0830 23:04:56.443528 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:04:56.443538 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Running
	I0830 23:04:56.443554 1258949 retry.go:31] will retry after 1.292577237s: missing components: kube-dns
	I0830 23:04:57.743378 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:04:57.743417 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending / Ready:ContainersNotReady (containers with unready status: [coredns]) / ContainersReady:ContainersNotReady (containers with unready status: [coredns])
	I0830 23:04:57.743425 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:04:57.743433 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Running
	I0830 23:04:57.743442 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:04:57.743453 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:04:57.743458 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Running
	I0830 23:04:57.743467 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:04:57.743476 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Running
	I0830 23:04:57.743495 1258949 retry.go:31] will retry after 1.910847026s: missing components: kube-dns
	I0830 23:04:59.661198 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:04:59.661234 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending / Ready:ContainersNotReady (containers with unready status: [coredns]) / ContainersReady:ContainersNotReady (containers with unready status: [coredns])
	I0830 23:04:59.661242 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:04:59.661253 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Running
	I0830 23:04:59.661258 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:04:59.661264 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:04:59.661269 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Running
	I0830 23:04:59.661274 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:04:59.661278 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Running
	I0830 23:04:59.661294 1258949 retry.go:31] will retry after 1.994210673s: missing components: kube-dns
	I0830 23:05:01.662558 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:05:01.662602 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending / Ready:ContainersNotReady (containers with unready status: [coredns]) / ContainersReady:ContainersNotReady (containers with unready status: [coredns])
	I0830 23:05:01.662610 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:05:01.662647 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Running
	I0830 23:05:01.662660 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:05:01.662666 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:05:01.662672 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Running
	I0830 23:05:01.662681 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:05:01.662686 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Running
	I0830 23:05:01.662703 1258949 retry.go:31] will retry after 2.275108918s: missing components: kube-dns
	I0830 23:05:03.944090 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:05:03.944124 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending / Ready:ContainersNotReady (containers with unready status: [coredns]) / ContainersReady:ContainersNotReady (containers with unready status: [coredns])
	I0830 23:05:03.944132 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:05:03.944139 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Running
	I0830 23:05:03.944145 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:05:03.944151 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:05:03.944158 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Running
	I0830 23:05:03.944169 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:05:03.944174 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Running
	I0830 23:05:03.944190 1258949 retry.go:31] will retry after 4.128672088s: missing components: kube-dns
	I0830 23:05:08.081305 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:05:08.081385 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Pending / Ready:ContainersNotReady (containers with unready status: [coredns]) / ContainersReady:ContainersNotReady (containers with unready status: [coredns])
	I0830 23:05:08.081409 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:05:08.081425 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Running
	I0830 23:05:08.081430 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:05:08.081448 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:05:08.081460 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Running
	I0830 23:05:08.081466 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:05:08.081470 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Running
	I0830 23:05:08.081495 1258949 retry.go:31] will retry after 5.431720131s: missing components: kube-dns
	I0830 23:05:13.519964 1258949 system_pods.go:86] 8 kube-system pods found
	I0830 23:05:13.519999 1258949 system_pods.go:89] "coredns-66bff467f8-j788v" [e589356a-6cad-43bd-b016-3894057fff15] Running
	I0830 23:05:13.520007 1258949 system_pods.go:89] "etcd-ingress-addon-legacy-235029" [0c76228a-29e3-44a1-ae8d-1bc83639745c] Running
	I0830 23:05:13.520012 1258949 system_pods.go:89] "kindnet-4jl78" [9a438dde-389a-408d-bf7b-d4d8b35b2572] Running
	I0830 23:05:13.520018 1258949 system_pods.go:89] "kube-apiserver-ingress-addon-legacy-235029" [71060b2e-b6cf-4a94-896d-c6ce51e86de5] Running
	I0830 23:05:13.520023 1258949 system_pods.go:89] "kube-controller-manager-ingress-addon-legacy-235029" [6fae8615-e002-4a6f-8842-544a28a68f5a] Running
	I0830 23:05:13.520028 1258949 system_pods.go:89] "kube-proxy-2bhx8" [ca73749f-fa3c-4439-a71d-566c0c1f0844] Running
	I0830 23:05:13.520034 1258949 system_pods.go:89] "kube-scheduler-ingress-addon-legacy-235029" [5f158346-07cf-4296-b6c0-baa1e6faf82e] Running
	I0830 23:05:13.520039 1258949 system_pods.go:89] "storage-provisioner" [789200f1-05da-4cc8-9564-c48050fb7389] Running
	I0830 23:05:13.520045 1258949 system_pods.go:126] duration metric: took 22.190515604s to wait for k8s-apps to be running ...
	I0830 23:05:13.520098 1258949 system_svc.go:44] waiting for kubelet service to be running ....
	I0830 23:05:13.520186 1258949 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0830 23:05:13.534220 1258949 system_svc.go:56] duration metric: took 14.112395ms WaitForService to wait for kubelet.
	I0830 23:05:13.534246 1258949 kubeadm.go:581] duration metric: took 22.927317445s to wait for : map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] ...
	I0830 23:05:13.534266 1258949 node_conditions.go:102] verifying NodePressure condition ...
	I0830 23:05:13.537618 1258949 node_conditions.go:122] node storage ephemeral capacity is 203034800Ki
	I0830 23:05:13.537658 1258949 node_conditions.go:123] node cpu capacity is 2
	I0830 23:05:13.537670 1258949 node_conditions.go:105] duration metric: took 3.399333ms to run NodePressure ...
	I0830 23:05:13.537702 1258949 start.go:228] waiting for startup goroutines ...
	I0830 23:05:13.537716 1258949 start.go:233] waiting for cluster config update ...
	I0830 23:05:13.537727 1258949 start.go:242] writing updated cluster config ...
	I0830 23:05:13.538042 1258949 ssh_runner.go:195] Run: rm -f paused
	I0830 23:05:13.598808 1258949 start.go:600] kubectl: 1.28.1, cluster: 1.18.20 (minor skew: 10)
	I0830 23:05:13.601308 1258949 out.go:177] 
	W0830 23:05:13.603148 1258949 out.go:239] ! /usr/local/bin/kubectl is version 1.28.1, which may have incompatibilities with Kubernetes 1.18.20.
	I0830 23:05:13.604730 1258949 out.go:177]   - Want kubectl v1.18.20? Try 'minikube kubectl -- get pods -A'
	I0830 23:05:13.606582 1258949 out.go:177] * Done! kubectl is now configured to use "ingress-addon-legacy-235029" cluster and "default" namespace by default
	
	* 
	* ==> container status <==
	* CONTAINER           IMAGE               CREATED              STATE               NAME                      ATTEMPT             POD ID              POD
	fb36ffc4bab1f       13753a81eccfd       18 seconds ago       Exited              hello-world-app           2                   82edfabe221ce       hello-world-app-5f5d8b66bb-6szz4
	13e0cd05b611f       fa0c6bb795403       41 seconds ago       Running             nginx                     0                   e55071f6f71fd       nginx
	1f95a85444839       d7f0cba3aa5bf       55 seconds ago       Exited              controller                0                   0b5712dee820f       ingress-nginx-controller-7fcf777cb7-jfv5q
	b376f51e09257       a883f7fc35610       About a minute ago   Exited              patch                     0                   4b4f81f5be909       ingress-nginx-admission-patch-77kp5
	549dc9b88c6fd       a883f7fc35610       About a minute ago   Exited              create                    0                   03aea956919ab       ingress-nginx-admission-create-jq4xk
	fc80a9fc9852a       6e17ba78cf3eb       About a minute ago   Running             coredns                   0                   a9731d5efbf97       coredns-66bff467f8-j788v
	04bb013421549       ba04bb24b9575       About a minute ago   Running             storage-provisioner       0                   c927bdc3d29d6       storage-provisioner
	f44fee6e9407e       b18bf71b941ba       About a minute ago   Running             kindnet-cni               0                   3509b5158dfa4       kindnet-4jl78
	d6417a8a2e5ee       565297bc6f7d4       About a minute ago   Running             kube-proxy                0                   f7c34719e948b       kube-proxy-2bhx8
	eebf8ed564e5c       2694cf044d665       About a minute ago   Running             kube-apiserver            0                   5200b94a7583c       kube-apiserver-ingress-addon-legacy-235029
	5b6430073e5fc       68a4fac29a865       About a minute ago   Running             kube-controller-manager   0                   c5d0d29757aa3       kube-controller-manager-ingress-addon-legacy-235029
	c2f9d005378fd       ab707b0a0ea33       About a minute ago   Running             etcd                      0                   df98c58884912       etcd-ingress-addon-legacy-235029
	3620e41650f7f       095f37015706d       About a minute ago   Running             kube-scheduler            0                   3c047f349e8dc       kube-scheduler-ingress-addon-legacy-235029
	
	* 
	* ==> containerd <==
	* Aug 30 23:06:01 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:01.570362282Z" level=info msg="RemoveContainer for \"126bc6c72d31035907a3402b7ab5b8e832901ce36a5efd712d4dd1898aef0a38\" returns successfully"
	Aug 30 23:06:10 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:10.221768377Z" level=info msg="StopContainer for \"1f95a85444839bea630cd8056fd2f3d5e99dd255db14b29e37aead92b85647d6\" with timeout 2 (s)"
	Aug 30 23:06:10 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:10.222546288Z" level=info msg="Stop container \"1f95a85444839bea630cd8056fd2f3d5e99dd255db14b29e37aead92b85647d6\" with signal terminated"
	Aug 30 23:06:10 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:10.235776055Z" level=info msg="StopContainer for \"1f95a85444839bea630cd8056fd2f3d5e99dd255db14b29e37aead92b85647d6\" with timeout 2 (s)"
	Aug 30 23:06:10 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:10.238035327Z" level=info msg="Skipping the sending of signal terminated to container \"1f95a85444839bea630cd8056fd2f3d5e99dd255db14b29e37aead92b85647d6\" because a prior stop with timeout>0 request already sent the signal"
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.235688581Z" level=info msg="Kill container \"1f95a85444839bea630cd8056fd2f3d5e99dd255db14b29e37aead92b85647d6\""
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.240346179Z" level=info msg="Kill container \"1f95a85444839bea630cd8056fd2f3d5e99dd255db14b29e37aead92b85647d6\""
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.347633924Z" level=info msg="shim disconnected" id=1f95a85444839bea630cd8056fd2f3d5e99dd255db14b29e37aead92b85647d6
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.347691747Z" level=warning msg="cleaning up after shim disconnected" id=1f95a85444839bea630cd8056fd2f3d5e99dd255db14b29e37aead92b85647d6 namespace=k8s.io
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.347703766Z" level=info msg="cleaning up dead shim"
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.359564756Z" level=warning msg="cleanup warnings time=\"2023-08-30T23:06:12Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4634 runtime=io.containerd.runc.v2\n"
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.362134672Z" level=info msg="StopContainer for \"1f95a85444839bea630cd8056fd2f3d5e99dd255db14b29e37aead92b85647d6\" returns successfully"
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.362260458Z" level=info msg="StopContainer for \"1f95a85444839bea630cd8056fd2f3d5e99dd255db14b29e37aead92b85647d6\" returns successfully"
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.362747474Z" level=info msg="StopPodSandbox for \"0b5712dee820ff57f1a04b857fa0cec24d6a9327b055097a5ea8426387aa2acb\""
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.362817333Z" level=info msg="Container to stop \"1f95a85444839bea630cd8056fd2f3d5e99dd255db14b29e37aead92b85647d6\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.363030149Z" level=info msg="StopPodSandbox for \"0b5712dee820ff57f1a04b857fa0cec24d6a9327b055097a5ea8426387aa2acb\""
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.363073229Z" level=info msg="Container to stop \"1f95a85444839bea630cd8056fd2f3d5e99dd255db14b29e37aead92b85647d6\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.405310028Z" level=info msg="shim disconnected" id=0b5712dee820ff57f1a04b857fa0cec24d6a9327b055097a5ea8426387aa2acb
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.405378204Z" level=warning msg="cleaning up after shim disconnected" id=0b5712dee820ff57f1a04b857fa0cec24d6a9327b055097a5ea8426387aa2acb namespace=k8s.io
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.405388796Z" level=info msg="cleaning up dead shim"
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.416683272Z" level=warning msg="cleanup warnings time=\"2023-08-30T23:06:12Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4672 runtime=io.containerd.runc.v2\n"
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.479354533Z" level=info msg="TearDown network for sandbox \"0b5712dee820ff57f1a04b857fa0cec24d6a9327b055097a5ea8426387aa2acb\" successfully"
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.479534894Z" level=info msg="StopPodSandbox for \"0b5712dee820ff57f1a04b857fa0cec24d6a9327b055097a5ea8426387aa2acb\" returns successfully"
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.482642059Z" level=info msg="TearDown network for sandbox \"0b5712dee820ff57f1a04b857fa0cec24d6a9327b055097a5ea8426387aa2acb\" successfully"
	Aug 30 23:06:12 ingress-addon-legacy-235029 containerd[836]: time="2023-08-30T23:06:12.482713107Z" level=info msg="StopPodSandbox for \"0b5712dee820ff57f1a04b857fa0cec24d6a9327b055097a5ea8426387aa2acb\" returns successfully"
	
	* 
	* ==> coredns [fc80a9fc9852a957816929f3ea1e5737fc4ab439b02b33be13fa1dff166985f6] <==
	* [INFO] 10.244.0.5:54324 - 25074 "A IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.001503903s
	[INFO] 10.244.0.5:54324 - 7274 "AAAA IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.003059855s
	[INFO] 10.244.0.5:44367 - 46555 "A IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.00093435s
	[INFO] 10.244.0.5:44367 - 49841 "AAAA IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000181672s
	[INFO] 10.244.0.5:39374 - 43635 "AAAA IN hello-world-app.default.svc.cluster.local.svc.cluster.local. udp 77 false 512" NXDOMAIN qr,aa,rd 170 0.000065641s
	[INFO] 10.244.0.5:44367 - 58054 "A IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.001811738s
	[INFO] 10.244.0.5:54324 - 101 "A IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.003418259s
	[INFO] 10.244.0.5:39374 - 6098 "A IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000104709s
	[INFO] 10.244.0.5:32952 - 46274 "A IN hello-world-app.default.svc.cluster.local.ingress-nginx.svc.cluster.local. udp 91 false 512" NXDOMAIN qr,aa,rd 184 0.000081443s
	[INFO] 10.244.0.5:44367 - 14841 "AAAA IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.001821641s
	[INFO] 10.244.0.5:39374 - 52744 "AAAA IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000639067s
	[INFO] 10.244.0.5:54324 - 54578 "AAAA IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.001254113s
	[INFO] 10.244.0.5:54324 - 57262 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000134933s
	[INFO] 10.244.0.5:32952 - 2324 "AAAA IN hello-world-app.default.svc.cluster.local.ingress-nginx.svc.cluster.local. udp 91 false 512" NXDOMAIN qr,aa,rd 184 0.000162442s
	[INFO] 10.244.0.5:44367 - 61051 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000114406s
	[INFO] 10.244.0.5:39374 - 29288 "A IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.002516956s
	[INFO] 10.244.0.5:32952 - 32898 "A IN hello-world-app.default.svc.cluster.local.svc.cluster.local. udp 77 false 512" NXDOMAIN qr,aa,rd 170 0.000035139s
	[INFO] 10.244.0.5:32952 - 30679 "AAAA IN hello-world-app.default.svc.cluster.local.svc.cluster.local. udp 77 false 512" NXDOMAIN qr,aa,rd 170 0.000038084s
	[INFO] 10.244.0.5:32952 - 4551 "A IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000044557s
	[INFO] 10.244.0.5:32952 - 26803 "AAAA IN hello-world-app.default.svc.cluster.local.cluster.local. udp 73 false 512" NXDOMAIN qr,aa,rd 166 0.000035146s
	[INFO] 10.244.0.5:39374 - 17286 "AAAA IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.001398883s
	[INFO] 10.244.0.5:32952 - 25583 "A IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.0008137s
	[INFO] 10.244.0.5:39374 - 60597 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000032111s
	[INFO] 10.244.0.5:32952 - 13489 "AAAA IN hello-world-app.default.svc.cluster.local.us-east-2.compute.internal. udp 86 false 512" NXDOMAIN qr,rd,ra 86 0.001552013s
	[INFO] 10.244.0.5:32952 - 33097 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000039379s
	
	* 
	* ==> describe nodes <==
	* Name:               ingress-addon-legacy-235029
	Roles:              master
	Labels:             beta.kubernetes.io/arch=arm64
	                    beta.kubernetes.io/os=linux
	                    kubernetes.io/arch=arm64
	                    kubernetes.io/hostname=ingress-addon-legacy-235029
	                    kubernetes.io/os=linux
	                    minikube.k8s.io/commit=dcfed3f069eb419c2ffae8f904d3fba5b9405fc5
	                    minikube.k8s.io/name=ingress-addon-legacy-235029
	                    minikube.k8s.io/primary=true
	                    minikube.k8s.io/updated_at=2023_08_30T23_04_36_0700
	                    minikube.k8s.io/version=v1.31.2
	                    node-role.kubernetes.io/master=
	Annotations:        kubeadm.alpha.kubernetes.io/cri-socket: /run/containerd/containerd.sock
	                    node.alpha.kubernetes.io/ttl: 0
	                    volumes.kubernetes.io/controller-managed-attach-detach: true
	CreationTimestamp:  Wed, 30 Aug 2023 23:04:32 +0000
	Taints:             <none>
	Unschedulable:      false
	Lease:
	  HolderIdentity:  ingress-addon-legacy-235029
	  AcquireTime:     <unset>
	  RenewTime:       Wed, 30 Aug 2023 23:06:09 +0000
	Conditions:
	  Type             Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message
	  ----             ------  -----------------                 ------------------                ------                       -------
	  MemoryPressure   False   Wed, 30 Aug 2023 23:06:09 +0000   Wed, 30 Aug 2023 23:04:25 +0000   KubeletHasSufficientMemory   kubelet has sufficient memory available
	  DiskPressure     False   Wed, 30 Aug 2023 23:06:09 +0000   Wed, 30 Aug 2023 23:04:25 +0000   KubeletHasNoDiskPressure     kubelet has no disk pressure
	  PIDPressure      False   Wed, 30 Aug 2023 23:06:09 +0000   Wed, 30 Aug 2023 23:04:25 +0000   KubeletHasSufficientPID      kubelet has sufficient PID available
	  Ready            True    Wed, 30 Aug 2023 23:06:09 +0000   Wed, 30 Aug 2023 23:04:49 +0000   KubeletReady                 kubelet is posting ready status
	Addresses:
	  InternalIP:  192.168.49.2
	  Hostname:    ingress-addon-legacy-235029
	Capacity:
	  cpu:                2
	  ephemeral-storage:  203034800Ki
	  hugepages-1Gi:      0
	  hugepages-2Mi:      0
	  hugepages-32Mi:     0
	  hugepages-64Ki:     0
	  memory:             8022572Ki
	  pods:               110
	Allocatable:
	  cpu:                2
	  ephemeral-storage:  203034800Ki
	  hugepages-1Gi:      0
	  hugepages-2Mi:      0
	  hugepages-32Mi:     0
	  hugepages-64Ki:     0
	  memory:             8022572Ki
	  pods:               110
	System Info:
	  Machine ID:                 49f6269204994d05952bd9c8b233ab83
	  System UUID:                dfa7e005-4267-4305-854a-2dc417e6f27b
	  Boot ID:                    98673563-8173-4281-afb4-eac1dfafdc23
	  Kernel Version:             5.15.0-1043-aws
	  OS Image:                   Ubuntu 22.04.3 LTS
	  Operating System:           linux
	  Architecture:               arm64
	  Container Runtime Version:  containerd://1.6.22
	  Kubelet Version:            v1.18.20
	  Kube-Proxy Version:         v1.18.20
	PodCIDR:                      10.244.0.0/24
	PodCIDRs:                     10.244.0.0/24
	Non-terminated Pods:          (10 in total)
	  Namespace                   Name                                                   CPU Requests  CPU Limits  Memory Requests  Memory Limits  AGE
	  ---------                   ----                                                   ------------  ----------  ---------------  -------------  ---
	  default                     hello-world-app-5f5d8b66bb-6szz4                       0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         35s
	  default                     nginx                                                  0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         44s
	  kube-system                 coredns-66bff467f8-j788v                               100m (5%!)(MISSING)     0 (0%!)(MISSING)      70Mi (0%!)(MISSING)        170Mi (2%!)(MISSING)     88s
	  kube-system                 etcd-ingress-addon-legacy-235029                       0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         99s
	  kube-system                 kindnet-4jl78                                          100m (5%!)(MISSING)     100m (5%!)(MISSING)   50Mi (0%!)(MISSING)        50Mi (0%!)(MISSING)      88s
	  kube-system                 kube-apiserver-ingress-addon-legacy-235029             250m (12%!)(MISSING)    0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         99s
	  kube-system                 kube-controller-manager-ingress-addon-legacy-235029    200m (10%!)(MISSING)    0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         99s
	  kube-system                 kube-proxy-2bhx8                                       0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         88s
	  kube-system                 kube-scheduler-ingress-addon-legacy-235029             100m (5%!)(MISSING)     0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         99s
	  kube-system                 storage-provisioner                                    0 (0%!)(MISSING)        0 (0%!)(MISSING)      0 (0%!)(MISSING)           0 (0%!)(MISSING)         87s
	Allocated resources:
	  (Total limits may be over 100 percent, i.e., overcommitted.)
	  Resource           Requests    Limits
	  --------           --------    ------
	  cpu                750m (37%!)(MISSING)  100m (5%!)(MISSING)
	  memory             120Mi (1%!)(MISSING)  220Mi (2%!)(MISSING)
	  ephemeral-storage  0 (0%!)(MISSING)      0 (0%!)(MISSING)
	  hugepages-1Gi      0 (0%!)(MISSING)      0 (0%!)(MISSING)
	  hugepages-2Mi      0 (0%!)(MISSING)      0 (0%!)(MISSING)
	  hugepages-32Mi     0 (0%!)(MISSING)      0 (0%!)(MISSING)
	  hugepages-64Ki     0 (0%!)(MISSING)      0 (0%!)(MISSING)
	Events:
	  Type    Reason                   Age                  From        Message
	  ----    ------                   ----                 ----        -------
	  Normal  NodeHasSufficientMemory  114s (x5 over 114s)  kubelet     Node ingress-addon-legacy-235029 status is now: NodeHasSufficientMemory
	  Normal  NodeHasNoDiskPressure    114s (x5 over 114s)  kubelet     Node ingress-addon-legacy-235029 status is now: NodeHasNoDiskPressure
	  Normal  NodeHasSufficientPID     114s (x5 over 114s)  kubelet     Node ingress-addon-legacy-235029 status is now: NodeHasSufficientPID
	  Normal  Starting                 99s                  kubelet     Starting kubelet.
	  Normal  NodeHasSufficientMemory  99s                  kubelet     Node ingress-addon-legacy-235029 status is now: NodeHasSufficientMemory
	  Normal  NodeHasNoDiskPressure    99s                  kubelet     Node ingress-addon-legacy-235029 status is now: NodeHasNoDiskPressure
	  Normal  NodeHasSufficientPID     99s                  kubelet     Node ingress-addon-legacy-235029 status is now: NodeHasSufficientPID
	  Normal  NodeAllocatableEnforced  99s                  kubelet     Updated Node Allocatable limit across pods
	  Normal  NodeReady                89s                  kubelet     Node ingress-addon-legacy-235029 status is now: NodeReady
	  Normal  Starting                 85s                  kube-proxy  Starting kube-proxy.
	
	* 
	* ==> dmesg <==
	* [  +0.001047] FS-Cache: O-key=[8] '6a415c0100000000'
	[  +0.000737] FS-Cache: N-cookie c=00000066 [p=0000005d fl=2 nc=0 na=1]
	[  +0.000924] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=00000000a41d18fb
	[  +0.001010] FS-Cache: N-key=[8] '6a415c0100000000'
	[  +0.002699] FS-Cache: Duplicate cookie detected
	[  +0.000740] FS-Cache: O-cookie c=00000060 [p=0000005d fl=226 nc=0 na=1]
	[  +0.000935] FS-Cache: O-cookie d=00000000d8a48a2b{9p.inode} n=000000009512bf07
	[  +0.001010] FS-Cache: O-key=[8] '6a415c0100000000'
	[  +0.000741] FS-Cache: N-cookie c=00000067 [p=0000005d fl=2 nc=0 na=1]
	[  +0.000916] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=00000000e0bc5e95
	[  +0.001060] FS-Cache: N-key=[8] '6a415c0100000000'
	[  +2.904353] FS-Cache: Duplicate cookie detected
	[  +0.000708] FS-Cache: O-cookie c=0000005e [p=0000005d fl=226 nc=0 na=1]
	[  +0.000962] FS-Cache: O-cookie d=00000000d8a48a2b{9p.inode} n=000000007bdd3326
	[  +0.001080] FS-Cache: O-key=[8] '69415c0100000000'
	[  +0.000675] FS-Cache: N-cookie c=00000069 [p=0000005d fl=2 nc=0 na=1]
	[  +0.000894] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=00000000a41d18fb
	[  +0.001068] FS-Cache: N-key=[8] '69415c0100000000'
	[  +0.462659] FS-Cache: Duplicate cookie detected
	[  +0.000688] FS-Cache: O-cookie c=00000063 [p=0000005d fl=226 nc=0 na=1]
	[  +0.000963] FS-Cache: O-cookie d=00000000d8a48a2b{9p.inode} n=00000000ec866464
	[  +0.001134] FS-Cache: O-key=[8] '6f415c0100000000'
	[  +0.000702] FS-Cache: N-cookie c=0000006a [p=0000005d fl=2 nc=0 na=1]
	[  +0.000917] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=000000001a64e3e4
	[  +0.001038] FS-Cache: N-key=[8] '6f415c0100000000'
	
	* 
	* ==> etcd [c2f9d005378fdd8eea90eaf14e17a4a30765cd674472cd14c8d79298cf976f67] <==
	* raft2023/08/30 23:04:26 INFO: aec36adc501070cc became follower at term 0
	raft2023/08/30 23:04:26 INFO: newRaft aec36adc501070cc [peers: [], term: 0, commit: 0, applied: 0, lastindex: 0, lastterm: 0]
	raft2023/08/30 23:04:26 INFO: aec36adc501070cc became follower at term 1
	raft2023/08/30 23:04:26 INFO: aec36adc501070cc switched to configuration voters=(12593026477526642892)
	2023-08-30 23:04:27.321012 W | auth: simple token is not cryptographically signed
	2023-08-30 23:04:27.597718 I | etcdserver: starting server... [version: 3.4.3, cluster version: to_be_decided]
	2023-08-30 23:04:27.713172 I | etcdserver: aec36adc501070cc as single-node; fast-forwarding 9 ticks (election ticks 10)
	raft2023/08/30 23:04:27 INFO: aec36adc501070cc switched to configuration voters=(12593026477526642892)
	2023-08-30 23:04:27.787584 I | etcdserver/membership: added member aec36adc501070cc [https://192.168.49.2:2380] to cluster fa54960ea34d58be
	2023-08-30 23:04:27.789803 I | embed: listening for peers on 192.168.49.2:2380
	2023-08-30 23:04:27.789971 I | embed: ClientTLS: cert = /var/lib/minikube/certs/etcd/server.crt, key = /var/lib/minikube/certs/etcd/server.key, trusted-ca = /var/lib/minikube/certs/etcd/ca.crt, client-cert-auth = true, crl-file = 
	2023-08-30 23:04:27.790319 I | embed: listening for metrics on http://127.0.0.1:2381
	raft2023/08/30 23:04:28 INFO: aec36adc501070cc is starting a new election at term 1
	raft2023/08/30 23:04:28 INFO: aec36adc501070cc became candidate at term 2
	raft2023/08/30 23:04:28 INFO: aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2
	raft2023/08/30 23:04:28 INFO: aec36adc501070cc became leader at term 2
	raft2023/08/30 23:04:28 INFO: raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2
	2023-08-30 23:04:28.498672 I | etcdserver: setting up the initial cluster version to 3.4
	2023-08-30 23:04:28.499318 N | etcdserver/membership: set the initial cluster version to 3.4
	2023-08-30 23:04:28.499463 I | etcdserver: published {Name:ingress-addon-legacy-235029 ClientURLs:[https://192.168.49.2:2379]} to cluster fa54960ea34d58be
	2023-08-30 23:04:28.500017 I | etcdserver/api: enabled capabilities for version 3.4
	2023-08-30 23:04:28.500740 I | embed: ready to serve client requests
	2023-08-30 23:04:28.501214 I | embed: ready to serve client requests
	2023-08-30 23:04:28.503059 I | embed: serving client requests on 192.168.49.2:2379
	2023-08-30 23:04:28.503483 I | embed: serving client requests on 127.0.0.1:2379
	
	* 
	* ==> kernel <==
	*  23:06:18 up  7:48,  0 users,  load average: 1.23, 1.75, 2.04
	Linux ingress-addon-legacy-235029 5.15.0-1043-aws #48~20.04.1-Ubuntu SMP Wed Aug 16 18:32:42 UTC 2023 aarch64 aarch64 aarch64 GNU/Linux
	PRETTY_NAME="Ubuntu 22.04.3 LTS"
	
	* 
	* ==> kindnet [f44fee6e9407e4261763301d99bd128a482aa4eae07e530bf4c9a38f8a8daad2] <==
	* I0830 23:04:53.182254       1 main.go:102] connected to apiserver: https://10.96.0.1:443
	I0830 23:04:53.182334       1 main.go:107] hostIP = 192.168.49.2
	podIP = 192.168.49.2
	I0830 23:04:53.182452       1 main.go:116] setting mtu 1500 for CNI 
	I0830 23:04:53.182467       1 main.go:146] kindnetd IP family: "ipv4"
	I0830 23:04:53.182478       1 main.go:150] noMask IPv4 subnets: [10.244.0.0/16]
	I0830 23:04:53.575875       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 23:04:53.576109       1 main.go:227] handling current node
	I0830 23:05:03.589422       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 23:05:03.589456       1 main.go:227] handling current node
	I0830 23:05:13.600609       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 23:05:13.600834       1 main.go:227] handling current node
	I0830 23:05:23.612658       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 23:05:23.612931       1 main.go:227] handling current node
	I0830 23:05:33.616638       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 23:05:33.616666       1 main.go:227] handling current node
	I0830 23:05:43.630899       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 23:05:43.630936       1 main.go:227] handling current node
	I0830 23:05:53.634798       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 23:05:53.634826       1 main.go:227] handling current node
	I0830 23:06:03.646608       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 23:06:03.646640       1 main.go:227] handling current node
	I0830 23:06:13.657943       1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
	I0830 23:06:13.657972       1 main.go:227] handling current node
	
	* 
	* ==> kube-apiserver [eebf8ed564e5c696f6b872e0831d7536e7baf714e1c9a30d5db2e538075a3b0b] <==
	* I0830 23:04:32.563056       1 dynamic_cafile_content.go:167] Starting request-header::/var/lib/minikube/certs/front-proxy-ca.crt
	E0830 23:04:32.588069       1 controller.go:152] Unable to remove old endpoints from kubernetes service: StorageError: key not found, Code: 1, Key: /registry/masterleases/192.168.49.2, ResourceVersion: 0, AdditionalErrorMsg: 
	I0830 23:04:32.734947       1 cache.go:39] Caches are synced for autoregister controller
	I0830 23:04:32.736078       1 shared_informer.go:230] Caches are synced for cluster_authentication_trust_controller 
	I0830 23:04:32.736431       1 cache.go:39] Caches are synced for AvailableConditionController controller
	I0830 23:04:32.738845       1 cache.go:39] Caches are synced for APIServiceRegistrationController controller
	I0830 23:04:32.769295       1 shared_informer.go:230] Caches are synced for crd-autoregister 
	I0830 23:04:33.530607       1 controller.go:130] OpenAPI AggregationController: action for item : Nothing (removed from the queue).
	I0830 23:04:33.530900       1 controller.go:130] OpenAPI AggregationController: action for item k8s_internal_local_delegation_chain_0000000000: Nothing (removed from the queue).
	I0830 23:04:33.537391       1 storage_scheduling.go:134] created PriorityClass system-node-critical with value 2000001000
	I0830 23:04:33.541883       1 storage_scheduling.go:134] created PriorityClass system-cluster-critical with value 2000000000
	I0830 23:04:33.542059       1 storage_scheduling.go:143] all system priority classes are created successfully or already exist.
	I0830 23:04:33.957978       1 controller.go:609] quota admission added evaluator for: roles.rbac.authorization.k8s.io
	I0830 23:04:33.993819       1 controller.go:609] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
	W0830 23:04:34.073858       1 lease.go:224] Resetting endpoints for master service "kubernetes" to [192.168.49.2]
	I0830 23:04:34.075044       1 controller.go:609] quota admission added evaluator for: endpoints
	I0830 23:04:34.080130       1 controller.go:609] quota admission added evaluator for: endpointslices.discovery.k8s.io
	I0830 23:04:34.991870       1 controller.go:609] quota admission added evaluator for: serviceaccounts
	I0830 23:04:35.797867       1 controller.go:609] quota admission added evaluator for: deployments.apps
	I0830 23:04:35.971525       1 controller.go:609] quota admission added evaluator for: daemonsets.apps
	I0830 23:04:39.222201       1 controller.go:609] quota admission added evaluator for: leases.coordination.k8s.io
	I0830 23:04:50.764131       1 controller.go:609] quota admission added evaluator for: replicasets.apps
	I0830 23:04:50.839822       1 controller.go:609] quota admission added evaluator for: controllerrevisions.apps
	I0830 23:05:14.477863       1 controller.go:609] quota admission added evaluator for: jobs.batch
	I0830 23:05:34.801058       1 controller.go:609] quota admission added evaluator for: ingresses.networking.k8s.io
	
	* 
	* ==> kube-controller-manager [5b6430073e5fce7b9aa2ecbdfddfe8420b20b564abdb562af732d61f2fa644f4] <==
	* I0830 23:04:50.837893       1 node_lifecycle_controller.go:1433] Initializing eviction metric for zone: 
	W0830 23:04:50.837981       1 node_lifecycle_controller.go:1048] Missing timestamp for Node ingress-addon-legacy-235029. Assuming now as a timestamp.
	I0830 23:04:50.838039       1 node_lifecycle_controller.go:1249] Controller detected that zone  is now in state Normal.
	I0830 23:04:50.838362       1 taint_manager.go:187] Starting NoExecuteTaintManager
	I0830 23:04:50.838758       1 event.go:278] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"ingress-addon-legacy-235029", UID:"5e6c5e5b-1372-4c4a-b117-2242e60b0b9b", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RegisteredNode' Node ingress-addon-legacy-235029 event: Registered Node ingress-addon-legacy-235029 in Controller
	I0830 23:04:50.843613       1 shared_informer.go:230] Caches are synced for HPA 
	I0830 23:04:50.843702       1 shared_informer.go:230] Caches are synced for job 
	I0830 23:04:50.881925       1 event.go:278] Event(v1.ObjectReference{Kind:"DaemonSet", Namespace:"kube-system", Name:"kindnet", UID:"6072dafc-f8e6-40fd-9aa6-2eb5a4648e22", APIVersion:"apps/v1", ResourceVersion:"231", FieldPath:""}): type: 'Normal' reason: 'SuccessfulCreate' Created pod: kindnet-4jl78
	I0830 23:04:50.881953       1 event.go:278] Event(v1.ObjectReference{Kind:"DaemonSet", Namespace:"kube-system", Name:"kube-proxy", UID:"f680dfa7-d601-4ffd-af91-90d4cd008843", APIVersion:"apps/v1", ResourceVersion:"218", FieldPath:""}): type: 'Normal' reason: 'SuccessfulCreate' Created pod: kube-proxy-2bhx8
	I0830 23:04:50.885196       1 shared_informer.go:230] Caches are synced for endpoint 
	I0830 23:04:50.887990       1 shared_informer.go:230] Caches are synced for bootstrap_signer 
	I0830 23:04:50.893528       1 shared_informer.go:230] Caches are synced for resource quota 
	I0830 23:04:50.943160       1 shared_informer.go:230] Caches are synced for resource quota 
	I0830 23:04:50.949807       1 shared_informer.go:230] Caches are synced for garbage collector 
	I0830 23:04:50.949829       1 garbagecollector.go:142] Garbage collector: all resource monitors have synced. Proceeding to collect garbage
	I0830 23:04:50.950659       1 shared_informer.go:230] Caches are synced for garbage collector 
	E0830 23:04:50.965614       1 daemon_controller.go:321] kube-system/kube-proxy failed with : error storing status for daemon set &v1.DaemonSet{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"kube-proxy", GenerateName:"", Namespace:"kube-system", SelfLink:"/apis/apps/v1/namespaces/kube-system/daemonsets/kube-proxy", UID:"f680dfa7-d601-4ffd-af91-90d4cd008843", ResourceVersion:"218", Generation:1, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63829033475, loc:(*time.Location)(0x6307ca0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-proxy"}, Annotations:map[string]string{"deprecated.daemonset.template.generation":"1"}, OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:"kubeadm", Operation:"Update", APIVersion:"apps/v1", Time:(*v1.Time)(0x4000bee0a0), FieldsType:"FieldsV1", FieldsV1:(*v1.Fields
V1)(0x4000bee0c0)}}}, Spec:v1.DaemonSetSpec{Selector:(*v1.LabelSelector)(0x4000bee0e0), Template:v1.PodTemplateSpec{ObjectMeta:v1.ObjectMeta{Name:"", GenerateName:"", Namespace:"", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-proxy"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"kube-proxy", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(n
il), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(0x4000675580), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}, v1.Volume{Name:"xtables-lock", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSou
rce)(0x4000bee100), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.Pr
ojectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}, v1.Volume{Name:"lib-modules", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(0x4000bee140), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolum
eSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}}, InitContainers:[]v1.Container(nil), Containers:[]v1.Container{v1.Container{Name:"kube-proxy", Image:"k8s.gcr.io/kube-proxy:v1.18.20", Command:[]string{"/usr/local/bin/kube-proxy", "--config=/var/lib/kube-proxy/config.conf", "--hostname-override=$(NODE_NAME)"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar{v1.EnvVar{Name:"NODE_NAME", Value:"", ValueFrom:(*v1.EnvVarSource)(0x4000bee180)}}, Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList
(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-proxy", ReadOnly:false, MountPath:"/var/lib/kube-proxy", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}, v1.VolumeMount{Name:"xtables-lock", ReadOnly:false, MountPath:"/run/xtables.lock", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}, v1.VolumeMount{Name:"lib-modules", ReadOnly:true, MountPath:"/lib/modules", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(0x40001fe230), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0x400091b5e8), Acti
veDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string{"kubernetes.io/os":"linux"}, ServiceAccountName:"kube-proxy", DeprecatedServiceAccount:"kube-proxy", AutomountServiceAccountToken:(*bool)(nil), NodeName:"", HostNetwork:true, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0x4000966310), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"CriticalAddonsOnly", Operator:"Exists", Value:"", Effect:"", TolerationSeconds:(*int64)(nil)}, v1.Toleration{Key:"", Operator:"Exists", Value:"", Effect:"", TolerationSeconds:(*int64)(nil)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"system-node-critical", Priority:(*int32)(nil), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(nil), PreemptionPoli
cy:(*v1.PreemptionPolicy)(nil), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil)}}, UpdateStrategy:v1.DaemonSetUpdateStrategy{Type:"RollingUpdate", RollingUpdate:(*v1.RollingUpdateDaemonSet)(0x400000eed8)}, MinReadySeconds:0, RevisionHistoryLimit:(*int32)(0x400091b638)}, Status:v1.DaemonSetStatus{CurrentNumberScheduled:0, NumberMisscheduled:0, DesiredNumberScheduled:0, NumberReady:0, ObservedGeneration:0, UpdatedNumberScheduled:0, NumberAvailable:0, NumberUnavailable:0, CollisionCount:(*int32)(nil), Conditions:[]v1.DaemonSetCondition(nil)}}: Operation cannot be fulfilled on daemonsets.apps "kube-proxy": the object has been modified; please apply your changes to the latest version and try again
	I0830 23:05:14.439780       1 event.go:278] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"ingress-nginx", Name:"ingress-nginx-controller", UID:"20f735f1-30e4-4731-871d-40083e26a863", APIVersion:"apps/v1", ResourceVersion:"461", FieldPath:""}): type: 'Normal' reason: 'ScalingReplicaSet' Scaled up replica set ingress-nginx-controller-7fcf777cb7 to 1
	I0830 23:05:14.479589       1 event.go:278] Event(v1.ObjectReference{Kind:"ReplicaSet", Namespace:"ingress-nginx", Name:"ingress-nginx-controller-7fcf777cb7", UID:"0adf6506-8d4d-4bb5-8ec3-d1c8410f9532", APIVersion:"apps/v1", ResourceVersion:"462", FieldPath:""}): type: 'Normal' reason: 'SuccessfulCreate' Created pod: ingress-nginx-controller-7fcf777cb7-jfv5q
	I0830 23:05:14.495445       1 event.go:278] Event(v1.ObjectReference{Kind:"Job", Namespace:"ingress-nginx", Name:"ingress-nginx-admission-create", UID:"67acb774-82a5-4fa8-a3cf-f775e400e3ba", APIVersion:"batch/v1", ResourceVersion:"466", FieldPath:""}): type: 'Normal' reason: 'SuccessfulCreate' Created pod: ingress-nginx-admission-create-jq4xk
	I0830 23:05:14.554154       1 event.go:278] Event(v1.ObjectReference{Kind:"Job", Namespace:"ingress-nginx", Name:"ingress-nginx-admission-patch", UID:"8b9382b9-1ea2-4c60-bb6e-020cf256a4a5", APIVersion:"batch/v1", ResourceVersion:"480", FieldPath:""}): type: 'Normal' reason: 'SuccessfulCreate' Created pod: ingress-nginx-admission-patch-77kp5
	I0830 23:05:17.395206       1 event.go:278] Event(v1.ObjectReference{Kind:"Job", Namespace:"ingress-nginx", Name:"ingress-nginx-admission-patch", UID:"8b9382b9-1ea2-4c60-bb6e-020cf256a4a5", APIVersion:"batch/v1", ResourceVersion:"488", FieldPath:""}): type: 'Normal' reason: 'Completed' Job completed
	I0830 23:05:17.436331       1 event.go:278] Event(v1.ObjectReference{Kind:"Job", Namespace:"ingress-nginx", Name:"ingress-nginx-admission-create", UID:"67acb774-82a5-4fa8-a3cf-f775e400e3ba", APIVersion:"batch/v1", ResourceVersion:"481", FieldPath:""}): type: 'Normal' reason: 'Completed' Job completed
	I0830 23:05:43.679980       1 event.go:278] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"default", Name:"hello-world-app", UID:"35df3b73-d985-407b-89f3-bcf401aea214", APIVersion:"apps/v1", ResourceVersion:"598", FieldPath:""}): type: 'Normal' reason: 'ScalingReplicaSet' Scaled up replica set hello-world-app-5f5d8b66bb to 1
	I0830 23:05:43.689394       1 event.go:278] Event(v1.ObjectReference{Kind:"ReplicaSet", Namespace:"default", Name:"hello-world-app-5f5d8b66bb", UID:"44d9d8b8-7365-4484-862f-e153e951e104", APIVersion:"apps/v1", ResourceVersion:"599", FieldPath:""}): type: 'Normal' reason: 'SuccessfulCreate' Created pod: hello-world-app-5f5d8b66bb-6szz4
	
	* 
	* ==> kube-proxy [d6417a8a2e5ee02d9a5271d260c75068b36269da053e03f17867f58918814d64] <==
	* W0830 23:04:53.139028       1 server_others.go:559] Unknown proxy mode "", assuming iptables proxy
	I0830 23:04:53.152688       1 node.go:136] Successfully retrieved node IP: 192.168.49.2
	I0830 23:04:53.152861       1 server_others.go:186] Using iptables Proxier.
	I0830 23:04:53.154903       1 server.go:583] Version: v1.18.20
	I0830 23:04:53.156702       1 config.go:315] Starting service config controller
	I0830 23:04:53.156883       1 shared_informer.go:223] Waiting for caches to sync for service config
	I0830 23:04:53.157028       1 config.go:133] Starting endpoints config controller
	I0830 23:04:53.157104       1 shared_informer.go:223] Waiting for caches to sync for endpoints config
	I0830 23:04:53.257193       1 shared_informer.go:230] Caches are synced for service config 
	I0830 23:04:53.257344       1 shared_informer.go:230] Caches are synced for endpoints config 
	
	* 
	* ==> kube-scheduler [3620e41650f7f7b24b04bde3e7f9d455910bf31fb10198e7e0d20a52d7e4dd43] <==
	* I0830 23:04:32.735564       1 registry.go:150] Registering EvenPodsSpread predicate and priority function
	I0830 23:04:32.738997       1 secure_serving.go:178] Serving securely on 127.0.0.1:10259
	I0830 23:04:32.739329       1 configmap_cafile_content.go:202] Starting client-ca::kube-system::extension-apiserver-authentication::client-ca-file
	I0830 23:04:32.739506       1 shared_informer.go:223] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
	I0830 23:04:32.739643       1 tlsconfig.go:240] Starting DynamicServingCertificateController
	E0830 23:04:32.744679       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
	E0830 23:04:32.744976       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
	E0830 23:04:32.745660       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
	E0830 23:04:32.745823       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
	E0830 23:04:32.745936       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
	E0830 23:04:32.746079       1 reflector.go:178] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:233: Failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
	E0830 23:04:32.746174       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
	E0830 23:04:32.746280       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
	E0830 23:04:32.746434       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1beta1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
	E0830 23:04:32.747551       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
	E0830 23:04:32.747815       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
	E0830 23:04:32.748491       1 reflector.go:178] k8s.io/apiserver/pkg/server/dynamiccertificates/configmap_cafile_content.go:206: Failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
	E0830 23:04:33.575337       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
	E0830 23:04:33.580008       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
	E0830 23:04:33.599483       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
	E0830 23:04:33.626400       1 reflector.go:178] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:233: Failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
	E0830 23:04:33.661044       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
	E0830 23:04:33.692506       1 reflector.go:178] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
	I0830 23:04:34.339741       1 shared_informer.go:230] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file 
	E0830 23:04:50.876467       1 factory.go:503] pod kube-system/coredns-66bff467f8-j788v is already present in the backoff queue
	
	* 
	* ==> kubelet <==
	* Aug 30 23:05:47 ingress-addon-legacy-235029 kubelet[1665]: E0830 23:05:47.532166    1665 pod_workers.go:191] Error syncing pod e6683f12-8586-46fd-a703-635f7edb93ee ("hello-world-app-5f5d8b66bb-6szz4_default(e6683f12-8586-46fd-a703-635f7edb93ee)"), skipping: failed to "StartContainer" for "hello-world-app" with CrashLoopBackOff: "back-off 10s restarting failed container=hello-world-app pod=hello-world-app-5f5d8b66bb-6szz4_default(e6683f12-8586-46fd-a703-635f7edb93ee)"
	Aug 30 23:05:48 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:05:48.534537    1665 topology_manager.go:221] [topologymanager] RemoveContainer - Container ID: a3e4e6c60da2d708d5b80be43672bfaa03e08e8ad104355c832c42f20aac27a7
	Aug 30 23:05:48 ingress-addon-legacy-235029 kubelet[1665]: E0830 23:05:48.534794    1665 pod_workers.go:191] Error syncing pod e6683f12-8586-46fd-a703-635f7edb93ee ("hello-world-app-5f5d8b66bb-6szz4_default(e6683f12-8586-46fd-a703-635f7edb93ee)"), skipping: failed to "StartContainer" for "hello-world-app" with CrashLoopBackOff: "back-off 10s restarting failed container=hello-world-app pod=hello-world-app-5f5d8b66bb-6szz4_default(e6683f12-8586-46fd-a703-635f7edb93ee)"
	Aug 30 23:05:57 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:05:57.266191    1665 topology_manager.go:221] [topologymanager] RemoveContainer - Container ID: 126bc6c72d31035907a3402b7ab5b8e832901ce36a5efd712d4dd1898aef0a38
	Aug 30 23:05:57 ingress-addon-legacy-235029 kubelet[1665]: E0830 23:05:57.266585    1665 pod_workers.go:191] Error syncing pod 57b05f7f-3468-42fe-9639-573649e437a2 ("kube-ingress-dns-minikube_kube-system(57b05f7f-3468-42fe-9639-573649e437a2)"), skipping: failed to "StartContainer" for "minikube-ingress-dns" with CrashLoopBackOff: "back-off 20s restarting failed container=minikube-ingress-dns pod=kube-ingress-dns-minikube_kube-system(57b05f7f-3468-42fe-9639-573649e437a2)"
	Aug 30 23:05:59 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:05:59.267722    1665 topology_manager.go:221] [topologymanager] RemoveContainer - Container ID: a3e4e6c60da2d708d5b80be43672bfaa03e08e8ad104355c832c42f20aac27a7
	Aug 30 23:05:59 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:05:59.555464    1665 topology_manager.go:221] [topologymanager] RemoveContainer - Container ID: a3e4e6c60da2d708d5b80be43672bfaa03e08e8ad104355c832c42f20aac27a7
	Aug 30 23:05:59 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:05:59.555831    1665 topology_manager.go:221] [topologymanager] RemoveContainer - Container ID: fb36ffc4bab1f759a5e18600a90c49e92aa56372e81e85aab287d58330754b01
	Aug 30 23:05:59 ingress-addon-legacy-235029 kubelet[1665]: E0830 23:05:59.556076    1665 pod_workers.go:191] Error syncing pod e6683f12-8586-46fd-a703-635f7edb93ee ("hello-world-app-5f5d8b66bb-6szz4_default(e6683f12-8586-46fd-a703-635f7edb93ee)"), skipping: failed to "StartContainer" for "hello-world-app" with CrashLoopBackOff: "back-off 20s restarting failed container=hello-world-app pod=hello-world-app-5f5d8b66bb-6szz4_default(e6683f12-8586-46fd-a703-635f7edb93ee)"
	Aug 30 23:05:59 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:05:59.689956    1665 reconciler.go:196] operationExecutor.UnmountVolume started for volume "minikube-ingress-dns-token-r9flz" (UniqueName: "kubernetes.io/secret/57b05f7f-3468-42fe-9639-573649e437a2-minikube-ingress-dns-token-r9flz") pod "57b05f7f-3468-42fe-9639-573649e437a2" (UID: "57b05f7f-3468-42fe-9639-573649e437a2")
	Aug 30 23:05:59 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:05:59.694081    1665 operation_generator.go:782] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b05f7f-3468-42fe-9639-573649e437a2-minikube-ingress-dns-token-r9flz" (OuterVolumeSpecName: "minikube-ingress-dns-token-r9flz") pod "57b05f7f-3468-42fe-9639-573649e437a2" (UID: "57b05f7f-3468-42fe-9639-573649e437a2"). InnerVolumeSpecName "minikube-ingress-dns-token-r9flz". PluginName "kubernetes.io/secret", VolumeGidValue ""
	Aug 30 23:05:59 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:05:59.790349    1665 reconciler.go:319] Volume detached for volume "minikube-ingress-dns-token-r9flz" (UniqueName: "kubernetes.io/secret/57b05f7f-3468-42fe-9639-573649e437a2-minikube-ingress-dns-token-r9flz") on node "ingress-addon-legacy-235029" DevicePath ""
	Aug 30 23:06:01 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:06:01.563047    1665 topology_manager.go:221] [topologymanager] RemoveContainer - Container ID: 126bc6c72d31035907a3402b7ab5b8e832901ce36a5efd712d4dd1898aef0a38
	Aug 30 23:06:10 ingress-addon-legacy-235029 kubelet[1665]: E0830 23:06:10.227449    1665 event.go:260] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"ingress-nginx-controller-7fcf777cb7-jfv5q.17804bc532a00cc3", GenerateName:"", Namespace:"ingress-nginx", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Pod", Namespace:"ingress-nginx", Name:"ingress-nginx-controller-7fcf777cb7-jfv5q", UID:"ab07db14-8cbf-40fb-8a69-a9b946ad39bf", APIVersion:"v1", ResourceVersion:"468", FieldPath:"spec.containers{controller}"}, Reason:"Killing", Message:"Stoppi
ng container controller", Source:v1.EventSource{Component:"kubelet", Host:"ingress-addon-legacy-235029"}, FirstTimestamp:v1.Time{Time:time.Time{wall:0xc13410b88d30b8c3, ext:94476613281, loc:(*time.Location)(0x6a0ef20)}}, LastTimestamp:v1.Time{Time:time.Time{wall:0xc13410b88d30b8c3, ext:94476613281, loc:(*time.Location)(0x6a0ef20)}}, Count:1, Type:"Normal", EventTime:v1.MicroTime{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "ingress-nginx-controller-7fcf777cb7-jfv5q.17804bc532a00cc3" is forbidden: unable to create new content in namespace ingress-nginx because it is being terminated' (will not retry!)
	Aug 30 23:06:10 ingress-addon-legacy-235029 kubelet[1665]: E0830 23:06:10.250420    1665 event.go:260] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"ingress-nginx-controller-7fcf777cb7-jfv5q.17804bc532a00cc3", GenerateName:"", Namespace:"ingress-nginx", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Pod", Namespace:"ingress-nginx", Name:"ingress-nginx-controller-7fcf777cb7-jfv5q", UID:"ab07db14-8cbf-40fb-8a69-a9b946ad39bf", APIVersion:"v1", ResourceVersion:"468", FieldPath:"spec.containers{controller}"}, Reason:"Killing", Message:"Stoppi
ng container controller", Source:v1.EventSource{Component:"kubelet", Host:"ingress-addon-legacy-235029"}, FirstTimestamp:v1.Time{Time:time.Time{wall:0xc13410b88d30b8c3, ext:94476613281, loc:(*time.Location)(0x6a0ef20)}}, LastTimestamp:v1.Time{Time:time.Time{wall:0xc13410b88e00c438, ext:94490247703, loc:(*time.Location)(0x6a0ef20)}}, Count:2, Type:"Normal", EventTime:v1.MicroTime{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "ingress-nginx-controller-7fcf777cb7-jfv5q.17804bc532a00cc3" is forbidden: unable to create new content in namespace ingress-nginx because it is being terminated' (will not retry!)
	Aug 30 23:06:12 ingress-addon-legacy-235029 kubelet[1665]: W0830 23:06:12.586593    1665 pod_container_deletor.go:77] Container "0b5712dee820ff57f1a04b857fa0cec24d6a9327b055097a5ea8426387aa2acb" not found in pod's containers
	Aug 30 23:06:13 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:06:13.266636    1665 topology_manager.go:221] [topologymanager] RemoveContainer - Container ID: fb36ffc4bab1f759a5e18600a90c49e92aa56372e81e85aab287d58330754b01
	Aug 30 23:06:13 ingress-addon-legacy-235029 kubelet[1665]: E0830 23:06:13.266991    1665 pod_workers.go:191] Error syncing pod e6683f12-8586-46fd-a703-635f7edb93ee ("hello-world-app-5f5d8b66bb-6szz4_default(e6683f12-8586-46fd-a703-635f7edb93ee)"), skipping: failed to "StartContainer" for "hello-world-app" with CrashLoopBackOff: "back-off 20s restarting failed container=hello-world-app pod=hello-world-app-5f5d8b66bb-6szz4_default(e6683f12-8586-46fd-a703-635f7edb93ee)"
	Aug 30 23:06:14 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:06:14.331632    1665 reconciler.go:196] operationExecutor.UnmountVolume started for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/ab07db14-8cbf-40fb-8a69-a9b946ad39bf-webhook-cert") pod "ab07db14-8cbf-40fb-8a69-a9b946ad39bf" (UID: "ab07db14-8cbf-40fb-8a69-a9b946ad39bf")
	Aug 30 23:06:14 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:06:14.333771    1665 reconciler.go:196] operationExecutor.UnmountVolume started for volume "ingress-nginx-token-jh4v6" (UniqueName: "kubernetes.io/secret/ab07db14-8cbf-40fb-8a69-a9b946ad39bf-ingress-nginx-token-jh4v6") pod "ab07db14-8cbf-40fb-8a69-a9b946ad39bf" (UID: "ab07db14-8cbf-40fb-8a69-a9b946ad39bf")
	Aug 30 23:06:14 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:06:14.338507    1665 operation_generator.go:782] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab07db14-8cbf-40fb-8a69-a9b946ad39bf-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "ab07db14-8cbf-40fb-8a69-a9b946ad39bf" (UID: "ab07db14-8cbf-40fb-8a69-a9b946ad39bf"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
	Aug 30 23:06:14 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:06:14.340200    1665 operation_generator.go:782] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab07db14-8cbf-40fb-8a69-a9b946ad39bf-ingress-nginx-token-jh4v6" (OuterVolumeSpecName: "ingress-nginx-token-jh4v6") pod "ab07db14-8cbf-40fb-8a69-a9b946ad39bf" (UID: "ab07db14-8cbf-40fb-8a69-a9b946ad39bf"). InnerVolumeSpecName "ingress-nginx-token-jh4v6". PluginName "kubernetes.io/secret", VolumeGidValue ""
	Aug 30 23:06:14 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:06:14.434348    1665 reconciler.go:319] Volume detached for volume "ingress-nginx-token-jh4v6" (UniqueName: "kubernetes.io/secret/ab07db14-8cbf-40fb-8a69-a9b946ad39bf-ingress-nginx-token-jh4v6") on node "ingress-addon-legacy-235029" DevicePath ""
	Aug 30 23:06:14 ingress-addon-legacy-235029 kubelet[1665]: I0830 23:06:14.434399    1665 reconciler.go:319] Volume detached for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/ab07db14-8cbf-40fb-8a69-a9b946ad39bf-webhook-cert") on node "ingress-addon-legacy-235029" DevicePath ""
	Aug 30 23:06:15 ingress-addon-legacy-235029 kubelet[1665]: W0830 23:06:15.271369    1665 kubelet_getters.go:297] Path "/var/lib/kubelet/pods/ab07db14-8cbf-40fb-8a69-a9b946ad39bf/volumes" does not exist
	
	* 
	* ==> storage-provisioner [04bb01342154960c9ffb503e39cb764de902683e3a2b6662e6473a3566c08427] <==
	* I0830 23:04:54.377690       1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
	I0830 23:04:54.390239       1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
	I0830 23:04:54.390965       1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
	I0830 23:04:54.398679       1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
	I0830 23:04:54.399071       1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_ingress-addon-legacy-235029_fb5fef1d-cb61-4f9c-a159-0b8a70d0e2c6!
	I0830 23:04:54.400552       1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"72e37890-f973-48c4-82f5-58843685d7d2", APIVersion:"v1", ResourceVersion:"401", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' ingress-addon-legacy-235029_fb5fef1d-cb61-4f9c-a159-0b8a70d0e2c6 became leader
	I0830 23:04:54.499590       1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_ingress-addon-legacy-235029_fb5fef1d-cb61-4f9c-a159-0b8a70d0e2c6!
	

                                                
                                                
-- /stdout --
helpers_test.go:254: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p ingress-addon-legacy-235029 -n ingress-addon-legacy-235029
helpers_test.go:261: (dbg) Run:  kubectl --context ingress-addon-legacy-235029 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:285: <<< TestIngressAddonLegacy/serial/ValidateIngressAddons FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestIngressAddonLegacy/serial/ValidateIngressAddons (54.97s)

                                                
                                    
x
+
TestMissingContainerUpgrade (182.93s)

                                                
                                                
=== RUN   TestMissingContainerUpgrade
=== PAUSE TestMissingContainerUpgrade

                                                
                                                

                                                
                                                
=== CONT  TestMissingContainerUpgrade
version_upgrade_test.go:321: (dbg) Run:  /tmp/minikube-v1.22.0.1348801557.exe start -p missing-upgrade-821485 --memory=2200 --driver=docker  --container-runtime=containerd
version_upgrade_test.go:321: (dbg) Done: /tmp/minikube-v1.22.0.1348801557.exe start -p missing-upgrade-821485 --memory=2200 --driver=docker  --container-runtime=containerd: (1m31.267551633s)
version_upgrade_test.go:330: (dbg) Run:  docker stop missing-upgrade-821485
version_upgrade_test.go:330: (dbg) Done: docker stop missing-upgrade-821485: (11.835461754s)
version_upgrade_test.go:335: (dbg) Run:  docker rm missing-upgrade-821485
version_upgrade_test.go:341: (dbg) Run:  out/minikube-linux-arm64 start -p missing-upgrade-821485 --memory=2200 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd
E0830 23:32:06.148308 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:32:44.061957 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
version_upgrade_test.go:341: (dbg) Non-zero exit: out/minikube-linux-arm64 start -p missing-upgrade-821485 --memory=2200 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd: exit status 90 (1m14.061515447s)

                                                
                                                
-- stdout --
	* [missing-upgrade-821485] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	  - MINIKUBE_LOCATION=17114
	  - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	  - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	  - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	  - MINIKUBE_BIN=out/minikube-linux-arm64
	  - MINIKUBE_FORCE_SYSTEMD=
	* Kubernetes 1.28.1 is now available. If you would like to upgrade, specify: --kubernetes-version=v1.28.1
	* Using the docker driver based on existing profile
	* Starting control plane node missing-upgrade-821485 in cluster missing-upgrade-821485
	* Pulling base image ...
	* docker "missing-upgrade-821485" container is missing, will recreate.
	* Creating docker container (CPUs=2, Memory=2200MB) ...
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	I0830 23:31:52.736720 1365243 out.go:296] Setting OutFile to fd 1 ...
	I0830 23:31:52.736917 1365243 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:31:52.736945 1365243 out.go:309] Setting ErrFile to fd 2...
	I0830 23:31:52.736963 1365243 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:31:52.737301 1365243 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
	I0830 23:31:52.737795 1365243 out.go:303] Setting JSON to false
	I0830 23:31:52.738879 1365243 start.go:128] hostinfo: {"hostname":"ip-172-31-31-251","uptime":29647,"bootTime":1693408666,"procs":285,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1043-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"982e3628-3742-4b3e-bb63-ac1b07660ec7"}
	I0830 23:31:52.738978 1365243 start.go:138] virtualization:  
	I0830 23:31:52.743745 1365243 out.go:177] * [missing-upgrade-821485] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	I0830 23:31:52.745838 1365243 out.go:177]   - MINIKUBE_LOCATION=17114
	I0830 23:31:52.745928 1365243 notify.go:220] Checking for updates...
	I0830 23:31:52.750460 1365243 out.go:177]   - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0830 23:31:52.752093 1365243 out.go:177]   - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	I0830 23:31:52.754204 1365243 out.go:177]   - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	I0830 23:31:52.755949 1365243 out.go:177]   - MINIKUBE_BIN=out/minikube-linux-arm64
	I0830 23:31:52.757634 1365243 out.go:177]   - MINIKUBE_FORCE_SYSTEMD=
	I0830 23:31:52.759777 1365243 config.go:182] Loaded profile config "missing-upgrade-821485": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.21.2
	I0830 23:31:52.761874 1365243 out.go:177] * Kubernetes 1.28.1 is now available. If you would like to upgrade, specify: --kubernetes-version=v1.28.1
	I0830 23:31:52.763498 1365243 driver.go:373] Setting default libvirt URI to qemu:///system
	I0830 23:31:52.800333 1365243 docker.go:121] docker version: linux-24.0.5:Docker Engine - Community
	I0830 23:31:52.800434 1365243 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 23:31:52.950364 1365243 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:1 ContainersRunning:1 ContainersPaused:0 ContainersStopped:0 Images:4 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:33 OomKillDisable:true NGoroutines:45 SystemTime:2023-08-30 23:31:52.939781488 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 23:31:52.950466 1365243 docker.go:294] overlay module found
	I0830 23:31:52.953092 1365243 out.go:177] * Using the docker driver based on existing profile
	I0830 23:31:52.955081 1365243 start.go:298] selected driver: docker
	I0830 23:31:52.955104 1365243 start.go:902] validating driver "docker" against &{Name:missing-upgrade-821485 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.99.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.21.2 ClusterName:missing-upgrade-821485 Namespace:default APIServerName:minikubeCA APIServer
Names:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[{Component:kubelet Key:cni-conf-dir Value:/etc/cni/net.mk}] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP:192.168.85.2 Port:8443 KubernetesVersion:v1.21.2 ContainerRuntime: ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString: Mount9PVersion: MountGID: MountIP: MountMSize:0 MountOptions:[] MountPort:0 MountType: MountUID: BinaryMirror: DisableOptimizations:false DisableMetrics:false
CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0}
	I0830 23:31:52.955244 1365243 start.go:913] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
	I0830 23:31:52.955877 1365243 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 23:31:53.077077 1365243 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:1 ContainersRunning:1 ContainersPaused:0 ContainersStopped:0 Images:4 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:33 OomKillDisable:true NGoroutines:45 SystemTime:2023-08-30 23:31:53.066847445 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 23:31:53.077478 1365243 cni.go:84] Creating CNI manager for ""
	I0830 23:31:53.077495 1365243 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
	I0830 23:31:53.077507 1365243 start_flags.go:319] config:
	{Name:missing-upgrade-821485 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.99.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.21.2 ClusterName:missing-upgrade-821485 Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRI
Socket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[{Component:kubelet Key:cni-conf-dir Value:/etc/cni/net.mk}] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP:192.168.85.2 Port:8443 KubernetesVersion:v1.21.2 ContainerRuntime: ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString: Mount9PVersion: MountGID: MountIP: MountMSize:0 MountOptions:[] MountPort:0 MountType: MountUID: BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAut
hSock: SSHAgentPID:0}
	I0830 23:31:53.080582 1365243 out.go:177] * Starting control plane node missing-upgrade-821485 in cluster missing-upgrade-821485
	I0830 23:31:53.082724 1365243 cache.go:122] Beginning downloading kic base image for docker with containerd
	I0830 23:31:53.084455 1365243 out.go:177] * Pulling base image ...
	I0830 23:31:53.085998 1365243 preload.go:132] Checking if preload exists for k8s version v1.21.2 and runtime containerd
	I0830 23:31:53.086057 1365243 preload.go:148] Found local preload: /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.21.2-containerd-overlay2-arm64.tar.lz4
	I0830 23:31:53.086071 1365243 cache.go:57] Caching tarball of preloaded images
	I0830 23:31:53.086159 1365243 preload.go:174] Found /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.21.2-containerd-overlay2-arm64.tar.lz4 in cache, skipping download
	I0830 23:31:53.086172 1365243 cache.go:60] Finished verifying existence of preloaded tar for  v1.21.2 on containerd
	I0830 23:31:53.086291 1365243 profile.go:148] Saving config to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/missing-upgrade-821485/config.json ...
	I0830 23:31:53.086508 1365243 image.go:79] Checking for gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 in local docker daemon
	I0830 23:31:53.105672 1365243 image.go:83] Found gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 in local docker daemon, skipping pull
	I0830 23:31:53.105700 1365243 cache.go:145] gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 exists in daemon, skipping load
	I0830 23:31:53.105720 1365243 cache.go:195] Successfully downloaded all kic artifacts
	I0830 23:31:53.105774 1365243 start.go:365] acquiring machines lock for missing-upgrade-821485: {Name:mk0aadb9f10896eff5c72577cf25f5b348680b76 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
	I0830 23:31:53.105847 1365243 start.go:369] acquired machines lock for "missing-upgrade-821485" in 42.091µs
	I0830 23:31:53.105871 1365243 start.go:96] Skipping create...Using existing machine configuration
	I0830 23:31:53.105880 1365243 fix.go:54] fixHost starting: 
	I0830 23:31:53.106163 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:53.124361 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:53.124421 1365243 fix.go:102] recreateIfNeeded on missing-upgrade-821485: state= err=unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:53.124441 1365243 fix.go:107] machineExists: false. err=machine does not exist
	I0830 23:31:53.126587 1365243 out.go:177] * docker "missing-upgrade-821485" container is missing, will recreate.
	I0830 23:31:53.128014 1365243 delete.go:124] DEMOLISHING missing-upgrade-821485 ...
	I0830 23:31:53.128116 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:53.151439 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	W0830 23:31:53.151512 1365243 stop.go:75] unable to get state: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:53.151550 1365243 delete.go:128] stophost failed (probably ok): ssh power off: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:53.152007 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:53.182602 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:53.182671 1365243 delete.go:82] Unable to get host status for missing-upgrade-821485, assuming it has already been deleted: state: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:53.182754 1365243 cli_runner.go:164] Run: docker container inspect -f {{.Id}} missing-upgrade-821485
	W0830 23:31:53.204567 1365243 cli_runner.go:211] docker container inspect -f {{.Id}} missing-upgrade-821485 returned with exit code 1
	I0830 23:31:53.204623 1365243 kic.go:367] could not find the container missing-upgrade-821485 to remove it. will try anyways
	I0830 23:31:53.204684 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:53.227796 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	W0830 23:31:53.227859 1365243 oci.go:84] error getting container status, will try to delete anyways: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:53.227927 1365243 cli_runner.go:164] Run: docker exec --privileged -t missing-upgrade-821485 /bin/bash -c "sudo init 0"
	W0830 23:31:53.258405 1365243 cli_runner.go:211] docker exec --privileged -t missing-upgrade-821485 /bin/bash -c "sudo init 0" returned with exit code 1
	I0830 23:31:53.258437 1365243 oci.go:647] error shutdown missing-upgrade-821485: docker exec --privileged -t missing-upgrade-821485 /bin/bash -c "sudo init 0": exit status 1
	stdout:
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:54.258611 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:54.276706 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:54.276789 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:54.276804 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:31:54.276835 1365243 retry.go:31] will retry after 444.223191ms: couldn't verify container is exited. %v: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:54.721269 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:54.738607 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:54.738672 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:54.738683 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:31:54.738712 1365243 retry.go:31] will retry after 634.565723ms: couldn't verify container is exited. %v: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:55.374312 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:55.395127 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:55.395184 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:55.395192 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:31:55.395214 1365243 retry.go:31] will retry after 1.256461515s: couldn't verify container is exited. %v: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:56.652170 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:56.670416 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:56.670479 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:56.670499 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:31:56.670522 1365243 retry.go:31] will retry after 2.366757653s: couldn't verify container is exited. %v: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:59.038293 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:59.056315 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:59.056382 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:59.056393 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:31:59.056418 1365243 retry.go:31] will retry after 3.625191599s: couldn't verify container is exited. %v: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:32:02.684782 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:32:02.701864 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:32:02.701932 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:32:02.701953 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:32:02.701977 1365243 retry.go:31] will retry after 5.483696825s: couldn't verify container is exited. %v: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:32:08.186602 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:32:08.213640 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:32:08.213701 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:32:08.213711 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:32:08.213738 1365243 retry.go:31] will retry after 3.196654656s: couldn't verify container is exited. %v: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:32:11.411526 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:32:11.437112 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:32:11.437210 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:32:11.437223 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:32:11.437281 1365243 oci.go:88] couldn't shut down missing-upgrade-821485 (might be okay): verify shutdown: couldn't verify container is exited. %v: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	 
	I0830 23:32:11.437365 1365243 cli_runner.go:164] Run: docker rm -f -v missing-upgrade-821485
	I0830 23:32:11.455349 1365243 cli_runner.go:164] Run: docker container inspect -f {{.Id}} missing-upgrade-821485
	W0830 23:32:11.472097 1365243 cli_runner.go:211] docker container inspect -f {{.Id}} missing-upgrade-821485 returned with exit code 1
	I0830 23:32:11.472198 1365243 cli_runner.go:164] Run: docker network inspect missing-upgrade-821485 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	I0830 23:32:11.490425 1365243 cli_runner.go:164] Run: docker network rm missing-upgrade-821485
	I0830 23:32:11.587805 1365243 fix.go:114] Sleeping 1 second for extra luck!
	I0830 23:32:12.587901 1365243 start.go:125] createHost starting for "" (driver="docker")
	I0830 23:32:12.590258 1365243 out.go:204] * Creating docker container (CPUs=2, Memory=2200MB) ...
	I0830 23:32:12.590389 1365243 start.go:159] libmachine.API.Create for "missing-upgrade-821485" (driver="docker")
	I0830 23:32:12.590424 1365243 client.go:168] LocalClient.Create starting
	I0830 23:32:12.590500 1365243 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem
	I0830 23:32:12.590554 1365243 main.go:141] libmachine: Decoding PEM data...
	I0830 23:32:12.590579 1365243 main.go:141] libmachine: Parsing certificate...
	I0830 23:32:12.590648 1365243 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem
	I0830 23:32:12.590683 1365243 main.go:141] libmachine: Decoding PEM data...
	I0830 23:32:12.590700 1365243 main.go:141] libmachine: Parsing certificate...
	I0830 23:32:12.591006 1365243 cli_runner.go:164] Run: docker network inspect missing-upgrade-821485 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	W0830 23:32:12.608394 1365243 cli_runner.go:211] docker network inspect missing-upgrade-821485 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
	I0830 23:32:12.608486 1365243 network_create.go:281] running [docker network inspect missing-upgrade-821485] to gather additional debugging logs...
	I0830 23:32:12.608508 1365243 cli_runner.go:164] Run: docker network inspect missing-upgrade-821485
	W0830 23:32:12.626064 1365243 cli_runner.go:211] docker network inspect missing-upgrade-821485 returned with exit code 1
	I0830 23:32:12.626104 1365243 network_create.go:284] error running [docker network inspect missing-upgrade-821485]: docker network inspect missing-upgrade-821485: exit status 1
	stdout:
	[]
	
	stderr:
	Error response from daemon: network missing-upgrade-821485 not found
	I0830 23:32:12.626135 1365243 network_create.go:286] output of [docker network inspect missing-upgrade-821485]: -- stdout --
	[]
	
	-- /stdout --
	** stderr ** 
	Error response from daemon: network missing-upgrade-821485 not found
	
	** /stderr **
	I0830 23:32:12.626200 1365243 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	I0830 23:32:12.644369 1365243 network.go:214] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-ed2d417202d5 IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:02:42:e9:35:f3:fe} reservation:<nil>}
	I0830 23:32:12.644763 1365243 network.go:214] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-b83085549c6c IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:02:42:58:7c:4f:97} reservation:<nil>}
	I0830 23:32:12.645253 1365243 network.go:214] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-4f518efc6da3 IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:02:42:89:2a:5e:24} reservation:<nil>}
	I0830 23:32:12.645663 1365243 network.go:214] skipping subnet 192.168.76.0/24 that is taken: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName:br-28e919bf1baa IfaceIPv4:192.168.76.1 IfaceMTU:1500 IfaceMAC:02:42:6e:68:af:9c} reservation:<nil>}
	I0830 23:32:12.646602 1365243 network.go:209] using free private subnet 192.168.85.0/24: &{IP:192.168.85.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.85.0/24 Gateway:192.168.85.1 ClientMin:192.168.85.2 ClientMax:192.168.85.254 Broadcast:192.168.85.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x4002a23d10}
	I0830 23:32:12.646627 1365243 network_create.go:123] attempt to create docker network missing-upgrade-821485 192.168.85.0/24 with gateway 192.168.85.1 and MTU of 1500 ...
	I0830 23:32:12.646693 1365243 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.85.0/24 --gateway=192.168.85.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=missing-upgrade-821485 missing-upgrade-821485
	I0830 23:32:12.719796 1365243 network_create.go:107] docker network missing-upgrade-821485 192.168.85.0/24 created
	I0830 23:32:12.719826 1365243 kic.go:117] calculated static IP "192.168.85.2" for the "missing-upgrade-821485" container
	I0830 23:32:12.719924 1365243 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
	I0830 23:32:12.737249 1365243 cli_runner.go:164] Run: docker volume create missing-upgrade-821485 --label name.minikube.sigs.k8s.io=missing-upgrade-821485 --label created_by.minikube.sigs.k8s.io=true
	I0830 23:32:12.756029 1365243 oci.go:103] Successfully created a docker volume missing-upgrade-821485
	I0830 23:32:12.756112 1365243 cli_runner.go:164] Run: docker run --rm --name missing-upgrade-821485-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=missing-upgrade-821485 --entrypoint /usr/bin/test -v missing-upgrade-821485:/var gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 -d /var/lib
	I0830 23:32:13.286154 1365243 oci.go:107] Successfully prepared a docker volume missing-upgrade-821485
	I0830 23:32:13.286210 1365243 preload.go:132] Checking if preload exists for k8s version v1.21.2 and runtime containerd
	I0830 23:32:13.286229 1365243 kic.go:190] Starting extracting preloaded images to volume ...
	I0830 23:32:13.286317 1365243 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.21.2-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v missing-upgrade-821485:/extractDir gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 -I lz4 -xf /preloaded.tar -C /extractDir
	I0830 23:32:18.094107 1365243 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.21.2-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v missing-upgrade-821485:/extractDir gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 -I lz4 -xf /preloaded.tar -C /extractDir: (4.807748151s)
	I0830 23:32:18.094142 1365243 kic.go:199] duration metric: took 4.807906 seconds to extract preloaded images to volume
	W0830 23:32:18.094300 1365243 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
	I0830 23:32:18.094409 1365243 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
	I0830 23:32:18.167517 1365243 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname missing-upgrade-821485 --name missing-upgrade-821485 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=missing-upgrade-821485 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=missing-upgrade-821485 --network missing-upgrade-821485 --ip 192.168.85.2 --volume missing-upgrade-821485:/var --security-opt apparmor=unconfined --memory=2200mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79
	I0830 23:32:18.557413 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Running}}
	I0830 23:32:18.579407 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	I0830 23:32:18.605656 1365243 cli_runner.go:164] Run: docker exec missing-upgrade-821485 stat /var/lib/dpkg/alternatives/iptables
	I0830 23:32:18.724768 1365243 oci.go:144] the created container "missing-upgrade-821485" has a running status.
	I0830 23:32:18.724798 1365243 kic.go:221] Creating ssh key for kic: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa...
	I0830 23:32:19.345344 1365243 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
	I0830 23:32:19.384250 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	I0830 23:32:19.409937 1365243 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
	I0830 23:32:19.409986 1365243 kic_runner.go:114] Args: [docker exec --privileged missing-upgrade-821485 chown docker:docker /home/docker/.ssh/authorized_keys]
	I0830 23:32:19.506973 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	I0830 23:32:19.536945 1365243 machine.go:88] provisioning docker machine ...
	I0830 23:32:19.536975 1365243 ubuntu.go:169] provisioning hostname "missing-upgrade-821485"
	I0830 23:32:19.537037 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:19.558760 1365243 main.go:141] libmachine: Using SSH client type: native
	I0830 23:32:19.559238 1365243 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34529 <nil> <nil>}
	I0830 23:32:19.559253 1365243 main.go:141] libmachine: About to run SSH command:
	sudo hostname missing-upgrade-821485 && echo "missing-upgrade-821485" | sudo tee /etc/hostname
	I0830 23:32:19.721990 1365243 main.go:141] libmachine: SSH cmd err, output: <nil>: missing-upgrade-821485
	
	I0830 23:32:19.722110 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:19.756584 1365243 main.go:141] libmachine: Using SSH client type: native
	I0830 23:32:19.757070 1365243 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34529 <nil> <nil>}
	I0830 23:32:19.757094 1365243 main.go:141] libmachine: About to run SSH command:
	
			if ! grep -xq '.*\smissing-upgrade-821485' /etc/hosts; then
				if grep -xq '127.0.1.1\s.*' /etc/hosts; then
					sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 missing-upgrade-821485/g' /etc/hosts;
				else 
					echo '127.0.1.1 missing-upgrade-821485' | sudo tee -a /etc/hosts; 
				fi
			fi
	I0830 23:32:19.915950 1365243 main.go:141] libmachine: SSH cmd err, output: <nil>: 
	I0830 23:32:19.916023 1365243 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/17114-1219981/.minikube CaCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/17114-1219981/.minikube}
	I0830 23:32:19.916095 1365243 ubuntu.go:177] setting up certificates
	I0830 23:32:19.916122 1365243 provision.go:83] configureAuth start
	I0830 23:32:19.916220 1365243 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" missing-upgrade-821485
	I0830 23:32:19.948839 1365243 provision.go:138] copyHostCerts
	I0830 23:32:19.948912 1365243 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem, removing ...
	I0830 23:32:19.948921 1365243 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem
	I0830 23:32:19.948996 1365243 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem (1082 bytes)
	I0830 23:32:19.949094 1365243 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem, removing ...
	I0830 23:32:19.949100 1365243 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem
	I0830 23:32:19.949166 1365243 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem (1123 bytes)
	I0830 23:32:19.949236 1365243 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem, removing ...
	I0830 23:32:19.949240 1365243 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem
	I0830 23:32:19.949266 1365243 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem (1679 bytes)
	I0830 23:32:19.949324 1365243 provision.go:112] generating server cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem org=jenkins.missing-upgrade-821485 san=[192.168.85.2 127.0.0.1 localhost 127.0.0.1 minikube missing-upgrade-821485]
	I0830 23:32:20.910285 1365243 provision.go:172] copyRemoteCerts
	I0830 23:32:20.910397 1365243 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
	I0830 23:32:20.910475 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:20.933385 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:21.027946 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
	I0830 23:32:21.054533 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem --> /etc/docker/server.pem (1241 bytes)
	I0830 23:32:21.082998 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
	I0830 23:32:21.110396 1365243 provision.go:86] duration metric: configureAuth took 1.194238702s
	I0830 23:32:21.110433 1365243 ubuntu.go:193] setting minikube options for container-runtime
	I0830 23:32:21.110630 1365243 config.go:182] Loaded profile config "missing-upgrade-821485": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.21.2
	I0830 23:32:21.110646 1365243 machine.go:91] provisioned docker machine in 1.573683801s
	I0830 23:32:21.110653 1365243 client.go:171] LocalClient.Create took 8.520220852s
	I0830 23:32:21.110671 1365243 start.go:167] duration metric: libmachine.API.Create for "missing-upgrade-821485" took 8.520280172s
	I0830 23:32:21.110682 1365243 start.go:300] post-start starting for "missing-upgrade-821485" (driver="docker")
	I0830 23:32:21.110692 1365243 start.go:329] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
	I0830 23:32:21.110750 1365243 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
	I0830 23:32:21.110794 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:21.138599 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:21.232506 1365243 ssh_runner.go:195] Run: cat /etc/os-release
	I0830 23:32:21.237238 1365243 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
	I0830 23:32:21.237265 1365243 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
	I0830 23:32:21.237277 1365243 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
	I0830 23:32:21.237289 1365243 info.go:137] Remote host: Ubuntu 20.04.2 LTS
	I0830 23:32:21.237299 1365243 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/addons for local assets ...
	I0830 23:32:21.237358 1365243 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/files for local assets ...
	I0830 23:32:21.237443 1365243 filesync.go:149] local asset: /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem -> 12253752.pem in /etc/ssl/certs
	I0830 23:32:21.237546 1365243 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
	I0830 23:32:21.248159 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem --> /etc/ssl/certs/12253752.pem (1708 bytes)
	I0830 23:32:21.282644 1365243 start.go:303] post-start completed in 171.942358ms
	I0830 23:32:21.283090 1365243 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" missing-upgrade-821485
	I0830 23:32:21.306835 1365243 profile.go:148] Saving config to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/missing-upgrade-821485/config.json ...
	I0830 23:32:21.307129 1365243 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0830 23:32:21.307181 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:21.328360 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:21.420052 1365243 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
	I0830 23:32:21.426005 1365243 start.go:128] duration metric: createHost completed in 8.838066552s
	I0830 23:32:21.426100 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:32:21.467765 1365243 fix.go:128] unexpected machine state, will restart: <nil>
	I0830 23:32:21.467788 1365243 machine.go:88] provisioning docker machine ...
	I0830 23:32:21.467805 1365243 ubuntu.go:169] provisioning hostname "missing-upgrade-821485"
	I0830 23:32:21.467870 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:21.490981 1365243 main.go:141] libmachine: Using SSH client type: native
	I0830 23:32:21.491443 1365243 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34529 <nil> <nil>}
	I0830 23:32:21.491456 1365243 main.go:141] libmachine: About to run SSH command:
	sudo hostname missing-upgrade-821485 && echo "missing-upgrade-821485" | sudo tee /etc/hostname
	I0830 23:32:21.630064 1365243 main.go:141] libmachine: SSH cmd err, output: <nil>: missing-upgrade-821485
	
	I0830 23:32:21.630146 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:21.656678 1365243 main.go:141] libmachine: Using SSH client type: native
	I0830 23:32:21.657116 1365243 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34529 <nil> <nil>}
	I0830 23:32:21.657172 1365243 main.go:141] libmachine: About to run SSH command:
	
			if ! grep -xq '.*\smissing-upgrade-821485' /etc/hosts; then
				if grep -xq '127.0.1.1\s.*' /etc/hosts; then
					sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 missing-upgrade-821485/g' /etc/hosts;
				else 
					echo '127.0.1.1 missing-upgrade-821485' | sudo tee -a /etc/hosts; 
				fi
			fi
	I0830 23:32:21.794803 1365243 main.go:141] libmachine: SSH cmd err, output: <nil>: 
	I0830 23:32:21.794830 1365243 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/17114-1219981/.minikube CaCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/17114-1219981/.minikube}
	I0830 23:32:21.794856 1365243 ubuntu.go:177] setting up certificates
	I0830 23:32:21.794870 1365243 provision.go:83] configureAuth start
	I0830 23:32:21.794935 1365243 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" missing-upgrade-821485
	I0830 23:32:21.818851 1365243 provision.go:138] copyHostCerts
	I0830 23:32:21.818915 1365243 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem, removing ...
	I0830 23:32:21.818935 1365243 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem
	I0830 23:32:21.819014 1365243 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem (1082 bytes)
	I0830 23:32:21.819123 1365243 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem, removing ...
	I0830 23:32:21.819133 1365243 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem
	I0830 23:32:21.819160 1365243 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem (1123 bytes)
	I0830 23:32:21.819220 1365243 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem, removing ...
	I0830 23:32:21.819229 1365243 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem
	I0830 23:32:21.819253 1365243 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem (1679 bytes)
	I0830 23:32:21.819336 1365243 provision.go:112] generating server cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem org=jenkins.missing-upgrade-821485 san=[192.168.85.2 127.0.0.1 localhost 127.0.0.1 minikube missing-upgrade-821485]
	I0830 23:32:22.102738 1365243 provision.go:172] copyRemoteCerts
	I0830 23:32:22.102806 1365243 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
	I0830 23:32:22.102853 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:22.121794 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:22.218890 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
	I0830 23:32:22.244766 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem --> /etc/docker/server.pem (1241 bytes)
	I0830 23:32:22.271374 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
	I0830 23:32:22.304362 1365243 provision.go:86] duration metric: configureAuth took 509.477014ms
	I0830 23:32:22.304389 1365243 ubuntu.go:193] setting minikube options for container-runtime
	I0830 23:32:22.304572 1365243 config.go:182] Loaded profile config "missing-upgrade-821485": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.21.2
	I0830 23:32:22.304586 1365243 machine.go:91] provisioned docker machine in 836.791852ms
	I0830 23:32:22.304593 1365243 start.go:300] post-start starting for "missing-upgrade-821485" (driver="docker")
	I0830 23:32:22.304602 1365243 start.go:329] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
	I0830 23:32:22.304653 1365243 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
	I0830 23:32:22.304695 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:22.327682 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:22.433578 1365243 ssh_runner.go:195] Run: cat /etc/os-release
	I0830 23:32:22.437960 1365243 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
	I0830 23:32:22.437989 1365243 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
	I0830 23:32:22.438002 1365243 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
	I0830 23:32:22.438010 1365243 info.go:137] Remote host: Ubuntu 20.04.2 LTS
	I0830 23:32:22.438020 1365243 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/addons for local assets ...
	I0830 23:32:22.438085 1365243 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/files for local assets ...
	I0830 23:32:22.438161 1365243 filesync.go:149] local asset: /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem -> 12253752.pem in /etc/ssl/certs
	I0830 23:32:22.438283 1365243 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
	I0830 23:32:22.448726 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem --> /etc/ssl/certs/12253752.pem (1708 bytes)
	I0830 23:32:22.480544 1365243 start.go:303] post-start completed in 175.936024ms
	I0830 23:32:22.480629 1365243 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0830 23:32:22.480690 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:22.507869 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:22.600238 1365243 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
	I0830 23:32:22.607154 1365243 fix.go:56] fixHost completed within 29.501258817s
	I0830 23:32:22.607175 1365243 start.go:83] releasing machines lock for "missing-upgrade-821485", held for 29.501315669s
	I0830 23:32:22.607244 1365243 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" missing-upgrade-821485
	I0830 23:32:22.626574 1365243 ssh_runner.go:195] Run: cat /version.json
	I0830 23:32:22.626625 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:22.626838 1365243 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
	I0830 23:32:22.626893 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:22.659612 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:22.664621 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	W0830 23:32:22.878282 1365243 start.go:419] Unable to open version.json: cat /version.json: Process exited with status 1
	stdout:
	
	stderr:
	cat: /version.json: No such file or directory
	I0830 23:32:22.878363 1365243 ssh_runner.go:195] Run: systemctl --version
	I0830 23:32:22.883858 1365243 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
	I0830 23:32:22.889584 1365243 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
	I0830 23:32:22.918673 1365243 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
	I0830 23:32:22.918815 1365243 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
	I0830 23:32:22.953301 1365243 cni.go:262] disabled [/etc/cni/net.d/100-crio-bridge.conf, /etc/cni/net.d/87-podman-bridge.conflist] bridge cni config(s)
	I0830 23:32:22.953326 1365243 start.go:466] detecting cgroup driver to use...
	I0830 23:32:22.953359 1365243 detect.go:196] detected "cgroupfs" cgroup driver on host os
	I0830 23:32:22.953412 1365243 ssh_runner.go:195] Run: sudo systemctl stop -f crio
	I0830 23:32:22.968880 1365243 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
	I0830 23:32:22.990880 1365243 docker.go:196] disabling cri-docker service (if available) ...
	I0830 23:32:22.990956 1365243 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
	I0830 23:32:23.003929 1365243 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
	I0830 23:32:23.017050 1365243 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
	W0830 23:32:23.031089 1365243 docker.go:206] Failed to disable socket "cri-docker.socket" (might be ok): sudo systemctl disable cri-docker.socket: Process exited with status 1
	stdout:
	
	stderr:
	Failed to disable unit: Unit file cri-docker.socket does not exist.
	I0830 23:32:23.031165 1365243 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
	I0830 23:32:23.135493 1365243 docker.go:212] disabling docker service ...
	I0830 23:32:23.135586 1365243 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
	I0830 23:32:23.159972 1365243 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
	I0830 23:32:23.173750 1365243 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
	I0830 23:32:23.276554 1365243 ssh_runner.go:195] Run: sudo systemctl mask docker.service
	I0830 23:32:23.384058 1365243 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
	I0830 23:32:23.396213 1365243 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
	" | sudo tee /etc/crictl.yaml"
	I0830 23:32:23.417196 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.4.1"|' /etc/containerd/config.toml"
	I0830 23:32:23.429390 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
	I0830 23:32:23.440691 1365243 containerd.go:145] configuring containerd to use "cgroupfs" as cgroup driver...
	I0830 23:32:23.440765 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
	I0830 23:32:23.452302 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
	I0830 23:32:23.463576 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
	I0830 23:32:23.475151 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
	I0830 23:32:23.486792 1365243 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
	I0830 23:32:23.497280 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
	I0830 23:32:23.508419 1365243 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
	I0830 23:32:23.518597 1365243 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
	I0830 23:32:23.528050 1365243 ssh_runner.go:195] Run: sudo systemctl daemon-reload
	I0830 23:32:23.626597 1365243 ssh_runner.go:195] Run: sudo systemctl restart containerd
	I0830 23:32:23.710815 1365243 start.go:513] Will wait 60s for socket path /run/containerd/containerd.sock
	I0830 23:32:23.710916 1365243 ssh_runner.go:195] Run: stat /run/containerd/containerd.sock
	I0830 23:32:23.715841 1365243 start.go:534] Will wait 60s for crictl version
	I0830 23:32:23.715905 1365243 ssh_runner.go:195] Run: which crictl
	I0830 23:32:23.720572 1365243 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
	I0830 23:32:23.754580 1365243 retry.go:31] will retry after 13.407237425s: Temporary Error: sudo /usr/bin/crictl version: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:32:23Z" level=fatal msg="getting the runtime version: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	I0830 23:32:37.166241 1365243 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
	I0830 23:32:37.196024 1365243 retry.go:31] will retry after 10.077638903s: Temporary Error: sudo /usr/bin/crictl version: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:32:37Z" level=fatal msg="getting the runtime version: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	I0830 23:32:47.273919 1365243 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
	I0830 23:32:47.306054 1365243 retry.go:31] will retry after 19.362516092s: Temporary Error: sudo /usr/bin/crictl version: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:32:47Z" level=fatal msg="getting the runtime version: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	I0830 23:33:06.669856 1365243 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
	I0830 23:33:06.702391 1365243 out.go:177] 
	W0830 23:33:06.704639 1365243 out.go:239] X Exiting due to RUNTIME_ENABLE: Failed to start container runtime: Temporary Error: sudo /usr/bin/crictl version: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:33:06Z" level=fatal msg="getting the runtime version: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	
	X Exiting due to RUNTIME_ENABLE: Failed to start container runtime: Temporary Error: sudo /usr/bin/crictl version: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:33:06Z" level=fatal msg="getting the runtime version: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	
	W0830 23:33:06.704654 1365243 out.go:239] * 
	* 
	W0830 23:33:06.705678 1365243 out.go:239] ╭─────────────────────────────────────────────────────────────────────────────────────────────╮
	│                                                                                             │
	│    * If the above advice does not help, please let us know:                                 │
	│      https://github.com/kubernetes/minikube/issues/new/choose                               │
	│                                                                                             │
	│    * Please run `minikube logs --file=logs.txt` and attach logs.txt to the GitHub issue.    │
	│                                                                                             │
	╰─────────────────────────────────────────────────────────────────────────────────────────────╯
	╭─────────────────────────────────────────────────────────────────────────────────────────────╮
	│                                                                                             │
	│    * If the above advice does not help, please let us know:                                 │
	│      https://github.com/kubernetes/minikube/issues/new/choose                               │
	│                                                                                             │
	│    * Please run `minikube logs --file=logs.txt` and attach logs.txt to the GitHub issue.    │
	│                                                                                             │
	╰─────────────────────────────────────────────────────────────────────────────────────────────╯
	I0830 23:33:06.708159 1365243 out.go:177] 

                                                
                                                
** /stderr **
version_upgrade_test.go:343: failed missing container upgrade from v1.22.0. args: out/minikube-linux-arm64 start -p missing-upgrade-821485 --memory=2200 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd : exit status 90
version_upgrade_test.go:345: *** TestMissingContainerUpgrade FAILED at 2023-08-30 23:33:06.752785697 +0000 UTC m=+2324.025716177
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======>  post-mortem[TestMissingContainerUpgrade]: docker inspect <======
helpers_test.go:231: (dbg) Run:  docker inspect missing-upgrade-821485
helpers_test.go:235: (dbg) docker inspect missing-upgrade-821485:

                                                
                                                
-- stdout --
	[
	    {
	        "Id": "6e3e92b6bf1ea9c42f982d403ca7da87649f079e5171d2463cad2a2f9216dd5f",
	        "Created": "2023-08-30T23:32:18.186237381Z",
	        "Path": "/usr/local/bin/entrypoint",
	        "Args": [
	            "/sbin/init"
	        ],
	        "State": {
	            "Status": "running",
	            "Running": true,
	            "Paused": false,
	            "Restarting": false,
	            "OOMKilled": false,
	            "Dead": false,
	            "Pid": 1366007,
	            "ExitCode": 0,
	            "Error": "",
	            "StartedAt": "2023-08-30T23:32:18.547224626Z",
	            "FinishedAt": "0001-01-01T00:00:00Z"
	        },
	        "Image": "sha256:ba5ae658d5b3f017bdb597cc46a1912d5eed54239e31b777788d204fdcbc4445",
	        "ResolvConfPath": "/var/lib/docker/containers/6e3e92b6bf1ea9c42f982d403ca7da87649f079e5171d2463cad2a2f9216dd5f/resolv.conf",
	        "HostnamePath": "/var/lib/docker/containers/6e3e92b6bf1ea9c42f982d403ca7da87649f079e5171d2463cad2a2f9216dd5f/hostname",
	        "HostsPath": "/var/lib/docker/containers/6e3e92b6bf1ea9c42f982d403ca7da87649f079e5171d2463cad2a2f9216dd5f/hosts",
	        "LogPath": "/var/lib/docker/containers/6e3e92b6bf1ea9c42f982d403ca7da87649f079e5171d2463cad2a2f9216dd5f/6e3e92b6bf1ea9c42f982d403ca7da87649f079e5171d2463cad2a2f9216dd5f-json.log",
	        "Name": "/missing-upgrade-821485",
	        "RestartCount": 0,
	        "Driver": "overlay2",
	        "Platform": "linux",
	        "MountLabel": "",
	        "ProcessLabel": "",
	        "AppArmorProfile": "unconfined",
	        "ExecIDs": null,
	        "HostConfig": {
	            "Binds": [
	                "/lib/modules:/lib/modules:ro",
	                "missing-upgrade-821485:/var"
	            ],
	            "ContainerIDFile": "",
	            "LogConfig": {
	                "Type": "json-file",
	                "Config": {}
	            },
	            "NetworkMode": "missing-upgrade-821485",
	            "PortBindings": {
	                "22/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "2376/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "32443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "5000/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "8443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ]
	            },
	            "RestartPolicy": {
	                "Name": "no",
	                "MaximumRetryCount": 0
	            },
	            "AutoRemove": false,
	            "VolumeDriver": "",
	            "VolumesFrom": null,
	            "ConsoleSize": [
	                0,
	                0
	            ],
	            "CapAdd": null,
	            "CapDrop": null,
	            "CgroupnsMode": "host",
	            "Dns": [],
	            "DnsOptions": [],
	            "DnsSearch": [],
	            "ExtraHosts": null,
	            "GroupAdd": null,
	            "IpcMode": "private",
	            "Cgroup": "",
	            "Links": null,
	            "OomScoreAdj": 0,
	            "PidMode": "",
	            "Privileged": true,
	            "PublishAllPorts": false,
	            "ReadonlyRootfs": false,
	            "SecurityOpt": [
	                "seccomp=unconfined",
	                "apparmor=unconfined",
	                "label=disable"
	            ],
	            "Tmpfs": {
	                "/run": "",
	                "/tmp": ""
	            },
	            "UTSMode": "",
	            "UsernsMode": "",
	            "ShmSize": 67108864,
	            "Runtime": "runc",
	            "Isolation": "",
	            "CpuShares": 0,
	            "Memory": 2306867200,
	            "NanoCpus": 2000000000,
	            "CgroupParent": "",
	            "BlkioWeight": 0,
	            "BlkioWeightDevice": [],
	            "BlkioDeviceReadBps": [],
	            "BlkioDeviceWriteBps": [],
	            "BlkioDeviceReadIOps": [],
	            "BlkioDeviceWriteIOps": [],
	            "CpuPeriod": 0,
	            "CpuQuota": 0,
	            "CpuRealtimePeriod": 0,
	            "CpuRealtimeRuntime": 0,
	            "CpusetCpus": "",
	            "CpusetMems": "",
	            "Devices": [],
	            "DeviceCgroupRules": null,
	            "DeviceRequests": null,
	            "MemoryReservation": 0,
	            "MemorySwap": 4613734400,
	            "MemorySwappiness": null,
	            "OomKillDisable": false,
	            "PidsLimit": null,
	            "Ulimits": null,
	            "CpuCount": 0,
	            "CpuPercent": 0,
	            "IOMaximumIOps": 0,
	            "IOMaximumBandwidth": 0,
	            "MaskedPaths": null,
	            "ReadonlyPaths": null
	        },
	        "GraphDriver": {
	            "Data": {
	                "LowerDir": "/var/lib/docker/overlay2/30a4b00c6c6318444e63a8bcdc1e78ceb808e0cc6ebe9feef5eaedb7b4ad468e-init/diff:/var/lib/docker/overlay2/6d37b79fc7350992a664000a2a91dd4411b139986a1cb0bf1c2e401164e2c8be/diff:/var/lib/docker/overlay2/312dd9b8593b82e0614fbb7b08c6d9288bd6b2c5963f8f713408e0297d5a1f75/diff:/var/lib/docker/overlay2/db680246e81f0fec575bbc0dfceae1645349322979230a2871db551415e7e283/diff:/var/lib/docker/overlay2/0e4f0266dede49a77fdb93a95fb3241cf015a6ade9f43269f5c2ffae9dd82f61/diff:/var/lib/docker/overlay2/6df76639fe736f3954ad729d60de4a582a317e3f51b8509d4e1697562c2668ef/diff:/var/lib/docker/overlay2/03017b302be08b49014a9cba53b02b74eb6b221755f540253eaa329b4b504764/diff:/var/lib/docker/overlay2/062c6b4678fe91c975670997c2b62d263dc482b06dbfea228f76b7e156b53ae3/diff:/var/lib/docker/overlay2/6d6faeba5098ea84ce25ed71dd4e23fa1b5980a00686f3539557ef6b73f685b4/diff:/var/lib/docker/overlay2/19d0c0180f367c7afa972dc96db3b7a037ba4fd40327e02d363f7a8540f7bed2/diff:/var/lib/docker/overlay2/bb50c6
ee59e7c02ae620708205df1f52667a312c2463a12e1d1bedc2e6371ee2/diff:/var/lib/docker/overlay2/2396d65122ac403fcf5a21095cf87a17d6825ae25d07ca3cb92bb1e155899545/diff:/var/lib/docker/overlay2/d6a62f5b20374bb02f467594feec31c42cc733f0da2f652519900fb497dc57df/diff:/var/lib/docker/overlay2/6d8082c1969ef764159e7cced934ac56922a46a9a6584c8214ff60a8504cc53d/diff:/var/lib/docker/overlay2/4a64b8c3cc812a8ecf005b155a93b3fdf230cd7cc5bf8c519a70d918f76f8183/diff:/var/lib/docker/overlay2/63d4f3fb47f6b2d59a30078b119a8c3640708eab4f81f15ab132f184e6523f6d/diff:/var/lib/docker/overlay2/0b8db9c016df03f6ce9f06a711fd78a1fcd2e7a69693d09acdf8ada73d313589/diff:/var/lib/docker/overlay2/c3d691b187be5566b22d449743241dbfb5b08aad08958f5b8a11bd2b60da4a59/diff:/var/lib/docker/overlay2/e20eb1cdbd5c398f8607aeac16d489f8bf77f3fe36acbc55fdcc4553bcc607a4/diff:/var/lib/docker/overlay2/c84da5186cfacdc6d0038ff332391c06b39584b802cc79ca9269c10b91fa2379/diff:/var/lib/docker/overlay2/e9ee59ec6960de2977fc73525e2c244d7dc2a934f64b5c21c8b9daf2a68e1e09/diff:/var/lib/d
ocker/overlay2/c5af1af545ef1a6d1485e2bc4bf6ab8bed86f6bafc40e97b1875b3b4ae60d9e5/diff:/var/lib/docker/overlay2/2af8c901b6fada41e6065488fab4f7756b2a2abc3d3e2f085ef993d5a60dc254/diff:/var/lib/docker/overlay2/484a2f82ed4f77d456f2a865814b13b5e205c77276ec56030c573b5c4ea3b2d1/diff:/var/lib/docker/overlay2/46d09892cf67a5b5325b5bc3b9cb69e8049c1d4a9ec44c584550ef105888b57e/diff:/var/lib/docker/overlay2/355acb00a79febfff77ea4445383b400388ff220316630a8743090a63f79a9b7/diff:/var/lib/docker/overlay2/8575844fdacf6b8b0fce5cfcbbc7900eed7a980960e57d4334f843c003854dea/diff:/var/lib/docker/overlay2/ecf530bda3fc3976c7fe3951062b761c889f74523948e4804ad7d568df21c5b0/diff:/var/lib/docker/overlay2/e4aa40dc2b793ac9287ab4b469a2b2014b1900371e1ee0c2f0c39e886c9831fd/diff:/var/lib/docker/overlay2/43b8d00a1677ebcc006199977630a93d0b28bf36f3bb5c47fc57cebb54a044ae/diff:/var/lib/docker/overlay2/7c1862cbe5141f1bb884c6f26a64ae7c872ed6f034d2cfc4f8ace5826f73b4ac/diff:/var/lib/docker/overlay2/63c40279f4daa425c835f8cf3f748bc3b1ad9bd62008e67fe6986fd4524
be1c5/diff:/var/lib/docker/overlay2/9688c4374d57bc2130f134b18bbc83016e298620559f4b2da558fbd3a51802f9/diff:/var/lib/docker/overlay2/d0ce8019900753a24a3d908802bc369f9944dc38825e135bb661d583a814b601/diff:/var/lib/docker/overlay2/7e72f786a28e1ede644d16ead1d0a341b1773a18e674889f889711733bb2585f/diff:/var/lib/docker/overlay2/f53da6ec1d2338a17d431f819027dd10596c3961ed80bd1878fc2fc4f310b0fd/diff:/var/lib/docker/overlay2/315ce817cf1eeab5630a9148c09730b283c49d5247e8bfdc247cc5d046181254/diff",
	                "MergedDir": "/var/lib/docker/overlay2/30a4b00c6c6318444e63a8bcdc1e78ceb808e0cc6ebe9feef5eaedb7b4ad468e/merged",
	                "UpperDir": "/var/lib/docker/overlay2/30a4b00c6c6318444e63a8bcdc1e78ceb808e0cc6ebe9feef5eaedb7b4ad468e/diff",
	                "WorkDir": "/var/lib/docker/overlay2/30a4b00c6c6318444e63a8bcdc1e78ceb808e0cc6ebe9feef5eaedb7b4ad468e/work"
	            },
	            "Name": "overlay2"
	        },
	        "Mounts": [
	            {
	                "Type": "bind",
	                "Source": "/lib/modules",
	                "Destination": "/lib/modules",
	                "Mode": "ro",
	                "RW": false,
	                "Propagation": "rprivate"
	            },
	            {
	                "Type": "volume",
	                "Name": "missing-upgrade-821485",
	                "Source": "/var/lib/docker/volumes/missing-upgrade-821485/_data",
	                "Destination": "/var",
	                "Driver": "local",
	                "Mode": "z",
	                "RW": true,
	                "Propagation": ""
	            }
	        ],
	        "Config": {
	            "Hostname": "missing-upgrade-821485",
	            "Domainname": "",
	            "User": "root",
	            "AttachStdin": false,
	            "AttachStdout": false,
	            "AttachStderr": false,
	            "ExposedPorts": {
	                "22/tcp": {},
	                "2376/tcp": {},
	                "32443/tcp": {},
	                "5000/tcp": {},
	                "8443/tcp": {}
	            },
	            "Tty": true,
	            "OpenStdin": false,
	            "StdinOnce": false,
	            "Env": [
	                "container=docker",
	                "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
	            ],
	            "Cmd": null,
	            "Image": "gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79",
	            "Volumes": null,
	            "WorkingDir": "",
	            "Entrypoint": [
	                "/usr/local/bin/entrypoint",
	                "/sbin/init"
	            ],
	            "OnBuild": null,
	            "Labels": {
	                "created_by.minikube.sigs.k8s.io": "true",
	                "mode.minikube.sigs.k8s.io": "missing-upgrade-821485",
	                "name.minikube.sigs.k8s.io": "missing-upgrade-821485",
	                "role.minikube.sigs.k8s.io": ""
	            },
	            "StopSignal": "SIGRTMIN+3"
	        },
	        "NetworkSettings": {
	            "Bridge": "",
	            "SandboxID": "afddaa51e3e889c2298f2ea392c6510f9ae4fcd31523e9e21fb833af0f2c7d99",
	            "HairpinMode": false,
	            "LinkLocalIPv6Address": "",
	            "LinkLocalIPv6PrefixLen": 0,
	            "Ports": {
	                "22/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34529"
	                    }
	                ],
	                "2376/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34528"
	                    }
	                ],
	                "32443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34525"
	                    }
	                ],
	                "5000/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34527"
	                    }
	                ],
	                "8443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "34526"
	                    }
	                ]
	            },
	            "SandboxKey": "/var/run/docker/netns/afddaa51e3e8",
	            "SecondaryIPAddresses": null,
	            "SecondaryIPv6Addresses": null,
	            "EndpointID": "",
	            "Gateway": "",
	            "GlobalIPv6Address": "",
	            "GlobalIPv6PrefixLen": 0,
	            "IPAddress": "",
	            "IPPrefixLen": 0,
	            "IPv6Gateway": "",
	            "MacAddress": "",
	            "Networks": {
	                "missing-upgrade-821485": {
	                    "IPAMConfig": {
	                        "IPv4Address": "192.168.85.2"
	                    },
	                    "Links": null,
	                    "Aliases": [
	                        "6e3e92b6bf1e",
	                        "missing-upgrade-821485"
	                    ],
	                    "NetworkID": "14bad5e56f8475a1bcc74192457ccc5404610b5f625d6edb1a7c17cc0b27eae1",
	                    "EndpointID": "c425f37411bf6aa3fc232cd70dd74a3f7db40a755de2bdc1fd90e8ef61f5567f",
	                    "Gateway": "192.168.85.1",
	                    "IPAddress": "192.168.85.2",
	                    "IPPrefixLen": 24,
	                    "IPv6Gateway": "",
	                    "GlobalIPv6Address": "",
	                    "GlobalIPv6PrefixLen": 0,
	                    "MacAddress": "02:42:c0:a8:55:02",
	                    "DriverOpts": null
	                }
	            }
	        }
	    }
	]

                                                
                                                
-- /stdout --
helpers_test.go:239: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p missing-upgrade-821485 -n missing-upgrade-821485
helpers_test.go:239: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.Host}} -p missing-upgrade-821485 -n missing-upgrade-821485: exit status 2 (332.280985ms)

                                                
                                                
-- stdout --
	Running

                                                
                                                
-- /stdout --
helpers_test.go:239: status error: exit status 2 (may be ok)
helpers_test.go:244: <<< TestMissingContainerUpgrade FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======>  post-mortem[TestMissingContainerUpgrade]: minikube logs <======
helpers_test.go:247: (dbg) Run:  out/minikube-linux-arm64 -p missing-upgrade-821485 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-arm64 -p missing-upgrade-821485 logs -n 25: (1.276610477s)
helpers_test.go:252: TestMissingContainerUpgrade logs: 
-- stdout --
	* 
	* ==> Audit <==
	* |---------|---------------------------------------|---------------------------|---------|---------|-------------------------------|-------------------------------|
	| Command |                 Args                  |          Profile          |  User   | Version |          Start Time           |           End Time            |
	|---------|---------------------------------------|---------------------------|---------|---------|-------------------------------|-------------------------------|
	| unpause | -p pause-498674                       | pause-498674              | jenkins | v1.31.2 | 30 Aug 23 23:25 UTC           | 30 Aug 23 23:25 UTC           |
	|         | --alsologtostderr -v=5                |                           |         |         |                               |                               |
	| pause   | -p pause-498674                       | pause-498674              | jenkins | v1.31.2 | 30 Aug 23 23:25 UTC           | 30 Aug 23 23:25 UTC           |
	|         | --alsologtostderr -v=5                |                           |         |         |                               |                               |
	| delete  | -p pause-498674                       | pause-498674              | jenkins | v1.31.2 | 30 Aug 23 23:25 UTC           | 30 Aug 23 23:25 UTC           |
	|         | --alsologtostderr -v=5                |                           |         |         |                               |                               |
	| delete  | -p pause-498674                       | pause-498674              | jenkins | v1.31.2 | 30 Aug 23 23:25 UTC           | 30 Aug 23 23:25 UTC           |
	| start   | -p force-systemd-flag-488784          | force-systemd-flag-488784 | jenkins | v1.31.2 | 30 Aug 23 23:25 UTC           | 30 Aug 23 23:26 UTC           |
	|         | --memory=2048 --force-systemd         |                           |         |         |                               |                               |
	|         | --alsologtostderr                     |                           |         |         |                               |                               |
	|         | -v=5 --driver=docker                  |                           |         |         |                               |                               |
	|         | --container-runtime=containerd        |                           |         |         |                               |                               |
	| ssh     | force-systemd-env-213826              | force-systemd-env-213826  | jenkins | v1.31.2 | 30 Aug 23 23:26 UTC           | 30 Aug 23 23:26 UTC           |
	|         | ssh cat                               |                           |         |         |                               |                               |
	|         | /etc/containerd/config.toml           |                           |         |         |                               |                               |
	| delete  | -p force-systemd-env-213826           | force-systemd-env-213826  | jenkins | v1.31.2 | 30 Aug 23 23:26 UTC           | 30 Aug 23 23:26 UTC           |
	| start   | -p cert-expiration-040323             | cert-expiration-040323    | jenkins | v1.31.2 | 30 Aug 23 23:26 UTC           | 30 Aug 23 23:26 UTC           |
	|         | --memory=2048                         |                           |         |         |                               |                               |
	|         | --cert-expiration=3m                  |                           |         |         |                               |                               |
	|         | --driver=docker                       |                           |         |         |                               |                               |
	|         | --container-runtime=containerd        |                           |         |         |                               |                               |
	| ssh     | force-systemd-flag-488784             | force-systemd-flag-488784 | jenkins | v1.31.2 | 30 Aug 23 23:26 UTC           | 30 Aug 23 23:26 UTC           |
	|         | ssh cat                               |                           |         |         |                               |                               |
	|         | /etc/containerd/config.toml           |                           |         |         |                               |                               |
	| delete  | -p force-systemd-flag-488784          | force-systemd-flag-488784 | jenkins | v1.31.2 | 30 Aug 23 23:26 UTC           | 30 Aug 23 23:26 UTC           |
	| start   | -p cert-options-372424                | cert-options-372424       | jenkins | v1.31.2 | 30 Aug 23 23:26 UTC           | 30 Aug 23 23:27 UTC           |
	|         | --memory=2048                         |                           |         |         |                               |                               |
	|         | --apiserver-ips=127.0.0.1             |                           |         |         |                               |                               |
	|         | --apiserver-ips=192.168.15.15         |                           |         |         |                               |                               |
	|         | --apiserver-names=localhost           |                           |         |         |                               |                               |
	|         | --apiserver-names=www.google.com      |                           |         |         |                               |                               |
	|         | --apiserver-port=8555                 |                           |         |         |                               |                               |
	|         | --driver=docker                       |                           |         |         |                               |                               |
	|         | --container-runtime=containerd        |                           |         |         |                               |                               |
	| ssh     | cert-options-372424 ssh               | cert-options-372424       | jenkins | v1.31.2 | 30 Aug 23 23:27 UTC           | 30 Aug 23 23:27 UTC           |
	|         | openssl x509 -text -noout -in         |                           |         |         |                               |                               |
	|         | /var/lib/minikube/certs/apiserver.crt |                           |         |         |                               |                               |
	| ssh     | -p cert-options-372424 -- sudo        | cert-options-372424       | jenkins | v1.31.2 | 30 Aug 23 23:27 UTC           | 30 Aug 23 23:27 UTC           |
	|         | cat /etc/kubernetes/admin.conf        |                           |         |         |                               |                               |
	| delete  | -p cert-options-372424                | cert-options-372424       | jenkins | v1.31.2 | 30 Aug 23 23:27 UTC           | 30 Aug 23 23:27 UTC           |
	| start   | -p stopped-upgrade-636929             | stopped-upgrade-636929    | jenkins | v1.22.0 | Wed, 30 Aug 2023 23:27:19 UTC | Wed, 30 Aug 2023 23:28:53 UTC |
	|         | --memory=2200                         |                           |         |         |                               |                               |
	|         | --vm-driver=docker                    |                           |         |         |                               |                               |
	|         | --container-runtime=containerd        |                           |         |         |                               |                               |
	| -p      | stopped-upgrade-636929 stop           | stopped-upgrade-636929    | jenkins | v1.22.0 | Wed, 30 Aug 2023 23:28:53 UTC | Wed, 30 Aug 2023 23:29:06 UTC |
	| start   | -p stopped-upgrade-636929             | stopped-upgrade-636929    | jenkins | v1.31.2 | 30 Aug 23 23:29 UTC           | 30 Aug 23 23:30 UTC           |
	|         | --memory=2200                         |                           |         |         |                               |                               |
	|         | --alsologtostderr                     |                           |         |         |                               |                               |
	|         | -v=1 --driver=docker                  |                           |         |         |                               |                               |
	|         | --container-runtime=containerd        |                           |         |         |                               |                               |
	| start   | -p cert-expiration-040323             | cert-expiration-040323    | jenkins | v1.31.2 | 30 Aug 23 23:29 UTC           | 30 Aug 23 23:30 UTC           |
	|         | --memory=2048                         |                           |         |         |                               |                               |
	|         | --cert-expiration=8760h               |                           |         |         |                               |                               |
	|         | --driver=docker                       |                           |         |         |                               |                               |
	|         | --container-runtime=containerd        |                           |         |         |                               |                               |
	| delete  | -p cert-expiration-040323             | cert-expiration-040323    | jenkins | v1.31.2 | 30 Aug 23 23:30 UTC           | 30 Aug 23 23:30 UTC           |
	| delete  | -p stopped-upgrade-636929             | stopped-upgrade-636929    | jenkins | v1.31.2 | 30 Aug 23 23:30 UTC           | 30 Aug 23 23:30 UTC           |
	| start   | -p kubernetes-upgrade-164934          | kubernetes-upgrade-164934 | jenkins | v1.31.2 | 30 Aug 23 23:30 UTC           | 30 Aug 23 23:31 UTC           |
	|         | --memory=2200                         |                           |         |         |                               |                               |
	|         | --kubernetes-version=v1.16.0          |                           |         |         |                               |                               |
	|         | --alsologtostderr                     |                           |         |         |                               |                               |
	|         | -v=1 --driver=docker                  |                           |         |         |                               |                               |
	|         | --container-runtime=containerd        |                           |         |         |                               |                               |
	| stop    | -p kubernetes-upgrade-164934          | kubernetes-upgrade-164934 | jenkins | v1.31.2 | 30 Aug 23 23:31 UTC           | 30 Aug 23 23:31 UTC           |
	| start   | -p kubernetes-upgrade-164934          | kubernetes-upgrade-164934 | jenkins | v1.31.2 | 30 Aug 23 23:31 UTC           |                               |
	|         | --memory=2200                         |                           |         |         |                               |                               |
	|         | --kubernetes-version=v1.28.1          |                           |         |         |                               |                               |
	|         | --alsologtostderr                     |                           |         |         |                               |                               |
	|         | -v=1 --driver=docker                  |                           |         |         |                               |                               |
	|         | --container-runtime=containerd        |                           |         |         |                               |                               |
	| start   | -p missing-upgrade-821485             | missing-upgrade-821485    | jenkins | v1.22.0 | Wed, 30 Aug 2023 23:30:09 UTC | Wed, 30 Aug 2023 23:31:40 UTC |
	|         | --memory=2200 --driver=docker         |                           |         |         |                               |                               |
	|         | --container-runtime=containerd        |                           |         |         |                               |                               |
	| start   | -p missing-upgrade-821485             | missing-upgrade-821485    | jenkins | v1.31.2 | 30 Aug 23 23:31 UTC           |                               |
	|         | --memory=2200                         |                           |         |         |                               |                               |
	|         | --alsologtostderr                     |                           |         |         |                               |                               |
	|         | -v=1 --driver=docker                  |                           |         |         |                               |                               |
	|         | --container-runtime=containerd        |                           |         |         |                               |                               |
	|---------|---------------------------------------|---------------------------|---------|---------|-------------------------------|-------------------------------|
	
	* 
	* ==> Last Start <==
	* Log file created at: 2023/08/30 23:31:52
	Running on machine: ip-172-31-31-251
	Binary: Built with gc go1.20.7 for linux/arm64
	Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
	I0830 23:31:52.736720 1365243 out.go:296] Setting OutFile to fd 1 ...
	I0830 23:31:52.736917 1365243 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:31:52.736945 1365243 out.go:309] Setting ErrFile to fd 2...
	I0830 23:31:52.736963 1365243 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:31:52.737301 1365243 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
	I0830 23:31:52.737795 1365243 out.go:303] Setting JSON to false
	I0830 23:31:52.738879 1365243 start.go:128] hostinfo: {"hostname":"ip-172-31-31-251","uptime":29647,"bootTime":1693408666,"procs":285,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1043-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"982e3628-3742-4b3e-bb63-ac1b07660ec7"}
	I0830 23:31:52.738978 1365243 start.go:138] virtualization:  
	I0830 23:31:52.743745 1365243 out.go:177] * [missing-upgrade-821485] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	I0830 23:31:52.745838 1365243 out.go:177]   - MINIKUBE_LOCATION=17114
	I0830 23:31:52.745928 1365243 notify.go:220] Checking for updates...
	I0830 23:31:52.750460 1365243 out.go:177]   - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0830 23:31:52.752093 1365243 out.go:177]   - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	I0830 23:31:52.754204 1365243 out.go:177]   - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	I0830 23:31:52.755949 1365243 out.go:177]   - MINIKUBE_BIN=out/minikube-linux-arm64
	I0830 23:31:52.757634 1365243 out.go:177]   - MINIKUBE_FORCE_SYSTEMD=
	I0830 23:31:52.759777 1365243 config.go:182] Loaded profile config "missing-upgrade-821485": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.21.2
	I0830 23:31:52.761874 1365243 out.go:177] * Kubernetes 1.28.1 is now available. If you would like to upgrade, specify: --kubernetes-version=v1.28.1
	I0830 23:31:52.763498 1365243 driver.go:373] Setting default libvirt URI to qemu:///system
	I0830 23:31:52.800333 1365243 docker.go:121] docker version: linux-24.0.5:Docker Engine - Community
	I0830 23:31:52.800434 1365243 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 23:31:52.950364 1365243 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:1 ContainersRunning:1 ContainersPaused:0 ContainersStopped:0 Images:4 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:33 OomKillDisable:true NGoroutines:45 SystemTime:2023-08-30 23:31:52.939781488 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 23:31:52.950466 1365243 docker.go:294] overlay module found
	I0830 23:31:52.953092 1365243 out.go:177] * Using the docker driver based on existing profile
	I0830 23:31:52.955081 1365243 start.go:298] selected driver: docker
	I0830 23:31:52.955104 1365243 start.go:902] validating driver "docker" against &{Name:missing-upgrade-821485 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.99.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.21.2 ClusterName:missing-upgrade-821485 Namespace:default APIServerName:minikubeCA APIServer
Names:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[{Component:kubelet Key:cni-conf-dir Value:/etc/cni/net.mk}] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP:192.168.85.2 Port:8443 KubernetesVersion:v1.21.2 ContainerRuntime: ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString: Mount9PVersion: MountGID: MountIP: MountMSize:0 MountOptions:[] MountPort:0 MountType: MountUID: BinaryMirror: DisableOptimizations:false DisableMetrics:false
CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0}
	I0830 23:31:52.955244 1365243 start.go:913] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
	I0830 23:31:52.955877 1365243 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 23:31:53.077077 1365243 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:1 ContainersRunning:1 ContainersPaused:0 ContainersStopped:0 Images:4 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:33 OomKillDisable:true NGoroutines:45 SystemTime:2023-08-30 23:31:53.066847445 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 23:31:53.077478 1365243 cni.go:84] Creating CNI manager for ""
	I0830 23:31:53.077495 1365243 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
	I0830 23:31:53.077507 1365243 start_flags.go:319] config:
	{Name:missing-upgrade-821485 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.99.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.21.2 ClusterName:missing-upgrade-821485 Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRI
Socket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[{Component:kubelet Key:cni-conf-dir Value:/etc/cni/net.mk}] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP:192.168.85.2 Port:8443 KubernetesVersion:v1.21.2 ContainerRuntime: ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString: Mount9PVersion: MountGID: MountIP: MountMSize:0 MountOptions:[] MountPort:0 MountType: MountUID: BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAut
hSock: SSHAgentPID:0}
	I0830 23:31:53.080582 1365243 out.go:177] * Starting control plane node missing-upgrade-821485 in cluster missing-upgrade-821485
	I0830 23:31:53.082724 1365243 cache.go:122] Beginning downloading kic base image for docker with containerd
	I0830 23:31:53.084455 1365243 out.go:177] * Pulling base image ...
	I0830 23:31:53.085998 1365243 preload.go:132] Checking if preload exists for k8s version v1.21.2 and runtime containerd
	I0830 23:31:53.086057 1365243 preload.go:148] Found local preload: /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.21.2-containerd-overlay2-arm64.tar.lz4
	I0830 23:31:53.086071 1365243 cache.go:57] Caching tarball of preloaded images
	I0830 23:31:53.086159 1365243 preload.go:174] Found /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.21.2-containerd-overlay2-arm64.tar.lz4 in cache, skipping download
	I0830 23:31:53.086172 1365243 cache.go:60] Finished verifying existence of preloaded tar for  v1.21.2 on containerd
	I0830 23:31:53.086291 1365243 profile.go:148] Saving config to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/missing-upgrade-821485/config.json ...
	I0830 23:31:53.086508 1365243 image.go:79] Checking for gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 in local docker daemon
	I0830 23:31:53.105672 1365243 image.go:83] Found gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 in local docker daemon, skipping pull
	I0830 23:31:53.105700 1365243 cache.go:145] gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 exists in daemon, skipping load
	I0830 23:31:53.105720 1365243 cache.go:195] Successfully downloaded all kic artifacts
	I0830 23:31:53.105774 1365243 start.go:365] acquiring machines lock for missing-upgrade-821485: {Name:mk0aadb9f10896eff5c72577cf25f5b348680b76 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
	I0830 23:31:53.105847 1365243 start.go:369] acquired machines lock for "missing-upgrade-821485" in 42.091µs
	I0830 23:31:53.105871 1365243 start.go:96] Skipping create...Using existing machine configuration
	I0830 23:31:53.105880 1365243 fix.go:54] fixHost starting: 
	I0830 23:31:53.106163 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:53.124361 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:53.124421 1365243 fix.go:102] recreateIfNeeded on missing-upgrade-821485: state= err=unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:53.124441 1365243 fix.go:107] machineExists: false. err=machine does not exist
	I0830 23:31:53.126587 1365243 out.go:177] * docker "missing-upgrade-821485" container is missing, will recreate.
	I0830 23:31:50.681105 1363857 ssh_runner.go:195] Run: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.28.1:$PATH" kubeadm init phase control-plane all --config /var/tmp/minikube/kubeadm.yaml"
	I0830 23:31:50.769321 1363857 ssh_runner.go:195] Run: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.28.1:$PATH" kubeadm init phase etcd local --config /var/tmp/minikube/kubeadm.yaml"
	I0830 23:31:50.866587 1363857 api_server.go:52] waiting for apiserver process to appear ...
	I0830 23:31:50.866714 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:50.890928 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:51.412343 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:51.911952 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:52.411637 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:52.912208 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:53.412529 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:53.911566 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:54.412069 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:54.911564 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:55.412300 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:53.128014 1365243 delete.go:124] DEMOLISHING missing-upgrade-821485 ...
	I0830 23:31:53.128116 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:53.151439 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	W0830 23:31:53.151512 1365243 stop.go:75] unable to get state: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:53.151550 1365243 delete.go:128] stophost failed (probably ok): ssh power off: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:53.152007 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:53.182602 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:53.182671 1365243 delete.go:82] Unable to get host status for missing-upgrade-821485, assuming it has already been deleted: state: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:53.182754 1365243 cli_runner.go:164] Run: docker container inspect -f {{.Id}} missing-upgrade-821485
	W0830 23:31:53.204567 1365243 cli_runner.go:211] docker container inspect -f {{.Id}} missing-upgrade-821485 returned with exit code 1
	I0830 23:31:53.204623 1365243 kic.go:367] could not find the container missing-upgrade-821485 to remove it. will try anyways
	I0830 23:31:53.204684 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:53.227796 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	W0830 23:31:53.227859 1365243 oci.go:84] error getting container status, will try to delete anyways: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:53.227927 1365243 cli_runner.go:164] Run: docker exec --privileged -t missing-upgrade-821485 /bin/bash -c "sudo init 0"
	W0830 23:31:53.258405 1365243 cli_runner.go:211] docker exec --privileged -t missing-upgrade-821485 /bin/bash -c "sudo init 0" returned with exit code 1
	I0830 23:31:53.258437 1365243 oci.go:647] error shutdown missing-upgrade-821485: docker exec --privileged -t missing-upgrade-821485 /bin/bash -c "sudo init 0": exit status 1
	stdout:
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:54.258611 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:54.276706 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:54.276789 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:54.276804 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:31:54.276835 1365243 retry.go:31] will retry after 444.223191ms: couldn't verify container is exited. %!v(MISSING): unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:54.721269 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:54.738607 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:54.738672 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:54.738683 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:31:54.738712 1365243 retry.go:31] will retry after 634.565723ms: couldn't verify container is exited. %!v(MISSING): unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:55.374312 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:55.395127 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:55.395184 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:55.395192 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:31:55.395214 1365243 retry.go:31] will retry after 1.256461515s: couldn't verify container is exited. %!v(MISSING): unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:56.652170 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:56.670416 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:56.670479 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:56.670499 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:31:56.670522 1365243 retry.go:31] will retry after 2.366757653s: couldn't verify container is exited. %!v(MISSING): unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:55.912290 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:56.412523 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:56.911559 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:57.411585 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:57.911549 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:58.411963 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:58.912239 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:59.411713 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:59.911957 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:00.411632 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:31:59.038293 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:31:59.056315 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:31:59.056382 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:31:59.056393 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:31:59.056418 1365243 retry.go:31] will retry after 3.625191599s: couldn't verify container is exited. %!v(MISSING): unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:32:02.684782 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:32:02.701864 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:32:02.701932 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:32:02.701953 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:32:02.701977 1365243 retry.go:31] will retry after 5.483696825s: couldn't verify container is exited. %!v(MISSING): unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:32:00.912525 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:01.412476 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:01.911606 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:02.411956 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:02.911654 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:03.412399 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:03.912212 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:04.412159 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:04.911491 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:05.411473 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:05.911595 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:06.411937 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:06.911539 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:07.411738 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:07.912441 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:08.411541 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:08.911578 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:09.411536 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:09.911553 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:10.411591 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:08.186602 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:32:08.213640 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:32:08.213701 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:32:08.213711 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:32:08.213738 1365243 retry.go:31] will retry after 3.196654656s: couldn't verify container is exited. %!v(MISSING): unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:32:11.411526 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:32:11.437112 1365243 cli_runner.go:211] docker container inspect missing-upgrade-821485 --format={{.State.Status}} returned with exit code 1
	I0830 23:32:11.437210 1365243 oci.go:659] temporary error verifying shutdown: unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	I0830 23:32:11.437223 1365243 oci.go:661] temporary error: container missing-upgrade-821485 status is  but expect it to be exited
	I0830 23:32:11.437281 1365243 oci.go:88] couldn't shut down missing-upgrade-821485 (might be okay): verify shutdown: couldn't verify container is exited. %!v(MISSING): unknown state "missing-upgrade-821485": docker container inspect missing-upgrade-821485 --format={{.State.Status}}: exit status 1
	stdout:
	
	
	stderr:
	Error response from daemon: No such container: missing-upgrade-821485
	 
	I0830 23:32:11.437365 1365243 cli_runner.go:164] Run: docker rm -f -v missing-upgrade-821485
	I0830 23:32:11.455349 1365243 cli_runner.go:164] Run: docker container inspect -f {{.Id}} missing-upgrade-821485
	W0830 23:32:11.472097 1365243 cli_runner.go:211] docker container inspect -f {{.Id}} missing-upgrade-821485 returned with exit code 1
	I0830 23:32:11.472198 1365243 cli_runner.go:164] Run: docker network inspect missing-upgrade-821485 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	I0830 23:32:11.490425 1365243 cli_runner.go:164] Run: docker network rm missing-upgrade-821485
	I0830 23:32:11.587805 1365243 fix.go:114] Sleeping 1 second for extra luck!
	I0830 23:32:12.587901 1365243 start.go:125] createHost starting for "" (driver="docker")
	I0830 23:32:12.590258 1365243 out.go:204] * Creating docker container (CPUs=2, Memory=2200MB) ...
	I0830 23:32:12.590389 1365243 start.go:159] libmachine.API.Create for "missing-upgrade-821485" (driver="docker")
	I0830 23:32:12.590424 1365243 client.go:168] LocalClient.Create starting
	I0830 23:32:12.590500 1365243 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem
	I0830 23:32:12.590554 1365243 main.go:141] libmachine: Decoding PEM data...
	I0830 23:32:12.590579 1365243 main.go:141] libmachine: Parsing certificate...
	I0830 23:32:12.590648 1365243 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem
	I0830 23:32:12.590683 1365243 main.go:141] libmachine: Decoding PEM data...
	I0830 23:32:12.590700 1365243 main.go:141] libmachine: Parsing certificate...
	I0830 23:32:12.591006 1365243 cli_runner.go:164] Run: docker network inspect missing-upgrade-821485 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	W0830 23:32:12.608394 1365243 cli_runner.go:211] docker network inspect missing-upgrade-821485 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
	I0830 23:32:12.608486 1365243 network_create.go:281] running [docker network inspect missing-upgrade-821485] to gather additional debugging logs...
	I0830 23:32:12.608508 1365243 cli_runner.go:164] Run: docker network inspect missing-upgrade-821485
	W0830 23:32:12.626064 1365243 cli_runner.go:211] docker network inspect missing-upgrade-821485 returned with exit code 1
	I0830 23:32:12.626104 1365243 network_create.go:284] error running [docker network inspect missing-upgrade-821485]: docker network inspect missing-upgrade-821485: exit status 1
	stdout:
	[]
	
	stderr:
	Error response from daemon: network missing-upgrade-821485 not found
	I0830 23:32:12.626135 1365243 network_create.go:286] output of [docker network inspect missing-upgrade-821485]: -- stdout --
	[]
	
	-- /stdout --
	** stderr ** 
	Error response from daemon: network missing-upgrade-821485 not found
	
	** /stderr **
	I0830 23:32:12.626200 1365243 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	I0830 23:32:12.644369 1365243 network.go:214] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-ed2d417202d5 IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:02:42:e9:35:f3:fe} reservation:<nil>}
	I0830 23:32:12.644763 1365243 network.go:214] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-b83085549c6c IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:02:42:58:7c:4f:97} reservation:<nil>}
	I0830 23:32:12.645253 1365243 network.go:214] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-4f518efc6da3 IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:02:42:89:2a:5e:24} reservation:<nil>}
	I0830 23:32:12.645663 1365243 network.go:214] skipping subnet 192.168.76.0/24 that is taken: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName:br-28e919bf1baa IfaceIPv4:192.168.76.1 IfaceMTU:1500 IfaceMAC:02:42:6e:68:af:9c} reservation:<nil>}
	I0830 23:32:12.646602 1365243 network.go:209] using free private subnet 192.168.85.0/24: &{IP:192.168.85.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.85.0/24 Gateway:192.168.85.1 ClientMin:192.168.85.2 ClientMax:192.168.85.254 Broadcast:192.168.85.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x4002a23d10}
	I0830 23:32:12.646627 1365243 network_create.go:123] attempt to create docker network missing-upgrade-821485 192.168.85.0/24 with gateway 192.168.85.1 and MTU of 1500 ...
	I0830 23:32:12.646693 1365243 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.85.0/24 --gateway=192.168.85.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=missing-upgrade-821485 missing-upgrade-821485
	I0830 23:32:12.719796 1365243 network_create.go:107] docker network missing-upgrade-821485 192.168.85.0/24 created
	I0830 23:32:12.719826 1365243 kic.go:117] calculated static IP "192.168.85.2" for the "missing-upgrade-821485" container
	I0830 23:32:12.719924 1365243 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
	I0830 23:32:10.911606 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:11.411580 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:11.912251 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:12.411741 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:12.911556 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:13.411534 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:13.912157 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:14.411549 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:14.912015 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:15.411550 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:12.737249 1365243 cli_runner.go:164] Run: docker volume create missing-upgrade-821485 --label name.minikube.sigs.k8s.io=missing-upgrade-821485 --label created_by.minikube.sigs.k8s.io=true
	I0830 23:32:12.756029 1365243 oci.go:103] Successfully created a docker volume missing-upgrade-821485
	I0830 23:32:12.756112 1365243 cli_runner.go:164] Run: docker run --rm --name missing-upgrade-821485-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=missing-upgrade-821485 --entrypoint /usr/bin/test -v missing-upgrade-821485:/var gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 -d /var/lib
	I0830 23:32:13.286154 1365243 oci.go:107] Successfully prepared a docker volume missing-upgrade-821485
	I0830 23:32:13.286210 1365243 preload.go:132] Checking if preload exists for k8s version v1.21.2 and runtime containerd
	I0830 23:32:13.286229 1365243 kic.go:190] Starting extracting preloaded images to volume ...
	I0830 23:32:13.286317 1365243 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.21.2-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v missing-upgrade-821485:/extractDir gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 -I lz4 -xf /preloaded.tar -C /extractDir
	I0830 23:32:15.911832 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:16.412415 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:16.911486 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:17.412371 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:17.912002 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:18.411605 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:18.911562 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:19.412055 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:19.912023 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:20.412126 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:18.094107 1365243 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.21.2-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v missing-upgrade-821485:/extractDir gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79 -I lz4 -xf /preloaded.tar -C /extractDir: (4.807748151s)
	I0830 23:32:18.094142 1365243 kic.go:199] duration metric: took 4.807906 seconds to extract preloaded images to volume
	W0830 23:32:18.094300 1365243 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
	I0830 23:32:18.094409 1365243 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
	I0830 23:32:18.167517 1365243 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname missing-upgrade-821485 --name missing-upgrade-821485 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=missing-upgrade-821485 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=missing-upgrade-821485 --network missing-upgrade-821485 --ip 192.168.85.2 --volume missing-upgrade-821485:/var --security-opt apparmor=unconfined --memory=2200mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase:v0.0.25@sha256:6f936e3443b95cd918d77623bf7b595653bb382766e280290a02b4a349e88b79
	I0830 23:32:18.557413 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Running}}
	I0830 23:32:18.579407 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	I0830 23:32:18.605656 1365243 cli_runner.go:164] Run: docker exec missing-upgrade-821485 stat /var/lib/dpkg/alternatives/iptables
	I0830 23:32:18.724768 1365243 oci.go:144] the created container "missing-upgrade-821485" has a running status.
	I0830 23:32:18.724798 1365243 kic.go:221] Creating ssh key for kic: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa...
	I0830 23:32:19.345344 1365243 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
	I0830 23:32:19.384250 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	I0830 23:32:19.409937 1365243 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
	I0830 23:32:19.409986 1365243 kic_runner.go:114] Args: [docker exec --privileged missing-upgrade-821485 chown docker:docker /home/docker/.ssh/authorized_keys]
	I0830 23:32:19.506973 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	I0830 23:32:19.536945 1365243 machine.go:88] provisioning docker machine ...
	I0830 23:32:19.536975 1365243 ubuntu.go:169] provisioning hostname "missing-upgrade-821485"
	I0830 23:32:19.537037 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:19.558760 1365243 main.go:141] libmachine: Using SSH client type: native
	I0830 23:32:19.559238 1365243 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34529 <nil> <nil>}
	I0830 23:32:19.559253 1365243 main.go:141] libmachine: About to run SSH command:
	sudo hostname missing-upgrade-821485 && echo "missing-upgrade-821485" | sudo tee /etc/hostname
	I0830 23:32:19.721990 1365243 main.go:141] libmachine: SSH cmd err, output: <nil>: missing-upgrade-821485
	
	I0830 23:32:19.722110 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:19.756584 1365243 main.go:141] libmachine: Using SSH client type: native
	I0830 23:32:19.757070 1365243 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34529 <nil> <nil>}
	I0830 23:32:19.757094 1365243 main.go:141] libmachine: About to run SSH command:
	
			if ! grep -xq '.*\smissing-upgrade-821485' /etc/hosts; then
				if grep -xq '127.0.1.1\s.*' /etc/hosts; then
					sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 missing-upgrade-821485/g' /etc/hosts;
				else 
					echo '127.0.1.1 missing-upgrade-821485' | sudo tee -a /etc/hosts; 
				fi
			fi
	I0830 23:32:19.915950 1365243 main.go:141] libmachine: SSH cmd err, output: <nil>: 
	I0830 23:32:19.916023 1365243 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/17114-1219981/.minikube CaCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/17114-1219981/.minikube}
	I0830 23:32:19.916095 1365243 ubuntu.go:177] setting up certificates
	I0830 23:32:19.916122 1365243 provision.go:83] configureAuth start
	I0830 23:32:19.916220 1365243 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" missing-upgrade-821485
	I0830 23:32:19.948839 1365243 provision.go:138] copyHostCerts
	I0830 23:32:19.948912 1365243 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem, removing ...
	I0830 23:32:19.948921 1365243 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem
	I0830 23:32:19.948996 1365243 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem (1082 bytes)
	I0830 23:32:19.949094 1365243 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem, removing ...
	I0830 23:32:19.949100 1365243 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem
	I0830 23:32:19.949166 1365243 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem (1123 bytes)
	I0830 23:32:19.949236 1365243 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem, removing ...
	I0830 23:32:19.949240 1365243 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem
	I0830 23:32:19.949266 1365243 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem (1679 bytes)
	I0830 23:32:19.949324 1365243 provision.go:112] generating server cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem org=jenkins.missing-upgrade-821485 san=[192.168.85.2 127.0.0.1 localhost 127.0.0.1 minikube missing-upgrade-821485]
	I0830 23:32:20.910285 1365243 provision.go:172] copyRemoteCerts
	I0830 23:32:20.910397 1365243 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
	I0830 23:32:20.910475 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:20.933385 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:21.027946 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
	I0830 23:32:21.054533 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem --> /etc/docker/server.pem (1241 bytes)
	I0830 23:32:21.082998 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
	I0830 23:32:21.110396 1365243 provision.go:86] duration metric: configureAuth took 1.194238702s
	I0830 23:32:21.110433 1365243 ubuntu.go:193] setting minikube options for container-runtime
	I0830 23:32:21.110630 1365243 config.go:182] Loaded profile config "missing-upgrade-821485": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.21.2
	I0830 23:32:21.110646 1365243 machine.go:91] provisioned docker machine in 1.573683801s
	I0830 23:32:21.110653 1365243 client.go:171] LocalClient.Create took 8.520220852s
	I0830 23:32:21.110671 1365243 start.go:167] duration metric: libmachine.API.Create for "missing-upgrade-821485" took 8.520280172s
	I0830 23:32:21.110682 1365243 start.go:300] post-start starting for "missing-upgrade-821485" (driver="docker")
	I0830 23:32:21.110692 1365243 start.go:329] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
	I0830 23:32:21.110750 1365243 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
	I0830 23:32:21.110794 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:21.138599 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:21.232506 1365243 ssh_runner.go:195] Run: cat /etc/os-release
	I0830 23:32:21.237238 1365243 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
	I0830 23:32:21.237265 1365243 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
	I0830 23:32:21.237277 1365243 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
	I0830 23:32:21.237289 1365243 info.go:137] Remote host: Ubuntu 20.04.2 LTS
	I0830 23:32:21.237299 1365243 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/addons for local assets ...
	I0830 23:32:21.237358 1365243 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/files for local assets ...
	I0830 23:32:21.237443 1365243 filesync.go:149] local asset: /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem -> 12253752.pem in /etc/ssl/certs
	I0830 23:32:21.237546 1365243 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
	I0830 23:32:21.248159 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem --> /etc/ssl/certs/12253752.pem (1708 bytes)
	I0830 23:32:21.282644 1365243 start.go:303] post-start completed in 171.942358ms
	I0830 23:32:21.283090 1365243 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" missing-upgrade-821485
	I0830 23:32:21.306835 1365243 profile.go:148] Saving config to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/missing-upgrade-821485/config.json ...
	I0830 23:32:21.307129 1365243 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0830 23:32:21.307181 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:21.328360 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:21.420052 1365243 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
	I0830 23:32:21.426005 1365243 start.go:128] duration metric: createHost completed in 8.838066552s
	I0830 23:32:21.426100 1365243 cli_runner.go:164] Run: docker container inspect missing-upgrade-821485 --format={{.State.Status}}
	W0830 23:32:21.467765 1365243 fix.go:128] unexpected machine state, will restart: <nil>
	I0830 23:32:21.467788 1365243 machine.go:88] provisioning docker machine ...
	I0830 23:32:21.467805 1365243 ubuntu.go:169] provisioning hostname "missing-upgrade-821485"
	I0830 23:32:21.467870 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:21.490981 1365243 main.go:141] libmachine: Using SSH client type: native
	I0830 23:32:21.491443 1365243 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34529 <nil> <nil>}
	I0830 23:32:21.491456 1365243 main.go:141] libmachine: About to run SSH command:
	sudo hostname missing-upgrade-821485 && echo "missing-upgrade-821485" | sudo tee /etc/hostname
	I0830 23:32:21.630064 1365243 main.go:141] libmachine: SSH cmd err, output: <nil>: missing-upgrade-821485
	
	I0830 23:32:21.630146 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:21.656678 1365243 main.go:141] libmachine: Using SSH client type: native
	I0830 23:32:21.657116 1365243 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3a0570] 0x3a2f00 <nil>  [] 0s} 127.0.0.1 34529 <nil> <nil>}
	I0830 23:32:21.657172 1365243 main.go:141] libmachine: About to run SSH command:
	
			if ! grep -xq '.*\smissing-upgrade-821485' /etc/hosts; then
				if grep -xq '127.0.1.1\s.*' /etc/hosts; then
					sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 missing-upgrade-821485/g' /etc/hosts;
				else 
					echo '127.0.1.1 missing-upgrade-821485' | sudo tee -a /etc/hosts; 
				fi
			fi
	I0830 23:32:21.794803 1365243 main.go:141] libmachine: SSH cmd err, output: <nil>: 
	I0830 23:32:21.794830 1365243 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/17114-1219981/.minikube CaCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/17114-1219981/.minikube}
	I0830 23:32:21.794856 1365243 ubuntu.go:177] setting up certificates
	I0830 23:32:21.794870 1365243 provision.go:83] configureAuth start
	I0830 23:32:21.794935 1365243 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" missing-upgrade-821485
	I0830 23:32:21.818851 1365243 provision.go:138] copyHostCerts
	I0830 23:32:21.818915 1365243 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem, removing ...
	I0830 23:32:21.818935 1365243 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem
	I0830 23:32:21.819014 1365243 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.pem (1082 bytes)
	I0830 23:32:21.819123 1365243 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem, removing ...
	I0830 23:32:21.819133 1365243 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem
	I0830 23:32:21.819160 1365243 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/cert.pem (1123 bytes)
	I0830 23:32:21.819220 1365243 exec_runner.go:144] found /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem, removing ...
	I0830 23:32:21.819229 1365243 exec_runner.go:203] rm: /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem
	I0830 23:32:21.819253 1365243 exec_runner.go:151] cp: /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/17114-1219981/.minikube/key.pem (1679 bytes)
	I0830 23:32:21.819336 1365243 provision.go:112] generating server cert: /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca-key.pem org=jenkins.missing-upgrade-821485 san=[192.168.85.2 127.0.0.1 localhost 127.0.0.1 minikube missing-upgrade-821485]
	I0830 23:32:22.102738 1365243 provision.go:172] copyRemoteCerts
	I0830 23:32:22.102806 1365243 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
	I0830 23:32:22.102853 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:22.121794 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:22.218890 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
	I0830 23:32:22.244766 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server.pem --> /etc/docker/server.pem (1241 bytes)
	I0830 23:32:22.271374 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
	I0830 23:32:22.304362 1365243 provision.go:86] duration metric: configureAuth took 509.477014ms
	I0830 23:32:22.304389 1365243 ubuntu.go:193] setting minikube options for container-runtime
	I0830 23:32:22.304572 1365243 config.go:182] Loaded profile config "missing-upgrade-821485": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.21.2
	I0830 23:32:22.304586 1365243 machine.go:91] provisioned docker machine in 836.791852ms
	I0830 23:32:22.304593 1365243 start.go:300] post-start starting for "missing-upgrade-821485" (driver="docker")
	I0830 23:32:22.304602 1365243 start.go:329] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
	I0830 23:32:22.304653 1365243 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
	I0830 23:32:22.304695 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:22.327682 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:22.433578 1365243 ssh_runner.go:195] Run: cat /etc/os-release
	I0830 23:32:22.437960 1365243 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
	I0830 23:32:22.437989 1365243 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
	I0830 23:32:22.438002 1365243 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
	I0830 23:32:22.438010 1365243 info.go:137] Remote host: Ubuntu 20.04.2 LTS
	I0830 23:32:22.438020 1365243 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/addons for local assets ...
	I0830 23:32:22.438085 1365243 filesync.go:126] Scanning /home/jenkins/minikube-integration/17114-1219981/.minikube/files for local assets ...
	I0830 23:32:22.438161 1365243 filesync.go:149] local asset: /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem -> 12253752.pem in /etc/ssl/certs
	I0830 23:32:22.438283 1365243 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
	I0830 23:32:22.448726 1365243 ssh_runner.go:362] scp /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/ssl/certs/12253752.pem --> /etc/ssl/certs/12253752.pem (1708 bytes)
	I0830 23:32:22.480544 1365243 start.go:303] post-start completed in 175.936024ms
	I0830 23:32:22.480629 1365243 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0830 23:32:22.480690 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:22.507869 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:22.600238 1365243 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
	I0830 23:32:22.607154 1365243 fix.go:56] fixHost completed within 29.501258817s
	I0830 23:32:22.607175 1365243 start.go:83] releasing machines lock for "missing-upgrade-821485", held for 29.501315669s
	I0830 23:32:22.607244 1365243 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" missing-upgrade-821485
	I0830 23:32:22.626574 1365243 ssh_runner.go:195] Run: cat /version.json
	I0830 23:32:22.626625 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:22.626838 1365243 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
	I0830 23:32:22.626893 1365243 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" missing-upgrade-821485
	I0830 23:32:22.659612 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:22.664621 1365243 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34529 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/missing-upgrade-821485/id_rsa Username:docker}
	I0830 23:32:20.911466 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:21.411538 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:21.912208 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:22.412234 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:22.912391 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:23.411888 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:23.911476 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:24.412341 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:24.912170 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:25.411927 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	W0830 23:32:22.878282 1365243 start.go:419] Unable to open version.json: cat /version.json: Process exited with status 1
	stdout:
	
	stderr:
	cat: /version.json: No such file or directory
	I0830 23:32:22.878363 1365243 ssh_runner.go:195] Run: systemctl --version
	I0830 23:32:22.883858 1365243 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
	I0830 23:32:22.889584 1365243 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
	I0830 23:32:22.918673 1365243 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
	I0830 23:32:22.918815 1365243 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%!p(MISSING), " -exec sh -c "sudo mv {} {}.mk_disabled" ;
	I0830 23:32:22.953301 1365243 cni.go:262] disabled [/etc/cni/net.d/100-crio-bridge.conf, /etc/cni/net.d/87-podman-bridge.conflist] bridge cni config(s)
	I0830 23:32:22.953326 1365243 start.go:466] detecting cgroup driver to use...
	I0830 23:32:22.953359 1365243 detect.go:196] detected "cgroupfs" cgroup driver on host os
	I0830 23:32:22.953412 1365243 ssh_runner.go:195] Run: sudo systemctl stop -f crio
	I0830 23:32:22.968880 1365243 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
	I0830 23:32:22.990880 1365243 docker.go:196] disabling cri-docker service (if available) ...
	I0830 23:32:22.990956 1365243 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
	I0830 23:32:23.003929 1365243 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
	I0830 23:32:23.017050 1365243 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
	W0830 23:32:23.031089 1365243 docker.go:206] Failed to disable socket "cri-docker.socket" (might be ok): sudo systemctl disable cri-docker.socket: Process exited with status 1
	stdout:
	
	stderr:
	Failed to disable unit: Unit file cri-docker.socket does not exist.
	I0830 23:32:23.031165 1365243 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
	I0830 23:32:23.135493 1365243 docker.go:212] disabling docker service ...
	I0830 23:32:23.135586 1365243 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
	I0830 23:32:23.159972 1365243 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
	I0830 23:32:23.173750 1365243 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
	I0830 23:32:23.276554 1365243 ssh_runner.go:195] Run: sudo systemctl mask docker.service
	I0830 23:32:23.384058 1365243 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
	I0830 23:32:23.396213 1365243 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///run/containerd/containerd.sock
	" | sudo tee /etc/crictl.yaml"
	I0830 23:32:23.417196 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.4.1"|' /etc/containerd/config.toml"
	I0830 23:32:23.429390 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
	I0830 23:32:23.440691 1365243 containerd.go:145] configuring containerd to use "cgroupfs" as cgroup driver...
	I0830 23:32:23.440765 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
	I0830 23:32:23.452302 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
	I0830 23:32:23.463576 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
	I0830 23:32:23.475151 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
	I0830 23:32:23.486792 1365243 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
	I0830 23:32:23.497280 1365243 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
	I0830 23:32:23.508419 1365243 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
	I0830 23:32:23.518597 1365243 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
	I0830 23:32:23.528050 1365243 ssh_runner.go:195] Run: sudo systemctl daemon-reload
	I0830 23:32:23.626597 1365243 ssh_runner.go:195] Run: sudo systemctl restart containerd
	I0830 23:32:23.710815 1365243 start.go:513] Will wait 60s for socket path /run/containerd/containerd.sock
	I0830 23:32:23.710916 1365243 ssh_runner.go:195] Run: stat /run/containerd/containerd.sock
	I0830 23:32:23.715841 1365243 start.go:534] Will wait 60s for crictl version
	I0830 23:32:23.715905 1365243 ssh_runner.go:195] Run: which crictl
	I0830 23:32:23.720572 1365243 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
	I0830 23:32:23.754580 1365243 retry.go:31] will retry after 13.407237425s: Temporary Error: sudo /usr/bin/crictl version: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:32:23Z" level=fatal msg="getting the runtime version: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	I0830 23:32:25.911676 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:26.411590 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:26.911728 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:27.412144 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:27.911578 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:28.411693 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:28.911711 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:29.412237 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:29.912244 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:30.412493 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:30.912205 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:31.411648 1363857 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:32:31.426344 1363857 api_server.go:72] duration metric: took 40.559756485s to wait for apiserver process to appear ...
	I0830 23:32:31.426364 1363857 api_server.go:88] waiting for apiserver healthz status ...
	I0830 23:32:31.426381 1363857 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
	I0830 23:32:37.166241 1365243 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
	I0830 23:32:37.196024 1365243 retry.go:31] will retry after 10.077638903s: Temporary Error: sudo /usr/bin/crictl version: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:32:37Z" level=fatal msg="getting the runtime version: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	I0830 23:32:36.427470 1363857 api_server.go:269] stopped: https://192.168.76.2:8443/healthz: Get "https://192.168.76.2:8443/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
	I0830 23:32:36.427526 1363857 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
	I0830 23:32:41.427844 1363857 api_server.go:269] stopped: https://192.168.76.2:8443/healthz: Get "https://192.168.76.2:8443/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
	I0830 23:32:41.928556 1363857 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
	I0830 23:32:47.273919 1365243 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
	I0830 23:32:47.306054 1365243 retry.go:31] will retry after 19.362516092s: Temporary Error: sudo /usr/bin/crictl version: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:32:47Z" level=fatal msg="getting the runtime version: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	I0830 23:32:46.929200 1363857 api_server.go:269] stopped: https://192.168.76.2:8443/healthz: Get "https://192.168.76.2:8443/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
	I0830 23:32:46.929239 1363857 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
	I0830 23:32:51.929622 1363857 api_server.go:269] stopped: https://192.168.76.2:8443/healthz: Get "https://192.168.76.2:8443/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
	I0830 23:32:51.929660 1363857 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
	I0830 23:32:52.649567 1363857 api_server.go:269] stopped: https://192.168.76.2:8443/healthz: Get "https://192.168.76.2:8443/healthz": read tcp 192.168.76.1:50782->192.168.76.2:8443: read: connection reset by peer
	I0830 23:32:52.649606 1363857 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
	I0830 23:32:52.649975 1363857 api_server.go:269] stopped: https://192.168.76.2:8443/healthz: Get "https://192.168.76.2:8443/healthz": dial tcp 192.168.76.2:8443: connect: connection refused
	I0830 23:32:52.928441 1363857 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
	I0830 23:32:52.928820 1363857 api_server.go:269] stopped: https://192.168.76.2:8443/healthz: Get "https://192.168.76.2:8443/healthz": dial tcp 192.168.76.2:8443: connect: connection refused
	I0830 23:32:53.428340 1363857 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
	I0830 23:32:58.429596 1363857 api_server.go:269] stopped: https://192.168.76.2:8443/healthz: Get "https://192.168.76.2:8443/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
	I0830 23:32:58.429636 1363857 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
	I0830 23:33:03.430362 1363857 api_server.go:269] stopped: https://192.168.76.2:8443/healthz: Get "https://192.168.76.2:8443/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
	I0830 23:33:03.430422 1363857 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
	I0830 23:33:06.669856 1365243 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
	I0830 23:33:06.702391 1365243 out.go:177] 
	W0830 23:33:06.704639 1365243 out.go:239] X Exiting due to RUNTIME_ENABLE: Failed to start container runtime: Temporary Error: sudo /usr/bin/crictl version: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:33:06Z" level=fatal msg="getting the runtime version: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	
	W0830 23:33:06.704654 1365243 out.go:239] * 
	W0830 23:33:06.705678 1365243 out.go:239] ╭─────────────────────────────────────────────────────────────────────────────────────────────╮
	│                                                                                             │
	│    * If the above advice does not help, please let us know:                                 │
	│      https://github.com/kubernetes/minikube/issues/new/choose                               │
	│                                                                                             │
	│    * Please run `minikube logs --file=logs.txt` and attach logs.txt to the GitHub issue.    │
	│                                                                                             │
	╰─────────────────────────────────────────────────────────────────────────────────────────────╯
	I0830 23:33:06.708159 1365243 out.go:177] 
	
	* 
	* ==> container status <==
	* 
	* ==> containerd <==
	* -- Logs begin at Wed 2023-08-30 23:32:19 UTC, end at Wed 2023-08-30 23:33:07 UTC. --
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.706170592Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.706230331Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.706288388Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.706351254Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.706486511Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.706597563Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707135439Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707230574Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707331239Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707540429Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707577769Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707596345Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707614100Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707629263Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707643375Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707656897Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707670861Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707722265Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707739323Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707754764Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.707767588Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.708005873Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.708054093Z" level=info msg=serving... address=/run/containerd/containerd.sock
	Aug 30 23:32:23 missing-upgrade-821485 systemd[1]: Started containerd container runtime.
	Aug 30 23:32:23 missing-upgrade-821485 containerd[628]: time="2023-08-30T23:32:23.709230518Z" level=info msg="containerd successfully booted in 0.036706s"
	
	* 
	* ==> describe nodes <==
	* 
	* ==> dmesg <==
	* [  +0.001047] FS-Cache: O-key=[8] '6a415c0100000000'
	[  +0.000737] FS-Cache: N-cookie c=00000066 [p=0000005d fl=2 nc=0 na=1]
	[  +0.000924] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=00000000a41d18fb
	[  +0.001010] FS-Cache: N-key=[8] '6a415c0100000000'
	[  +0.002699] FS-Cache: Duplicate cookie detected
	[  +0.000740] FS-Cache: O-cookie c=00000060 [p=0000005d fl=226 nc=0 na=1]
	[  +0.000935] FS-Cache: O-cookie d=00000000d8a48a2b{9p.inode} n=000000009512bf07
	[  +0.001010] FS-Cache: O-key=[8] '6a415c0100000000'
	[  +0.000741] FS-Cache: N-cookie c=00000067 [p=0000005d fl=2 nc=0 na=1]
	[  +0.000916] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=00000000e0bc5e95
	[  +0.001060] FS-Cache: N-key=[8] '6a415c0100000000'
	[  +2.904353] FS-Cache: Duplicate cookie detected
	[  +0.000708] FS-Cache: O-cookie c=0000005e [p=0000005d fl=226 nc=0 na=1]
	[  +0.000962] FS-Cache: O-cookie d=00000000d8a48a2b{9p.inode} n=000000007bdd3326
	[  +0.001080] FS-Cache: O-key=[8] '69415c0100000000'
	[  +0.000675] FS-Cache: N-cookie c=00000069 [p=0000005d fl=2 nc=0 na=1]
	[  +0.000894] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=00000000a41d18fb
	[  +0.001068] FS-Cache: N-key=[8] '69415c0100000000'
	[  +0.462659] FS-Cache: Duplicate cookie detected
	[  +0.000688] FS-Cache: O-cookie c=00000063 [p=0000005d fl=226 nc=0 na=1]
	[  +0.000963] FS-Cache: O-cookie d=00000000d8a48a2b{9p.inode} n=00000000ec866464
	[  +0.001134] FS-Cache: O-key=[8] '6f415c0100000000'
	[  +0.000702] FS-Cache: N-cookie c=0000006a [p=0000005d fl=2 nc=0 na=1]
	[  +0.000917] FS-Cache: N-cookie d=00000000d8a48a2b{9p.inode} n=000000001a64e3e4
	[  +0.001038] FS-Cache: N-key=[8] '6f415c0100000000'
	
	* 
	* ==> kernel <==
	*  23:33:08 up  8:15,  0 users,  load average: 1.62, 2.55, 2.21
	Linux missing-upgrade-821485 5.15.0-1043-aws #48~20.04.1-Ubuntu SMP Wed Aug 16 18:32:42 UTC 2023 aarch64 aarch64 aarch64 GNU/Linux
	PRETTY_NAME="Ubuntu 20.04.2 LTS"
	
	* 
	* ==> kubelet <==
	* -- Logs begin at Wed 2023-08-30 23:32:19 UTC, end at Wed 2023-08-30 23:33:08 UTC. --
	-- No entries --
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	E0830 23:33:07.426385 1367223 logs.go:281] Failed to list containers for "kube-apiserver": crictl list: sudo crictl ps -a --quiet --name=kube-apiserver: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:33:07Z" level=fatal msg="listing containers: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	E0830 23:33:07.456984 1367223 logs.go:281] Failed to list containers for "etcd": crictl list: sudo crictl ps -a --quiet --name=etcd: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:33:07Z" level=fatal msg="listing containers: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	E0830 23:33:07.488103 1367223 logs.go:281] Failed to list containers for "coredns": crictl list: sudo crictl ps -a --quiet --name=coredns: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:33:07Z" level=fatal msg="listing containers: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	E0830 23:33:07.520069 1367223 logs.go:281] Failed to list containers for "kube-scheduler": crictl list: sudo crictl ps -a --quiet --name=kube-scheduler: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:33:07Z" level=fatal msg="listing containers: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	E0830 23:33:07.553056 1367223 logs.go:281] Failed to list containers for "kube-proxy": crictl list: sudo crictl ps -a --quiet --name=kube-proxy: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:33:07Z" level=fatal msg="listing containers: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	E0830 23:33:07.582701 1367223 logs.go:281] Failed to list containers for "kube-controller-manager": crictl list: sudo crictl ps -a --quiet --name=kube-controller-manager: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:33:07Z" level=fatal msg="listing containers: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	E0830 23:33:07.611763 1367223 logs.go:281] Failed to list containers for "kindnet": crictl list: sudo crictl ps -a --quiet --name=kindnet: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:33:07Z" level=fatal msg="listing containers: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	E0830 23:33:07.650572 1367223 logs.go:281] Failed to list containers for "storage-provisioner": crictl list: sudo crictl ps -a --quiet --name=storage-provisioner: Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:33:07Z" level=fatal msg="listing containers: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	E0830 23:33:07.968975 1367223 logs.go:195] command /bin/bash -c "sudo `which crictl || echo crictl` ps -a || sudo docker ps -a" failed with error: /bin/bash -c "sudo `which crictl || echo crictl` ps -a || sudo docker ps -a": Process exited with status 1
	stdout:
	
	stderr:
	time="2023-08-30T23:33:07Z" level=fatal msg="listing containers: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService"
	Cannot connect to the Docker daemon at unix:///var/run/docker.sock. Is the docker daemon running?
	 output: "\n** stderr ** \ntime=\"2023-08-30T23:33:07Z\" level=fatal msg=\"listing containers: rpc error: code = Unimplemented desc = unknown service runtime.v1alpha2.RuntimeService\"\nCannot connect to the Docker daemon at unix:///var/run/docker.sock. Is the docker daemon running?\n\n** /stderr **"
	E0830 23:33:08.305762 1367223 logs.go:195] command /bin/bash -c "sudo /var/lib/minikube/binaries/v1.21.2/kubectl describe nodes --kubeconfig=/var/lib/minikube/kubeconfig" failed with error: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.21.2/kubectl describe nodes --kubeconfig=/var/lib/minikube/kubeconfig": Process exited with status 1
	stdout:
	
	stderr:
	The connection to the server localhost:8443 was refused - did you specify the right host or port?
	 output: "\n** stderr ** \nThe connection to the server localhost:8443 was refused - did you specify the right host or port?\n\n** /stderr **"
	! unable to fetch logs for: container status, describe nodes

                                                
                                                
** /stderr **
helpers_test.go:254: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p missing-upgrade-821485 -n missing-upgrade-821485
helpers_test.go:254: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.APIServer}} -p missing-upgrade-821485 -n missing-upgrade-821485: exit status 2 (346.016838ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
helpers_test.go:254: status error: exit status 2 (may be ok)
helpers_test.go:256: "missing-upgrade-821485" apiserver is not running, skipping kubectl commands (state="Stopped")
helpers_test.go:175: Cleaning up "missing-upgrade-821485" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p missing-upgrade-821485
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p missing-upgrade-821485: (2.358289426s)
--- FAIL: TestMissingContainerUpgrade (182.93s)

                                                
                                    

Test pass (267/304)

Order passed test Duration
3 TestDownloadOnly/v1.16.0/json-events 10.52
4 TestDownloadOnly/v1.16.0/preload-exists 0
8 TestDownloadOnly/v1.16.0/LogsDuration 0.16
10 TestDownloadOnly/v1.28.1/json-events 8.93
11 TestDownloadOnly/v1.28.1/preload-exists 0
15 TestDownloadOnly/v1.28.1/LogsDuration 0.08
16 TestDownloadOnly/DeleteAll 0.24
17 TestDownloadOnly/DeleteAlwaysSucceeds 0.15
19 TestBinaryMirror 0.61
22 TestAddons/Setup 141.91
24 TestAddons/parallel/Registry 15.16
27 TestAddons/parallel/MetricsServer 5.91
30 TestAddons/parallel/CSI 58.04
31 TestAddons/parallel/Headlamp 11.42
32 TestAddons/parallel/CloudSpanner 5.71
35 TestAddons/serial/GCPAuth/Namespaces 0.26
36 TestAddons/StoppedEnableDisable 12.59
37 TestCertOptions 38.98
38 TestCertExpiration 229.78
40 TestForceSystemdFlag 43.4
41 TestForceSystemdEnv 43
42 TestDockerEnvContainerd 51.03
47 TestErrorSpam/setup 32.24
48 TestErrorSpam/start 0.87
49 TestErrorSpam/status 1.12
50 TestErrorSpam/pause 1.9
51 TestErrorSpam/unpause 2.09
52 TestErrorSpam/stop 1.52
55 TestFunctional/serial/CopySyncFile 0
56 TestFunctional/serial/StartWithProxy 56.93
57 TestFunctional/serial/AuditLog 0
58 TestFunctional/serial/SoftStart 6.33
59 TestFunctional/serial/KubeContext 0.06
60 TestFunctional/serial/KubectlGetPods 0.11
63 TestFunctional/serial/CacheCmd/cache/add_remote 4.16
64 TestFunctional/serial/CacheCmd/cache/add_local 1.44
65 TestFunctional/serial/CacheCmd/cache/CacheDelete 0.08
66 TestFunctional/serial/CacheCmd/cache/list 0.07
67 TestFunctional/serial/CacheCmd/cache/verify_cache_inside_node 0.35
68 TestFunctional/serial/CacheCmd/cache/cache_reload 2.5
69 TestFunctional/serial/CacheCmd/cache/delete 0.14
70 TestFunctional/serial/MinikubeKubectlCmd 0.15
71 TestFunctional/serial/MinikubeKubectlCmdDirectly 0.16
72 TestFunctional/serial/ExtraConfig 46.11
73 TestFunctional/serial/ComponentHealth 0.11
74 TestFunctional/serial/LogsCmd 1.84
75 TestFunctional/serial/LogsFileCmd 1.94
76 TestFunctional/serial/InvalidService 4.84
78 TestFunctional/parallel/ConfigCmd 0.52
79 TestFunctional/parallel/DashboardCmd 10.33
80 TestFunctional/parallel/DryRun 0.67
81 TestFunctional/parallel/InternationalLanguage 0.3
82 TestFunctional/parallel/StatusCmd 1.55
86 TestFunctional/parallel/ServiceCmdConnect 6.76
87 TestFunctional/parallel/AddonsCmd 0.22
88 TestFunctional/parallel/PersistentVolumeClaim 24.88
90 TestFunctional/parallel/SSHCmd 0.85
91 TestFunctional/parallel/CpCmd 1.61
93 TestFunctional/parallel/FileSync 0.4
94 TestFunctional/parallel/CertSync 2.42
98 TestFunctional/parallel/NodeLabels 0.11
100 TestFunctional/parallel/NonActiveRuntimeDisabled 0.82
102 TestFunctional/parallel/License 0.43
103 TestFunctional/parallel/Version/short 0.09
104 TestFunctional/parallel/Version/components 1.24
105 TestFunctional/parallel/ImageCommands/ImageListShort 0.42
106 TestFunctional/parallel/ImageCommands/ImageListTable 0.34
107 TestFunctional/parallel/ImageCommands/ImageListJson 0.33
108 TestFunctional/parallel/ImageCommands/ImageListYaml 0.33
109 TestFunctional/parallel/ImageCommands/ImageBuild 3.6
110 TestFunctional/parallel/ImageCommands/Setup 2.11
111 TestFunctional/parallel/UpdateContextCmd/no_changes 0.25
112 TestFunctional/parallel/UpdateContextCmd/no_minikube_cluster 0.33
113 TestFunctional/parallel/UpdateContextCmd/no_clusters 0.26
115 TestFunctional/parallel/ServiceCmd/DeployApp 10.35
118 TestFunctional/parallel/ServiceCmd/List 0.44
119 TestFunctional/parallel/ServiceCmd/JSONOutput 0.55
120 TestFunctional/parallel/ServiceCmd/HTTPS 0.54
121 TestFunctional/parallel/ServiceCmd/Format 0.53
123 TestFunctional/parallel/ServiceCmd/URL 0.49
124 TestFunctional/parallel/ImageCommands/ImageRemove 0.68
126 TestFunctional/parallel/ImageCommands/ImageSaveDaemon 0.75
128 TestFunctional/parallel/TunnelCmd/serial/RunSecondTunnel 0.78
129 TestFunctional/parallel/TunnelCmd/serial/StartTunnel 0
131 TestFunctional/parallel/TunnelCmd/serial/WaitService/Setup 8.44
132 TestFunctional/parallel/TunnelCmd/serial/WaitService/IngressIP 0.14
133 TestFunctional/parallel/TunnelCmd/serial/AccessDirect 0
137 TestFunctional/parallel/TunnelCmd/serial/DeleteTunnel 0.11
138 TestFunctional/parallel/ProfileCmd/profile_not_create 0.58
139 TestFunctional/parallel/ProfileCmd/profile_list 0.52
140 TestFunctional/parallel/ProfileCmd/profile_json_output 0.47
141 TestFunctional/parallel/MountCmd/any-port 7.74
142 TestFunctional/parallel/MountCmd/specific-port 2.76
143 TestFunctional/parallel/MountCmd/VerifyCleanup 2.18
144 TestFunctional/delete_addon-resizer_images 0.08
145 TestFunctional/delete_my-image_image 0.02
146 TestFunctional/delete_minikube_cached_images 0.02
150 TestIngressAddonLegacy/StartLegacyK8sCluster 94.83
152 TestIngressAddonLegacy/serial/ValidateIngressAddonActivation 10
153 TestIngressAddonLegacy/serial/ValidateIngressDNSAddonActivation 0.67
157 TestJSONOutput/start/Command 85.64
158 TestJSONOutput/start/Audit 0
160 TestJSONOutput/start/parallel/DistinctCurrentSteps 0
161 TestJSONOutput/start/parallel/IncreasingCurrentSteps 0
163 TestJSONOutput/pause/Command 0.85
164 TestJSONOutput/pause/Audit 0
166 TestJSONOutput/pause/parallel/DistinctCurrentSteps 0
167 TestJSONOutput/pause/parallel/IncreasingCurrentSteps 0
169 TestJSONOutput/unpause/Command 0.74
170 TestJSONOutput/unpause/Audit 0
172 TestJSONOutput/unpause/parallel/DistinctCurrentSteps 0
173 TestJSONOutput/unpause/parallel/IncreasingCurrentSteps 0
175 TestJSONOutput/stop/Command 1.32
176 TestJSONOutput/stop/Audit 0
178 TestJSONOutput/stop/parallel/DistinctCurrentSteps 0
179 TestJSONOutput/stop/parallel/IncreasingCurrentSteps 0
180 TestErrorJSONOutput 0.26
182 TestKicCustomNetwork/create_custom_network 44.9
183 TestKicCustomNetwork/use_default_bridge_network 33.12
184 TestKicExistingNetwork 32.97
185 TestKicCustomSubnet 35.39
186 TestKicStaticIP 36.74
187 TestMainNoArgs 0.07
188 TestMinikubeProfile 73.55
191 TestMountStart/serial/StartWithMountFirst 7.03
192 TestMountStart/serial/VerifyMountFirst 0.29
193 TestMountStart/serial/StartWithMountSecond 9.68
194 TestMountStart/serial/VerifyMountSecond 0.31
195 TestMountStart/serial/DeleteFirst 1.7
196 TestMountStart/serial/VerifyMountPostDelete 0.29
197 TestMountStart/serial/Stop 1.24
198 TestMountStart/serial/RestartStopped 7.66
199 TestMountStart/serial/VerifyMountPostStop 0.3
202 TestMultiNode/serial/FreshStart2Nodes 77.75
203 TestMultiNode/serial/DeployApp2Nodes 4.92
204 TestMultiNode/serial/PingHostFrom2Pods 1.17
205 TestMultiNode/serial/AddNode 21.69
206 TestMultiNode/serial/ProfileList 0.37
207 TestMultiNode/serial/CopyFile 11.52
208 TestMultiNode/serial/StopNode 2.46
209 TestMultiNode/serial/StartAfterStop 13.2
210 TestMultiNode/serial/RestartKeepsNodes 124.83
211 TestMultiNode/serial/DeleteNode 5.48
212 TestMultiNode/serial/StopMultiNode 24.28
213 TestMultiNode/serial/RestartMultiNode 78.55
214 TestMultiNode/serial/ValidateNameConflict 39.73
219 TestPreload 150
221 TestScheduledStopUnix 107.5
224 TestInsufficientStorage 14.56
225 TestRunningBinaryUpgrade 119.3
227 TestKubernetesUpgrade 422.67
231 TestNoKubernetes/serial/StartNoK8sWithVersion 0.1
232 TestPause/serial/Start 97.39
233 TestNoKubernetes/serial/StartWithK8s 44.3
234 TestNoKubernetes/serial/StartWithStopK8s 16.38
235 TestNoKubernetes/serial/Start 6
236 TestNoKubernetes/serial/VerifyK8sNotRunning 0.33
237 TestNoKubernetes/serial/ProfileList 1.03
238 TestNoKubernetes/serial/Stop 1.26
239 TestNoKubernetes/serial/StartNoArgs 6.57
240 TestNoKubernetes/serial/VerifyK8sNotRunningSecond 0.32
248 TestNetworkPlugins/group/false 4.26
252 TestPause/serial/SecondStartNoReconfiguration 6.53
253 TestPause/serial/Pause 0.93
254 TestPause/serial/VerifyStatus 0.46
255 TestPause/serial/Unpause 0.91
256 TestPause/serial/PauseAgain 1.27
257 TestPause/serial/DeletePaused 3.01
258 TestPause/serial/VerifyDeletedResources 0.47
259 TestStoppedBinaryUpgrade/Setup 1.21
260 TestStoppedBinaryUpgrade/Upgrade 178.44
261 TestStoppedBinaryUpgrade/MinikubeLogs 1.73
269 TestNetworkPlugins/group/auto/Start 95.42
270 TestNetworkPlugins/group/auto/KubeletFlags 0.47
271 TestNetworkPlugins/group/auto/NetCatPod 9.61
272 TestNetworkPlugins/group/auto/DNS 0.29
273 TestNetworkPlugins/group/auto/Localhost 0.25
274 TestNetworkPlugins/group/auto/HairPin 0.2
275 TestNetworkPlugins/group/kindnet/Start 91.98
276 TestNetworkPlugins/group/calico/Start 72.97
277 TestNetworkPlugins/group/calico/ControllerPod 5.04
278 TestNetworkPlugins/group/calico/KubeletFlags 0.31
279 TestNetworkPlugins/group/calico/NetCatPod 9.43
280 TestNetworkPlugins/group/calico/DNS 0.24
281 TestNetworkPlugins/group/calico/Localhost 0.21
282 TestNetworkPlugins/group/calico/HairPin 0.21
283 TestNetworkPlugins/group/kindnet/ControllerPod 5.04
284 TestNetworkPlugins/group/kindnet/KubeletFlags 0.47
285 TestNetworkPlugins/group/kindnet/NetCatPod 9.65
286 TestNetworkPlugins/group/kindnet/DNS 0.28
287 TestNetworkPlugins/group/kindnet/Localhost 0.25
288 TestNetworkPlugins/group/kindnet/HairPin 0.26
289 TestNetworkPlugins/group/custom-flannel/Start 63.69
290 TestNetworkPlugins/group/enable-default-cni/Start 55.62
291 TestNetworkPlugins/group/custom-flannel/KubeletFlags 0.33
292 TestNetworkPlugins/group/custom-flannel/NetCatPod 10.43
293 TestNetworkPlugins/group/enable-default-cni/KubeletFlags 0.34
294 TestNetworkPlugins/group/custom-flannel/DNS 0.3
295 TestNetworkPlugins/group/enable-default-cni/NetCatPod 8.54
296 TestNetworkPlugins/group/custom-flannel/Localhost 0.27
297 TestNetworkPlugins/group/custom-flannel/HairPin 0.24
298 TestNetworkPlugins/group/enable-default-cni/DNS 0.29
299 TestNetworkPlugins/group/enable-default-cni/Localhost 0.27
300 TestNetworkPlugins/group/enable-default-cni/HairPin 0.26
301 TestNetworkPlugins/group/flannel/Start 65.98
302 TestNetworkPlugins/group/bridge/Start 52.41
303 TestNetworkPlugins/group/bridge/KubeletFlags 0.42
304 TestNetworkPlugins/group/bridge/NetCatPod 8.4
305 TestNetworkPlugins/group/flannel/ControllerPod 5.03
306 TestNetworkPlugins/group/bridge/DNS 0.21
307 TestNetworkPlugins/group/bridge/Localhost 0.18
308 TestNetworkPlugins/group/bridge/HairPin 0.18
309 TestNetworkPlugins/group/flannel/KubeletFlags 0.45
310 TestNetworkPlugins/group/flannel/NetCatPod 9.35
311 TestNetworkPlugins/group/flannel/DNS 0.25
312 TestNetworkPlugins/group/flannel/Localhost 0.27
313 TestNetworkPlugins/group/flannel/HairPin 0.27
315 TestStartStop/group/old-k8s-version/serial/FirstStart 134.08
317 TestStartStop/group/no-preload/serial/FirstStart 90.41
318 TestStartStop/group/no-preload/serial/DeployApp 8.47
319 TestStartStop/group/no-preload/serial/EnableAddonWhileActive 1.24
320 TestStartStop/group/no-preload/serial/Stop 12.21
321 TestStartStop/group/no-preload/serial/EnableAddonAfterStop 0.22
322 TestStartStop/group/no-preload/serial/SecondStart 347.45
323 TestStartStop/group/old-k8s-version/serial/DeployApp 9.53
324 TestStartStop/group/old-k8s-version/serial/EnableAddonWhileActive 1.23
325 TestStartStop/group/old-k8s-version/serial/Stop 12.47
326 TestStartStop/group/old-k8s-version/serial/EnableAddonAfterStop 0.29
327 TestStartStop/group/old-k8s-version/serial/SecondStart 617.31
328 TestStartStop/group/no-preload/serial/UserAppExistsAfterStop 10.03
329 TestStartStop/group/no-preload/serial/AddonExistsAfterStop 5.12
330 TestStartStop/group/no-preload/serial/VerifyKubernetesImages 0.4
331 TestStartStop/group/no-preload/serial/Pause 4.85
333 TestStartStop/group/embed-certs/serial/FirstStart 85.68
334 TestStartStop/group/embed-certs/serial/DeployApp 9.5
335 TestStartStop/group/embed-certs/serial/EnableAddonWhileActive 1.28
336 TestStartStop/group/embed-certs/serial/Stop 12.19
337 TestStartStop/group/embed-certs/serial/EnableAddonAfterStop 0.21
338 TestStartStop/group/embed-certs/serial/SecondStart 342.19
339 TestStartStop/group/old-k8s-version/serial/UserAppExistsAfterStop 5.02
340 TestStartStop/group/old-k8s-version/serial/AddonExistsAfterStop 5.12
341 TestStartStop/group/old-k8s-version/serial/VerifyKubernetesImages 0.37
342 TestStartStop/group/old-k8s-version/serial/Pause 3.66
344 TestStartStop/group/default-k8s-diff-port/serial/FirstStart 60.29
345 TestStartStop/group/default-k8s-diff-port/serial/DeployApp 8.5
346 TestStartStop/group/default-k8s-diff-port/serial/EnableAddonWhileActive 1.26
347 TestStartStop/group/default-k8s-diff-port/serial/Stop 12.2
348 TestStartStop/group/default-k8s-diff-port/serial/EnableAddonAfterStop 0.22
349 TestStartStop/group/default-k8s-diff-port/serial/SecondStart 342.48
350 TestStartStop/group/embed-certs/serial/UserAppExistsAfterStop 14.03
351 TestStartStop/group/embed-certs/serial/AddonExistsAfterStop 5.12
352 TestStartStop/group/embed-certs/serial/VerifyKubernetesImages 0.38
353 TestStartStop/group/embed-certs/serial/Pause 3.52
355 TestStartStop/group/newest-cni/serial/FirstStart 46.46
356 TestStartStop/group/newest-cni/serial/DeployApp 0
357 TestStartStop/group/newest-cni/serial/EnableAddonWhileActive 1.47
358 TestStartStop/group/newest-cni/serial/Stop 1.32
359 TestStartStop/group/newest-cni/serial/EnableAddonAfterStop 0.23
360 TestStartStop/group/newest-cni/serial/SecondStart 31.95
361 TestStartStop/group/newest-cni/serial/UserAppExistsAfterStop 0
362 TestStartStop/group/newest-cni/serial/AddonExistsAfterStop 0
363 TestStartStop/group/newest-cni/serial/VerifyKubernetesImages 0.38
364 TestStartStop/group/newest-cni/serial/Pause 3.51
365 TestStartStop/group/default-k8s-diff-port/serial/UserAppExistsAfterStop 9.03
366 TestStartStop/group/default-k8s-diff-port/serial/AddonExistsAfterStop 5.14
367 TestStartStop/group/default-k8s-diff-port/serial/VerifyKubernetesImages 0.37
368 TestStartStop/group/default-k8s-diff-port/serial/Pause 3.36
x
+
TestDownloadOnly/v1.16.0/json-events (10.52s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.16.0/json-events
aaa_download_only_test.go:69: (dbg) Run:  out/minikube-linux-arm64 start -o=json --download-only -p download-only-635688 --force --alsologtostderr --kubernetes-version=v1.16.0 --container-runtime=containerd --driver=docker  --container-runtime=containerd
aaa_download_only_test.go:69: (dbg) Done: out/minikube-linux-arm64 start -o=json --download-only -p download-only-635688 --force --alsologtostderr --kubernetes-version=v1.16.0 --container-runtime=containerd --driver=docker  --container-runtime=containerd: (10.51991917s)
--- PASS: TestDownloadOnly/v1.16.0/json-events (10.52s)

                                                
                                    
x
+
TestDownloadOnly/v1.16.0/preload-exists (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.16.0/preload-exists
--- PASS: TestDownloadOnly/v1.16.0/preload-exists (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.16.0/LogsDuration (0.16s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.16.0/LogsDuration
aaa_download_only_test.go:169: (dbg) Run:  out/minikube-linux-arm64 logs -p download-only-635688
aaa_download_only_test.go:169: (dbg) Non-zero exit: out/minikube-linux-arm64 logs -p download-only-635688: exit status 85 (163.192491ms)

                                                
                                                
-- stdout --
	* 
	* ==> Audit <==
	* |---------|--------------------------------|----------------------|---------|---------|---------------------|----------|
	| Command |              Args              |       Profile        |  User   | Version |     Start Time      | End Time |
	|---------|--------------------------------|----------------------|---------|---------|---------------------|----------|
	| start   | -o=json --download-only        | download-only-635688 | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC |          |
	|         | -p download-only-635688        |                      |         |         |                     |          |
	|         | --force --alsologtostderr      |                      |         |         |                     |          |
	|         | --kubernetes-version=v1.16.0   |                      |         |         |                     |          |
	|         | --container-runtime=containerd |                      |         |         |                     |          |
	|         | --driver=docker                |                      |         |         |                     |          |
	|         | --container-runtime=containerd |                      |         |         |                     |          |
	|---------|--------------------------------|----------------------|---------|---------|---------------------|----------|
	
	* 
	* ==> Last Start <==
	* Log file created at: 2023/08/30 22:54:22
	Running on machine: ip-172-31-31-251
	Binary: Built with gc go1.20.7 for linux/arm64
	Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
	I0830 22:54:22.843296 1225380 out.go:296] Setting OutFile to fd 1 ...
	I0830 22:54:22.843510 1225380 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 22:54:22.843539 1225380 out.go:309] Setting ErrFile to fd 2...
	I0830 22:54:22.843559 1225380 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 22:54:22.843854 1225380 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
	W0830 22:54:22.843999 1225380 root.go:314] Error reading config file at /home/jenkins/minikube-integration/17114-1219981/.minikube/config/config.json: open /home/jenkins/minikube-integration/17114-1219981/.minikube/config/config.json: no such file or directory
	I0830 22:54:22.844418 1225380 out.go:303] Setting JSON to true
	I0830 22:54:22.845389 1225380 start.go:128] hostinfo: {"hostname":"ip-172-31-31-251","uptime":27397,"bootTime":1693408666,"procs":186,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1043-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"982e3628-3742-4b3e-bb63-ac1b07660ec7"}
	I0830 22:54:22.845481 1225380 start.go:138] virtualization:  
	I0830 22:54:22.848319 1225380 out.go:97] [download-only-635688] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	I0830 22:54:22.850276 1225380 out.go:169] MINIKUBE_LOCATION=17114
	W0830 22:54:22.848552 1225380 preload.go:295] Failed to list preload files: open /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball: no such file or directory
	I0830 22:54:22.848627 1225380 notify.go:220] Checking for updates...
	I0830 22:54:22.853777 1225380 out.go:169] MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0830 22:54:22.855687 1225380 out.go:169] KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	I0830 22:54:22.857233 1225380 out.go:169] MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	I0830 22:54:22.859046 1225380 out.go:169] MINIKUBE_BIN=out/minikube-linux-arm64
	W0830 22:54:22.862246 1225380 out.go:272] minikube skips various validations when --force is supplied; this may lead to unexpected behavior
	I0830 22:54:22.862514 1225380 driver.go:373] Setting default libvirt URI to qemu:///system
	I0830 22:54:22.886516 1225380 docker.go:121] docker version: linux-24.0.5:Docker Engine - Community
	I0830 22:54:22.886609 1225380 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 22:54:22.965824 1225380 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:28 OomKillDisable:true NGoroutines:42 SystemTime:2023-08-30 22:54:22.955169401 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 22:54:22.965928 1225380 docker.go:294] overlay module found
	I0830 22:54:22.967800 1225380 out.go:97] Using the docker driver based on user configuration
	I0830 22:54:22.967823 1225380 start.go:298] selected driver: docker
	I0830 22:54:22.967829 1225380 start.go:902] validating driver "docker" against <nil>
	I0830 22:54:22.967951 1225380 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 22:54:23.034035 1225380 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:28 OomKillDisable:true NGoroutines:42 SystemTime:2023-08-30 22:54:23.024745339 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 22:54:23.034201 1225380 start_flags.go:305] no existing cluster config was found, will generate one from the flags 
	I0830 22:54:23.034475 1225380 start_flags.go:382] Using suggested 2200MB memory alloc based on sys=7834MB, container=7834MB
	I0830 22:54:23.034628 1225380 start_flags.go:901] Wait components to verify : map[apiserver:true system_pods:true]
	I0830 22:54:23.036794 1225380 out.go:169] Using Docker driver with root privileges
	I0830 22:54:23.038457 1225380 cni.go:84] Creating CNI manager for ""
	I0830 22:54:23.038472 1225380 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
	I0830 22:54:23.038487 1225380 start_flags.go:314] Found "CNI" CNI - setting NetworkPlugin=cni
	I0830 22:54:23.038505 1225380 start_flags.go:319] config:
	{Name:download-only-635688 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.16.0 ClusterName:download-only-635688 Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRu
ntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0}
	I0830 22:54:23.040460 1225380 out.go:97] Starting control plane node download-only-635688 in cluster download-only-635688
	I0830 22:54:23.040494 1225380 cache.go:122] Beginning downloading kic base image for docker with containerd
	I0830 22:54:23.042301 1225380 out.go:97] Pulling base image ...
	I0830 22:54:23.042324 1225380 preload.go:132] Checking if preload exists for k8s version v1.16.0 and runtime containerd
	I0830 22:54:23.042423 1225380 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec in local docker daemon
	I0830 22:54:23.059381 1225380 cache.go:150] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec to local cache
	I0830 22:54:23.059603 1225380 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec in local cache directory
	I0830 22:54:23.059736 1225380 image.go:118] Writing gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec to local cache
	I0830 22:54:23.116015 1225380 preload.go:119] Found remote preload: https://storage.googleapis.com/minikube-preloaded-volume-tarballs/v18/v1.16.0/preloaded-images-k8s-v18-v1.16.0-containerd-overlay2-arm64.tar.lz4
	I0830 22:54:23.116042 1225380 cache.go:57] Caching tarball of preloaded images
	I0830 22:54:23.116206 1225380 preload.go:132] Checking if preload exists for k8s version v1.16.0 and runtime containerd
	I0830 22:54:23.118549 1225380 out.go:97] Downloading Kubernetes v1.16.0 preload ...
	I0830 22:54:23.118574 1225380 preload.go:238] getting checksum for preloaded-images-k8s-v18-v1.16.0-containerd-overlay2-arm64.tar.lz4 ...
	I0830 22:54:23.247814 1225380 download.go:107] Downloading: https://storage.googleapis.com/minikube-preloaded-volume-tarballs/v18/v1.16.0/preloaded-images-k8s-v18-v1.16.0-containerd-overlay2-arm64.tar.lz4?checksum=md5:1f1e2324dbd6e4f3d8734226d9194e9f -> /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.16.0-containerd-overlay2-arm64.tar.lz4
	I0830 22:54:27.527508 1225380 cache.go:153] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec as a tarball
	I0830 22:54:31.705661 1225380 preload.go:249] saving checksum for preloaded-images-k8s-v18-v1.16.0-containerd-overlay2-arm64.tar.lz4 ...
	I0830 22:54:31.705762 1225380 preload.go:256] verifying checksum of /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.16.0-containerd-overlay2-arm64.tar.lz4 ...
	I0830 22:54:32.744119 1225380 cache.go:60] Finished verifying existence of preloaded tar for  v1.16.0 on containerd
	I0830 22:54:32.744482 1225380 profile.go:148] Saving config to /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/download-only-635688/config.json ...
	I0830 22:54:32.744514 1225380 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/download-only-635688/config.json: {Name:mk9c8e9dc59a3ea4032d58f4fcc3b0642b94c548 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0830 22:54:32.744698 1225380 preload.go:132] Checking if preload exists for k8s version v1.16.0 and runtime containerd
	I0830 22:54:32.745460 1225380 download.go:107] Downloading: https://dl.k8s.io/release/v1.16.0/bin/linux/arm64/kubectl?checksum=file:https://dl.k8s.io/release/v1.16.0/bin/linux/arm64/kubectl.sha1 -> /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/linux/arm64/v1.16.0/kubectl
	
	* 
	* The control plane node "" does not exist.
	  To start a cluster, run: "minikube start -p download-only-635688"

                                                
                                                
-- /stdout --
aaa_download_only_test.go:170: minikube logs failed with error: exit status 85
--- PASS: TestDownloadOnly/v1.16.0/LogsDuration (0.16s)

                                                
                                    
x
+
TestDownloadOnly/v1.28.1/json-events (8.93s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.28.1/json-events
aaa_download_only_test.go:69: (dbg) Run:  out/minikube-linux-arm64 start -o=json --download-only -p download-only-635688 --force --alsologtostderr --kubernetes-version=v1.28.1 --container-runtime=containerd --driver=docker  --container-runtime=containerd
aaa_download_only_test.go:69: (dbg) Done: out/minikube-linux-arm64 start -o=json --download-only -p download-only-635688 --force --alsologtostderr --kubernetes-version=v1.28.1 --container-runtime=containerd --driver=docker  --container-runtime=containerd: (8.929134935s)
--- PASS: TestDownloadOnly/v1.28.1/json-events (8.93s)

                                                
                                    
x
+
TestDownloadOnly/v1.28.1/preload-exists (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.28.1/preload-exists
--- PASS: TestDownloadOnly/v1.28.1/preload-exists (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.28.1/LogsDuration (0.08s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.28.1/LogsDuration
aaa_download_only_test.go:169: (dbg) Run:  out/minikube-linux-arm64 logs -p download-only-635688
aaa_download_only_test.go:169: (dbg) Non-zero exit: out/minikube-linux-arm64 logs -p download-only-635688: exit status 85 (82.359277ms)

                                                
                                                
-- stdout --
	* 
	* ==> Audit <==
	* |---------|--------------------------------|----------------------|---------|---------|---------------------|----------|
	| Command |              Args              |       Profile        |  User   | Version |     Start Time      | End Time |
	|---------|--------------------------------|----------------------|---------|---------|---------------------|----------|
	| start   | -o=json --download-only        | download-only-635688 | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC |          |
	|         | -p download-only-635688        |                      |         |         |                     |          |
	|         | --force --alsologtostderr      |                      |         |         |                     |          |
	|         | --kubernetes-version=v1.16.0   |                      |         |         |                     |          |
	|         | --container-runtime=containerd |                      |         |         |                     |          |
	|         | --driver=docker                |                      |         |         |                     |          |
	|         | --container-runtime=containerd |                      |         |         |                     |          |
	| start   | -o=json --download-only        | download-only-635688 | jenkins | v1.31.2 | 30 Aug 23 22:54 UTC |          |
	|         | -p download-only-635688        |                      |         |         |                     |          |
	|         | --force --alsologtostderr      |                      |         |         |                     |          |
	|         | --kubernetes-version=v1.28.1   |                      |         |         |                     |          |
	|         | --container-runtime=containerd |                      |         |         |                     |          |
	|         | --driver=docker                |                      |         |         |                     |          |
	|         | --container-runtime=containerd |                      |         |         |                     |          |
	|---------|--------------------------------|----------------------|---------|---------|---------------------|----------|
	
	* 
	* ==> Last Start <==
	* Log file created at: 2023/08/30 22:54:33
	Running on machine: ip-172-31-31-251
	Binary: Built with gc go1.20.7 for linux/arm64
	Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
	I0830 22:54:33.532655 1225454 out.go:296] Setting OutFile to fd 1 ...
	I0830 22:54:33.532859 1225454 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 22:54:33.532889 1225454 out.go:309] Setting ErrFile to fd 2...
	I0830 22:54:33.532909 1225454 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 22:54:33.533256 1225454 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
	W0830 22:54:33.533406 1225454 root.go:314] Error reading config file at /home/jenkins/minikube-integration/17114-1219981/.minikube/config/config.json: open /home/jenkins/minikube-integration/17114-1219981/.minikube/config/config.json: no such file or directory
	I0830 22:54:33.533661 1225454 out.go:303] Setting JSON to true
	I0830 22:54:33.534562 1225454 start.go:128] hostinfo: {"hostname":"ip-172-31-31-251","uptime":27408,"bootTime":1693408666,"procs":183,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1043-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"982e3628-3742-4b3e-bb63-ac1b07660ec7"}
	I0830 22:54:33.534657 1225454 start.go:138] virtualization:  
	I0830 22:54:33.548109 1225454 out.go:97] [download-only-635688] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	I0830 22:54:33.557519 1225454 out.go:169] MINIKUBE_LOCATION=17114
	I0830 22:54:33.548479 1225454 notify.go:220] Checking for updates...
	I0830 22:54:33.575439 1225454 out.go:169] MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0830 22:54:33.584421 1225454 out.go:169] KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	I0830 22:54:33.594713 1225454 out.go:169] MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	I0830 22:54:33.607292 1225454 out.go:169] MINIKUBE_BIN=out/minikube-linux-arm64
	W0830 22:54:33.624820 1225454 out.go:272] minikube skips various validations when --force is supplied; this may lead to unexpected behavior
	I0830 22:54:33.625349 1225454 config.go:182] Loaded profile config "download-only-635688": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.16.0
	W0830 22:54:33.625469 1225454 start.go:810] api.Load failed for download-only-635688: filestore "download-only-635688": Docker machine "download-only-635688" does not exist. Use "docker-machine ls" to list machines. Use "docker-machine create" to add a new one.
	I0830 22:54:33.625573 1225454 driver.go:373] Setting default libvirt URI to qemu:///system
	W0830 22:54:33.625598 1225454 start.go:810] api.Load failed for download-only-635688: filestore "download-only-635688": Docker machine "download-only-635688" does not exist. Use "docker-machine ls" to list machines. Use "docker-machine create" to add a new one.
	I0830 22:54:33.652588 1225454 docker.go:121] docker version: linux-24.0.5:Docker Engine - Community
	I0830 22:54:33.652681 1225454 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 22:54:33.738617 1225454 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:28 OomKillDisable:true NGoroutines:38 SystemTime:2023-08-30 22:54:33.727974904 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 22:54:33.738722 1225454 docker.go:294] overlay module found
	I0830 22:54:33.764846 1225454 out.go:97] Using the docker driver based on existing profile
	I0830 22:54:33.764882 1225454 start.go:298] selected driver: docker
	I0830 22:54:33.764890 1225454 start.go:902] validating driver "docker" against &{Name:download-only-635688 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.16.0 ClusterName:download-only-635688 Namespace:default APIServerName:
minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.16.0 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnet
ClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0}
	I0830 22:54:33.765122 1225454 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 22:54:33.835468 1225454 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:28 OomKillDisable:true NGoroutines:38 SystemTime:2023-08-30 22:54:33.825983051 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 22:54:33.836024 1225454 cni.go:84] Creating CNI manager for ""
	I0830 22:54:33.836039 1225454 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
	I0830 22:54:33.836050 1225454 start_flags.go:319] config:
	{Name:download-only-635688 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:2200 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.28.1 ClusterName:download-only-635688 Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRu
ntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.16.0 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0}
	I0830 22:54:33.846018 1225454 out.go:97] Starting control plane node download-only-635688 in cluster download-only-635688
	I0830 22:54:33.846072 1225454 cache.go:122] Beginning downloading kic base image for docker with containerd
	I0830 22:54:33.861327 1225454 out.go:97] Pulling base image ...
	I0830 22:54:33.861380 1225454 preload.go:132] Checking if preload exists for k8s version v1.28.1 and runtime containerd
	I0830 22:54:33.861459 1225454 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec in local docker daemon
	I0830 22:54:33.878098 1225454 cache.go:150] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec to local cache
	I0830 22:54:33.878218 1225454 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec in local cache directory
	I0830 22:54:33.878240 1225454 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec in local cache directory, skipping pull
	I0830 22:54:33.878245 1225454 image.go:105] gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec exists in cache, skipping pull
	I0830 22:54:33.878253 1225454 cache.go:153] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec as a tarball
	I0830 22:54:33.917494 1225454 preload.go:119] Found remote preload: https://storage.googleapis.com/minikube-preloaded-volume-tarballs/v18/v1.28.1/preloaded-images-k8s-v18-v1.28.1-containerd-overlay2-arm64.tar.lz4
	I0830 22:54:33.917518 1225454 cache.go:57] Caching tarball of preloaded images
	I0830 22:54:33.917684 1225454 preload.go:132] Checking if preload exists for k8s version v1.28.1 and runtime containerd
	I0830 22:54:33.958519 1225454 out.go:97] Downloading Kubernetes v1.28.1 preload ...
	I0830 22:54:33.958550 1225454 preload.go:238] getting checksum for preloaded-images-k8s-v18-v1.28.1-containerd-overlay2-arm64.tar.lz4 ...
	I0830 22:54:34.080859 1225454 download.go:107] Downloading: https://storage.googleapis.com/minikube-preloaded-volume-tarballs/v18/v1.28.1/preloaded-images-k8s-v18-v1.28.1-containerd-overlay2-arm64.tar.lz4?checksum=md5:a95a45d80ac0b4b5848efd127ce0fe53 -> /home/jenkins/minikube-integration/17114-1219981/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.28.1-containerd-overlay2-arm64.tar.lz4
	
	* 
	* The control plane node "" does not exist.
	  To start a cluster, run: "minikube start -p download-only-635688"

                                                
                                                
-- /stdout --
aaa_download_only_test.go:170: minikube logs failed with error: exit status 85
--- PASS: TestDownloadOnly/v1.28.1/LogsDuration (0.08s)

                                                
                                    
x
+
TestDownloadOnly/DeleteAll (0.24s)

                                                
                                                
=== RUN   TestDownloadOnly/DeleteAll
aaa_download_only_test.go:187: (dbg) Run:  out/minikube-linux-arm64 delete --all
--- PASS: TestDownloadOnly/DeleteAll (0.24s)

                                                
                                    
x
+
TestDownloadOnly/DeleteAlwaysSucceeds (0.15s)

                                                
                                                
=== RUN   TestDownloadOnly/DeleteAlwaysSucceeds
aaa_download_only_test.go:199: (dbg) Run:  out/minikube-linux-arm64 delete -p download-only-635688
--- PASS: TestDownloadOnly/DeleteAlwaysSucceeds (0.15s)

                                                
                                    
x
+
TestBinaryMirror (0.61s)

                                                
                                                
=== RUN   TestBinaryMirror
aaa_download_only_test.go:304: (dbg) Run:  out/minikube-linux-arm64 start --download-only -p binary-mirror-621146 --alsologtostderr --binary-mirror http://127.0.0.1:44437 --driver=docker  --container-runtime=containerd
helpers_test.go:175: Cleaning up "binary-mirror-621146" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p binary-mirror-621146
--- PASS: TestBinaryMirror (0.61s)

                                                
                                    
x
+
TestAddons/Setup (141.91s)

                                                
                                                
=== RUN   TestAddons/Setup
addons_test.go:88: (dbg) Run:  out/minikube-linux-arm64 start -p addons-015166 --wait=true --memory=4000 --alsologtostderr --addons=registry --addons=metrics-server --addons=volumesnapshots --addons=csi-hostpath-driver --addons=gcp-auth --addons=cloud-spanner --addons=inspektor-gadget --driver=docker  --container-runtime=containerd --addons=ingress --addons=ingress-dns
addons_test.go:88: (dbg) Done: out/minikube-linux-arm64 start -p addons-015166 --wait=true --memory=4000 --alsologtostderr --addons=registry --addons=metrics-server --addons=volumesnapshots --addons=csi-hostpath-driver --addons=gcp-auth --addons=cloud-spanner --addons=inspektor-gadget --driver=docker  --container-runtime=containerd --addons=ingress --addons=ingress-dns: (2m21.906070894s)
--- PASS: TestAddons/Setup (141.91s)

                                                
                                    
x
+
TestAddons/parallel/Registry (15.16s)

                                                
                                                
=== RUN   TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/Registry
addons_test.go:306: registry stabilized in 41.987929ms
addons_test.go:308: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-gcmd4" [5548c52c-f180-40e9-8007-9ae9bc562acc] Running
addons_test.go:308: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 5.022670968s
addons_test.go:311: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-r82mq" [e38598dc-2305-4c69-95fc-5496df205f2d] Running
addons_test.go:311: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.016830408s
addons_test.go:316: (dbg) Run:  kubectl --context addons-015166 delete po -l run=registry-test --now
addons_test.go:321: (dbg) Run:  kubectl --context addons-015166 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:321: (dbg) Done: kubectl --context addons-015166 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": (3.702932817s)
addons_test.go:335: (dbg) Run:  out/minikube-linux-arm64 -p addons-015166 ip
2023/08/30 22:57:20 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:364: (dbg) Run:  out/minikube-linux-arm64 -p addons-015166 addons disable registry --alsologtostderr -v=1
addons_test.go:364: (dbg) Done: out/minikube-linux-arm64 -p addons-015166 addons disable registry --alsologtostderr -v=1: (1.028465618s)
--- PASS: TestAddons/parallel/Registry (15.16s)

                                                
                                    
x
+
TestAddons/parallel/MetricsServer (5.91s)

                                                
                                                
=== RUN   TestAddons/parallel/MetricsServer
=== PAUSE TestAddons/parallel/MetricsServer

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/MetricsServer
addons_test.go:383: metrics-server stabilized in 3.7827ms
addons_test.go:385: (dbg) TestAddons/parallel/MetricsServer: waiting 6m0s for pods matching "k8s-app=metrics-server" in namespace "kube-system" ...
helpers_test.go:344: "metrics-server-7c66d45ddc-mdsn9" [41e36ea8-5467-4930-a72f-892ee5740e66] Running
addons_test.go:385: (dbg) TestAddons/parallel/MetricsServer: k8s-app=metrics-server healthy within 5.01672768s
addons_test.go:391: (dbg) Run:  kubectl --context addons-015166 top pods -n kube-system
addons_test.go:408: (dbg) Run:  out/minikube-linux-arm64 -p addons-015166 addons disable metrics-server --alsologtostderr -v=1
--- PASS: TestAddons/parallel/MetricsServer (5.91s)

                                                
                                    
x
+
TestAddons/parallel/CSI (58.04s)

                                                
                                                
=== RUN   TestAddons/parallel/CSI
=== PAUSE TestAddons/parallel/CSI

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/CSI
addons_test.go:537: csi-hostpath-driver pods stabilized in 5.7905ms
addons_test.go:540: (dbg) Run:  kubectl --context addons-015166 create -f testdata/csi-hostpath-driver/pvc.yaml
addons_test.go:545: (dbg) TestAddons/parallel/CSI: waiting 6m0s for pvc "hpvc" in namespace "default" ...
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc -o jsonpath={.status.phase} -n default
addons_test.go:550: (dbg) Run:  kubectl --context addons-015166 create -f testdata/csi-hostpath-driver/pv-pod.yaml
addons_test.go:555: (dbg) TestAddons/parallel/CSI: waiting 6m0s for pods matching "app=task-pv-pod" in namespace "default" ...
helpers_test.go:344: "task-pv-pod" [49b3a806-f215-4cb0-aa88-6644cff09819] Pending
helpers_test.go:344: "task-pv-pod" [49b3a806-f215-4cb0-aa88-6644cff09819] Pending / Ready:ContainersNotReady (containers with unready status: [task-pv-container]) / ContainersReady:ContainersNotReady (containers with unready status: [task-pv-container])
helpers_test.go:344: "task-pv-pod" [49b3a806-f215-4cb0-aa88-6644cff09819] Running
addons_test.go:555: (dbg) TestAddons/parallel/CSI: app=task-pv-pod healthy within 11.041969448s
addons_test.go:560: (dbg) Run:  kubectl --context addons-015166 create -f testdata/csi-hostpath-driver/snapshot.yaml
addons_test.go:565: (dbg) TestAddons/parallel/CSI: waiting 6m0s for volume snapshot "new-snapshot-demo" in namespace "default" ...
helpers_test.go:419: (dbg) Run:  kubectl --context addons-015166 get volumesnapshot new-snapshot-demo -o jsonpath={.status.readyToUse} -n default
helpers_test.go:419: (dbg) Run:  kubectl --context addons-015166 get volumesnapshot new-snapshot-demo -o jsonpath={.status.readyToUse} -n default
addons_test.go:570: (dbg) Run:  kubectl --context addons-015166 delete pod task-pv-pod
addons_test.go:576: (dbg) Run:  kubectl --context addons-015166 delete pvc hpvc
addons_test.go:582: (dbg) Run:  kubectl --context addons-015166 create -f testdata/csi-hostpath-driver/pvc-restore.yaml
addons_test.go:587: (dbg) TestAddons/parallel/CSI: waiting 6m0s for pvc "hpvc-restore" in namespace "default" ...
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-015166 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
addons_test.go:592: (dbg) Run:  kubectl --context addons-015166 create -f testdata/csi-hostpath-driver/pv-pod-restore.yaml
addons_test.go:597: (dbg) TestAddons/parallel/CSI: waiting 6m0s for pods matching "app=task-pv-pod-restore" in namespace "default" ...
helpers_test.go:344: "task-pv-pod-restore" [6b24d1c0-53ae-4f1b-a0cc-17528e535b21] Pending / Ready:ContainersNotReady (containers with unready status: [task-pv-container]) / ContainersReady:ContainersNotReady (containers with unready status: [task-pv-container])
helpers_test.go:344: "task-pv-pod-restore" [6b24d1c0-53ae-4f1b-a0cc-17528e535b21] Running
addons_test.go:597: (dbg) TestAddons/parallel/CSI: app=task-pv-pod-restore healthy within 8.0182478s
addons_test.go:602: (dbg) Run:  kubectl --context addons-015166 delete pod task-pv-pod-restore
addons_test.go:602: (dbg) Done: kubectl --context addons-015166 delete pod task-pv-pod-restore: (1.114942732s)
addons_test.go:606: (dbg) Run:  kubectl --context addons-015166 delete pvc hpvc-restore
addons_test.go:610: (dbg) Run:  kubectl --context addons-015166 delete volumesnapshot new-snapshot-demo
addons_test.go:614: (dbg) Run:  out/minikube-linux-arm64 -p addons-015166 addons disable csi-hostpath-driver --alsologtostderr -v=1
addons_test.go:614: (dbg) Done: out/minikube-linux-arm64 -p addons-015166 addons disable csi-hostpath-driver --alsologtostderr -v=1: (6.898670308s)
addons_test.go:618: (dbg) Run:  out/minikube-linux-arm64 -p addons-015166 addons disable volumesnapshots --alsologtostderr -v=1
--- PASS: TestAddons/parallel/CSI (58.04s)

                                                
                                    
x
+
TestAddons/parallel/Headlamp (11.42s)

                                                
                                                
=== RUN   TestAddons/parallel/Headlamp
=== PAUSE TestAddons/parallel/Headlamp

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/Headlamp
addons_test.go:800: (dbg) Run:  out/minikube-linux-arm64 addons enable headlamp -p addons-015166 --alsologtostderr -v=1
addons_test.go:800: (dbg) Done: out/minikube-linux-arm64 addons enable headlamp -p addons-015166 --alsologtostderr -v=1: (1.342120414s)
addons_test.go:805: (dbg) TestAddons/parallel/Headlamp: waiting 8m0s for pods matching "app.kubernetes.io/name=headlamp" in namespace "headlamp" ...
helpers_test.go:344: "headlamp-699c48fb74-k524x" [859b08e9-f5ac-4ce4-8dce-4208df73a05b] Pending
helpers_test.go:344: "headlamp-699c48fb74-k524x" [859b08e9-f5ac-4ce4-8dce-4208df73a05b] Pending / Ready:ContainersNotReady (containers with unready status: [headlamp]) / ContainersReady:ContainersNotReady (containers with unready status: [headlamp])
helpers_test.go:344: "headlamp-699c48fb74-k524x" [859b08e9-f5ac-4ce4-8dce-4208df73a05b] Running
addons_test.go:805: (dbg) TestAddons/parallel/Headlamp: app.kubernetes.io/name=headlamp healthy within 10.073202888s
--- PASS: TestAddons/parallel/Headlamp (11.42s)

                                                
                                    
x
+
TestAddons/parallel/CloudSpanner (5.71s)

                                                
                                                
=== RUN   TestAddons/parallel/CloudSpanner
=== PAUSE TestAddons/parallel/CloudSpanner

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/CloudSpanner
addons_test.go:833: (dbg) TestAddons/parallel/CloudSpanner: waiting 6m0s for pods matching "app=cloud-spanner-emulator" in namespace "default" ...
helpers_test.go:344: "cloud-spanner-emulator-6dcc56475c-2zgfz" [d265d563-fa3c-4a8b-ba45-88e70a5e9138] Running
addons_test.go:833: (dbg) TestAddons/parallel/CloudSpanner: app=cloud-spanner-emulator healthy within 5.027295448s
addons_test.go:836: (dbg) Run:  out/minikube-linux-arm64 addons disable cloud-spanner -p addons-015166
--- PASS: TestAddons/parallel/CloudSpanner (5.71s)

                                                
                                    
x
+
TestAddons/serial/GCPAuth/Namespaces (0.26s)

                                                
                                                
=== RUN   TestAddons/serial/GCPAuth/Namespaces
addons_test.go:626: (dbg) Run:  kubectl --context addons-015166 create ns new-namespace
addons_test.go:640: (dbg) Run:  kubectl --context addons-015166 get secret gcp-auth -n new-namespace
--- PASS: TestAddons/serial/GCPAuth/Namespaces (0.26s)

                                                
                                    
x
+
TestAddons/StoppedEnableDisable (12.59s)

                                                
                                                
=== RUN   TestAddons/StoppedEnableDisable
addons_test.go:148: (dbg) Run:  out/minikube-linux-arm64 stop -p addons-015166
addons_test.go:148: (dbg) Done: out/minikube-linux-arm64 stop -p addons-015166: (12.278227565s)
addons_test.go:152: (dbg) Run:  out/minikube-linux-arm64 addons enable dashboard -p addons-015166
addons_test.go:156: (dbg) Run:  out/minikube-linux-arm64 addons disable dashboard -p addons-015166
addons_test.go:161: (dbg) Run:  out/minikube-linux-arm64 addons disable gvisor -p addons-015166
--- PASS: TestAddons/StoppedEnableDisable (12.59s)

                                                
                                    
x
+
TestCertOptions (38.98s)

                                                
                                                
=== RUN   TestCertOptions
=== PAUSE TestCertOptions

                                                
                                                

                                                
                                                
=== CONT  TestCertOptions
cert_options_test.go:49: (dbg) Run:  out/minikube-linux-arm64 start -p cert-options-372424 --memory=2048 --apiserver-ips=127.0.0.1 --apiserver-ips=192.168.15.15 --apiserver-names=localhost --apiserver-names=www.google.com --apiserver-port=8555 --driver=docker  --container-runtime=containerd
E0830 23:26:47.355759 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
cert_options_test.go:49: (dbg) Done: out/minikube-linux-arm64 start -p cert-options-372424 --memory=2048 --apiserver-ips=127.0.0.1 --apiserver-ips=192.168.15.15 --apiserver-names=localhost --apiserver-names=www.google.com --apiserver-port=8555 --driver=docker  --container-runtime=containerd: (36.279026457s)
cert_options_test.go:60: (dbg) Run:  out/minikube-linux-arm64 -p cert-options-372424 ssh "openssl x509 -text -noout -in /var/lib/minikube/certs/apiserver.crt"
cert_options_test.go:88: (dbg) Run:  kubectl --context cert-options-372424 config view
cert_options_test.go:100: (dbg) Run:  out/minikube-linux-arm64 ssh -p cert-options-372424 -- "sudo cat /etc/kubernetes/admin.conf"
helpers_test.go:175: Cleaning up "cert-options-372424" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p cert-options-372424
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p cert-options-372424: (1.992672386s)
--- PASS: TestCertOptions (38.98s)

                                                
                                    
x
+
TestCertExpiration (229.78s)

                                                
                                                
=== RUN   TestCertExpiration
=== PAUSE TestCertExpiration

                                                
                                                

                                                
                                                
=== CONT  TestCertExpiration
cert_options_test.go:123: (dbg) Run:  out/minikube-linux-arm64 start -p cert-expiration-040323 --memory=2048 --cert-expiration=3m --driver=docker  --container-runtime=containerd
cert_options_test.go:123: (dbg) Done: out/minikube-linux-arm64 start -p cert-expiration-040323 --memory=2048 --cert-expiration=3m --driver=docker  --container-runtime=containerd: (37.982300279s)
E0830 23:27:06.148296 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
cert_options_test.go:131: (dbg) Run:  out/minikube-linux-arm64 start -p cert-expiration-040323 --memory=2048 --cert-expiration=8760h --driver=docker  --container-runtime=containerd
cert_options_test.go:131: (dbg) Done: out/minikube-linux-arm64 start -p cert-expiration-040323 --memory=2048 --cert-expiration=8760h --driver=docker  --container-runtime=containerd: (9.15591095s)
helpers_test.go:175: Cleaning up "cert-expiration-040323" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p cert-expiration-040323
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p cert-expiration-040323: (2.636976686s)
--- PASS: TestCertExpiration (229.78s)

                                                
                                    
x
+
TestForceSystemdFlag (43.4s)

                                                
                                                
=== RUN   TestForceSystemdFlag
=== PAUSE TestForceSystemdFlag

                                                
                                                

                                                
                                                
=== CONT  TestForceSystemdFlag
docker_test.go:91: (dbg) Run:  out/minikube-linux-arm64 start -p force-systemd-flag-488784 --memory=2048 --force-systemd --alsologtostderr -v=5 --driver=docker  --container-runtime=containerd
docker_test.go:91: (dbg) Done: out/minikube-linux-arm64 start -p force-systemd-flag-488784 --memory=2048 --force-systemd --alsologtostderr -v=5 --driver=docker  --container-runtime=containerd: (40.795508081s)
docker_test.go:121: (dbg) Run:  out/minikube-linux-arm64 -p force-systemd-flag-488784 ssh "cat /etc/containerd/config.toml"
helpers_test.go:175: Cleaning up "force-systemd-flag-488784" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p force-systemd-flag-488784
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p force-systemd-flag-488784: (2.133479134s)
--- PASS: TestForceSystemdFlag (43.40s)

                                                
                                    
x
+
TestForceSystemdEnv (43s)

                                                
                                                
=== RUN   TestForceSystemdEnv
=== PAUSE TestForceSystemdEnv

                                                
                                                

                                                
                                                
=== CONT  TestForceSystemdEnv
docker_test.go:155: (dbg) Run:  out/minikube-linux-arm64 start -p force-systemd-env-213826 --memory=2048 --alsologtostderr -v=5 --driver=docker  --container-runtime=containerd
docker_test.go:155: (dbg) Done: out/minikube-linux-arm64 start -p force-systemd-env-213826 --memory=2048 --alsologtostderr -v=5 --driver=docker  --container-runtime=containerd: (40.452012868s)
docker_test.go:121: (dbg) Run:  out/minikube-linux-arm64 -p force-systemd-env-213826 ssh "cat /etc/containerd/config.toml"
helpers_test.go:175: Cleaning up "force-systemd-env-213826" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p force-systemd-env-213826
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p force-systemd-env-213826: (2.110655441s)
--- PASS: TestForceSystemdEnv (43.00s)

                                                
                                    
x
+
TestDockerEnvContainerd (51.03s)

                                                
                                                
=== RUN   TestDockerEnvContainerd
docker_test.go:170: running with containerd true linux arm64
docker_test.go:181: (dbg) Run:  out/minikube-linux-arm64 start -p dockerenv-006199 --driver=docker  --container-runtime=containerd
docker_test.go:181: (dbg) Done: out/minikube-linux-arm64 start -p dockerenv-006199 --driver=docker  --container-runtime=containerd: (34.359207614s)
docker_test.go:189: (dbg) Run:  /bin/bash -c "out/minikube-linux-arm64 docker-env --ssh-host --ssh-add -p dockerenv-006199"
docker_test.go:189: (dbg) Done: /bin/bash -c "out/minikube-linux-arm64 docker-env --ssh-host --ssh-add -p dockerenv-006199": (1.552128556s)
docker_test.go:220: (dbg) Run:  /bin/bash -c "SSH_AUTH_SOCK="/tmp/ssh-hsjsdfnMqJvA/agent.1242267" SSH_AGENT_PID="1242273" DOCKER_HOST=ssh://docker@127.0.0.1:34319 docker version"
docker_test.go:243: (dbg) Run:  /bin/bash -c "SSH_AUTH_SOCK="/tmp/ssh-hsjsdfnMqJvA/agent.1242267" SSH_AGENT_PID="1242273" DOCKER_HOST=ssh://docker@127.0.0.1:34319 DOCKER_BUILDKIT=0 docker build -t local/minikube-dockerenv-containerd-test:latest testdata/docker-env"
docker_test.go:243: (dbg) Done: /bin/bash -c "SSH_AUTH_SOCK="/tmp/ssh-hsjsdfnMqJvA/agent.1242267" SSH_AGENT_PID="1242273" DOCKER_HOST=ssh://docker@127.0.0.1:34319 DOCKER_BUILDKIT=0 docker build -t local/minikube-dockerenv-containerd-test:latest testdata/docker-env": (1.615388971s)
docker_test.go:250: (dbg) Run:  /bin/bash -c "SSH_AUTH_SOCK="/tmp/ssh-hsjsdfnMqJvA/agent.1242267" SSH_AGENT_PID="1242273" DOCKER_HOST=ssh://docker@127.0.0.1:34319 docker image ls"
helpers_test.go:175: Cleaning up "dockerenv-006199" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p dockerenv-006199
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p dockerenv-006199: (2.352455959s)
--- PASS: TestDockerEnvContainerd (51.03s)

                                                
                                    
x
+
TestErrorSpam/setup (32.24s)

                                                
                                                
=== RUN   TestErrorSpam/setup
error_spam_test.go:81: (dbg) Run:  out/minikube-linux-arm64 start -p nospam-563088 -n=1 --memory=2250 --wait=false --log_dir=/tmp/nospam-563088 --driver=docker  --container-runtime=containerd
error_spam_test.go:81: (dbg) Done: out/minikube-linux-arm64 start -p nospam-563088 -n=1 --memory=2250 --wait=false --log_dir=/tmp/nospam-563088 --driver=docker  --container-runtime=containerd: (32.240861542s)
--- PASS: TestErrorSpam/setup (32.24s)

                                                
                                    
x
+
TestErrorSpam/start (0.87s)

                                                
                                                
=== RUN   TestErrorSpam/start
error_spam_test.go:216: Cleaning up 1 logfile(s) ...
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 start --dry-run
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 start --dry-run
error_spam_test.go:182: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 start --dry-run
--- PASS: TestErrorSpam/start (0.87s)

                                                
                                    
x
+
TestErrorSpam/status (1.12s)

                                                
                                                
=== RUN   TestErrorSpam/status
error_spam_test.go:216: Cleaning up 0 logfile(s) ...
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 status
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 status
error_spam_test.go:182: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 status
--- PASS: TestErrorSpam/status (1.12s)

                                                
                                    
x
+
TestErrorSpam/pause (1.9s)

                                                
                                                
=== RUN   TestErrorSpam/pause
error_spam_test.go:216: Cleaning up 0 logfile(s) ...
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 pause
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 pause
error_spam_test.go:182: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 pause
--- PASS: TestErrorSpam/pause (1.90s)

                                                
                                    
x
+
TestErrorSpam/unpause (2.09s)

                                                
                                                
=== RUN   TestErrorSpam/unpause
error_spam_test.go:216: Cleaning up 0 logfile(s) ...
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 unpause
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 unpause
error_spam_test.go:182: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 unpause
--- PASS: TestErrorSpam/unpause (2.09s)

                                                
                                    
x
+
TestErrorSpam/stop (1.52s)

                                                
                                                
=== RUN   TestErrorSpam/stop
error_spam_test.go:216: Cleaning up 0 logfile(s) ...
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 stop
error_spam_test.go:159: (dbg) Done: out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 stop: (1.285109342s)
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 stop
error_spam_test.go:182: (dbg) Run:  out/minikube-linux-arm64 -p nospam-563088 --log_dir /tmp/nospam-563088 stop
--- PASS: TestErrorSpam/stop (1.52s)

                                                
                                    
x
+
TestFunctional/serial/CopySyncFile (0s)

                                                
                                                
=== RUN   TestFunctional/serial/CopySyncFile
functional_test.go:1851: local sync path: /home/jenkins/minikube-integration/17114-1219981/.minikube/files/etc/test/nested/copy/1225375/hosts
--- PASS: TestFunctional/serial/CopySyncFile (0.00s)

                                                
                                    
x
+
TestFunctional/serial/StartWithProxy (56.93s)

                                                
                                                
=== RUN   TestFunctional/serial/StartWithProxy
functional_test.go:2230: (dbg) Run:  out/minikube-linux-arm64 start -p functional-479614 --memory=4000 --apiserver-port=8441 --wait=all --driver=docker  --container-runtime=containerd
functional_test.go:2230: (dbg) Done: out/minikube-linux-arm64 start -p functional-479614 --memory=4000 --apiserver-port=8441 --wait=all --driver=docker  --container-runtime=containerd: (56.932043695s)
--- PASS: TestFunctional/serial/StartWithProxy (56.93s)

                                                
                                    
x
+
TestFunctional/serial/AuditLog (0s)

                                                
                                                
=== RUN   TestFunctional/serial/AuditLog
--- PASS: TestFunctional/serial/AuditLog (0.00s)

                                                
                                    
x
+
TestFunctional/serial/SoftStart (6.33s)

                                                
                                                
=== RUN   TestFunctional/serial/SoftStart
functional_test.go:655: (dbg) Run:  out/minikube-linux-arm64 start -p functional-479614 --alsologtostderr -v=8
functional_test.go:655: (dbg) Done: out/minikube-linux-arm64 start -p functional-479614 --alsologtostderr -v=8: (6.328277814s)
functional_test.go:659: soft start took 6.328760714s for "functional-479614" cluster.
--- PASS: TestFunctional/serial/SoftStart (6.33s)

                                                
                                    
x
+
TestFunctional/serial/KubeContext (0.06s)

                                                
                                                
=== RUN   TestFunctional/serial/KubeContext
functional_test.go:677: (dbg) Run:  kubectl config current-context
--- PASS: TestFunctional/serial/KubeContext (0.06s)

                                                
                                    
x
+
TestFunctional/serial/KubectlGetPods (0.11s)

                                                
                                                
=== RUN   TestFunctional/serial/KubectlGetPods
functional_test.go:692: (dbg) Run:  kubectl --context functional-479614 get po -A
--- PASS: TestFunctional/serial/KubectlGetPods (0.11s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/add_remote (4.16s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/add_remote
functional_test.go:1045: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 cache add registry.k8s.io/pause:3.1
functional_test.go:1045: (dbg) Done: out/minikube-linux-arm64 -p functional-479614 cache add registry.k8s.io/pause:3.1: (1.408101029s)
functional_test.go:1045: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 cache add registry.k8s.io/pause:3.3
functional_test.go:1045: (dbg) Done: out/minikube-linux-arm64 -p functional-479614 cache add registry.k8s.io/pause:3.3: (1.443785858s)
functional_test.go:1045: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 cache add registry.k8s.io/pause:latest
functional_test.go:1045: (dbg) Done: out/minikube-linux-arm64 -p functional-479614 cache add registry.k8s.io/pause:latest: (1.304712165s)
--- PASS: TestFunctional/serial/CacheCmd/cache/add_remote (4.16s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/add_local (1.44s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/add_local
functional_test.go:1073: (dbg) Run:  docker build -t minikube-local-cache-test:functional-479614 /tmp/TestFunctionalserialCacheCmdcacheadd_local3478393342/001
functional_test.go:1085: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 cache add minikube-local-cache-test:functional-479614
functional_test.go:1090: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 cache delete minikube-local-cache-test:functional-479614
functional_test.go:1079: (dbg) Run:  docker rmi minikube-local-cache-test:functional-479614
--- PASS: TestFunctional/serial/CacheCmd/cache/add_local (1.44s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/CacheDelete (0.08s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/CacheDelete
functional_test.go:1098: (dbg) Run:  out/minikube-linux-arm64 cache delete registry.k8s.io/pause:3.3
--- PASS: TestFunctional/serial/CacheCmd/cache/CacheDelete (0.08s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/list (0.07s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/list
functional_test.go:1106: (dbg) Run:  out/minikube-linux-arm64 cache list
--- PASS: TestFunctional/serial/CacheCmd/cache/list (0.07s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/verify_cache_inside_node (0.35s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/verify_cache_inside_node
functional_test.go:1120: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh sudo crictl images
--- PASS: TestFunctional/serial/CacheCmd/cache/verify_cache_inside_node (0.35s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/cache_reload (2.5s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/cache_reload
functional_test.go:1143: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh sudo crictl rmi registry.k8s.io/pause:latest
functional_test.go:1149: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh sudo crictl inspecti registry.k8s.io/pause:latest
functional_test.go:1149: (dbg) Non-zero exit: out/minikube-linux-arm64 -p functional-479614 ssh sudo crictl inspecti registry.k8s.io/pause:latest: exit status 1 (336.965457ms)

                                                
                                                
-- stdout --
	FATA[0000] no such image "registry.k8s.io/pause:latest" present 

                                                
                                                
-- /stdout --
** stderr ** 
	ssh: Process exited with status 1

                                                
                                                
** /stderr **
functional_test.go:1154: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 cache reload
functional_test.go:1154: (dbg) Done: out/minikube-linux-arm64 -p functional-479614 cache reload: (1.437005635s)
functional_test.go:1159: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh sudo crictl inspecti registry.k8s.io/pause:latest
--- PASS: TestFunctional/serial/CacheCmd/cache/cache_reload (2.50s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/delete (0.14s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/delete
functional_test.go:1168: (dbg) Run:  out/minikube-linux-arm64 cache delete registry.k8s.io/pause:3.1
functional_test.go:1168: (dbg) Run:  out/minikube-linux-arm64 cache delete registry.k8s.io/pause:latest
--- PASS: TestFunctional/serial/CacheCmd/cache/delete (0.14s)

                                                
                                    
x
+
TestFunctional/serial/MinikubeKubectlCmd (0.15s)

                                                
                                                
=== RUN   TestFunctional/serial/MinikubeKubectlCmd
functional_test.go:712: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 kubectl -- --context functional-479614 get pods
--- PASS: TestFunctional/serial/MinikubeKubectlCmd (0.15s)

                                                
                                    
x
+
TestFunctional/serial/MinikubeKubectlCmdDirectly (0.16s)

                                                
                                                
=== RUN   TestFunctional/serial/MinikubeKubectlCmdDirectly
functional_test.go:737: (dbg) Run:  out/kubectl --context functional-479614 get pods
--- PASS: TestFunctional/serial/MinikubeKubectlCmdDirectly (0.16s)

                                                
                                    
x
+
TestFunctional/serial/ExtraConfig (46.11s)

                                                
                                                
=== RUN   TestFunctional/serial/ExtraConfig
functional_test.go:753: (dbg) Run:  out/minikube-linux-arm64 start -p functional-479614 --extra-config=apiserver.enable-admission-plugins=NamespaceAutoProvision --wait=all
E0830 23:02:06.147813 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:02:06.153442 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:02:06.163656 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:02:06.183925 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:02:06.224156 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:02:06.304444 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:02:06.464796 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:02:06.785408 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:02:07.426356 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:02:08.706564 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:02:11.266867 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:02:16.387650 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:02:26.627822 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
functional_test.go:753: (dbg) Done: out/minikube-linux-arm64 start -p functional-479614 --extra-config=apiserver.enable-admission-plugins=NamespaceAutoProvision --wait=all: (46.112626603s)
functional_test.go:757: restart took 46.112732623s for "functional-479614" cluster.
--- PASS: TestFunctional/serial/ExtraConfig (46.11s)

                                                
                                    
x
+
TestFunctional/serial/ComponentHealth (0.11s)

                                                
                                                
=== RUN   TestFunctional/serial/ComponentHealth
functional_test.go:806: (dbg) Run:  kubectl --context functional-479614 get po -l tier=control-plane -n kube-system -o=json
functional_test.go:821: etcd phase: Running
functional_test.go:831: etcd status: Ready
functional_test.go:821: kube-apiserver phase: Running
functional_test.go:831: kube-apiserver status: Ready
functional_test.go:821: kube-controller-manager phase: Running
functional_test.go:831: kube-controller-manager status: Ready
functional_test.go:821: kube-scheduler phase: Running
functional_test.go:831: kube-scheduler status: Ready
--- PASS: TestFunctional/serial/ComponentHealth (0.11s)

                                                
                                    
x
+
TestFunctional/serial/LogsCmd (1.84s)

                                                
                                                
=== RUN   TestFunctional/serial/LogsCmd
functional_test.go:1232: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 logs
functional_test.go:1232: (dbg) Done: out/minikube-linux-arm64 -p functional-479614 logs: (1.843910444s)
--- PASS: TestFunctional/serial/LogsCmd (1.84s)

                                                
                                    
x
+
TestFunctional/serial/LogsFileCmd (1.94s)

                                                
                                                
=== RUN   TestFunctional/serial/LogsFileCmd
functional_test.go:1246: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 logs --file /tmp/TestFunctionalserialLogsFileCmd2917554865/001/logs.txt
functional_test.go:1246: (dbg) Done: out/minikube-linux-arm64 -p functional-479614 logs --file /tmp/TestFunctionalserialLogsFileCmd2917554865/001/logs.txt: (1.940045374s)
--- PASS: TestFunctional/serial/LogsFileCmd (1.94s)

                                                
                                    
x
+
TestFunctional/serial/InvalidService (4.84s)

                                                
                                                
=== RUN   TestFunctional/serial/InvalidService
functional_test.go:2317: (dbg) Run:  kubectl --context functional-479614 apply -f testdata/invalidsvc.yaml
functional_test.go:2331: (dbg) Run:  out/minikube-linux-arm64 service invalid-svc -p functional-479614
functional_test.go:2331: (dbg) Non-zero exit: out/minikube-linux-arm64 service invalid-svc -p functional-479614: exit status 115 (422.56787ms)

                                                
                                                
-- stdout --
	|-----------|-------------|-------------|---------------------------|
	| NAMESPACE |    NAME     | TARGET PORT |            URL            |
	|-----------|-------------|-------------|---------------------------|
	| default   | invalid-svc |          80 | http://192.168.49.2:31764 |
	|-----------|-------------|-------------|---------------------------|
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	X Exiting due to SVC_UNREACHABLE: service not available: no running pod for service invalid-svc found
	* 
	╭─────────────────────────────────────────────────────────────────────────────────────────────╮
	│                                                                                             │
	│    * If the above advice does not help, please let us know:                                 │
	│      https://github.com/kubernetes/minikube/issues/new/choose                               │
	│                                                                                             │
	│    * Please run `minikube logs --file=logs.txt` and attach logs.txt to the GitHub issue.    │
	│    * Please also attach the following file to the GitHub issue:                             │
	│    * - /tmp/minikube_service_96b204199e3191fa1740d4430b018a3c8028d52d_0.log                 │
	│                                                                                             │
	╰─────────────────────────────────────────────────────────────────────────────────────────────╯

                                                
                                                
** /stderr **
functional_test.go:2323: (dbg) Run:  kubectl --context functional-479614 delete -f testdata/invalidsvc.yaml
functional_test.go:2323: (dbg) Done: kubectl --context functional-479614 delete -f testdata/invalidsvc.yaml: (1.095801612s)
--- PASS: TestFunctional/serial/InvalidService (4.84s)

                                                
                                    
x
+
TestFunctional/parallel/ConfigCmd (0.52s)

                                                
                                                
=== RUN   TestFunctional/parallel/ConfigCmd
=== PAUSE TestFunctional/parallel/ConfigCmd

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ConfigCmd
functional_test.go:1195: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 config unset cpus
functional_test.go:1195: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 config get cpus
functional_test.go:1195: (dbg) Non-zero exit: out/minikube-linux-arm64 -p functional-479614 config get cpus: exit status 14 (80.488331ms)

                                                
                                                
** stderr ** 
	Error: specified key could not be found in config

                                                
                                                
** /stderr **
functional_test.go:1195: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 config set cpus 2
functional_test.go:1195: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 config get cpus
functional_test.go:1195: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 config unset cpus
functional_test.go:1195: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 config get cpus
functional_test.go:1195: (dbg) Non-zero exit: out/minikube-linux-arm64 -p functional-479614 config get cpus: exit status 14 (85.675653ms)

                                                
                                                
** stderr ** 
	Error: specified key could not be found in config

                                                
                                                
** /stderr **
--- PASS: TestFunctional/parallel/ConfigCmd (0.52s)

                                                
                                    
x
+
TestFunctional/parallel/DashboardCmd (10.33s)

                                                
                                                
=== RUN   TestFunctional/parallel/DashboardCmd
=== PAUSE TestFunctional/parallel/DashboardCmd

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/DashboardCmd
functional_test.go:901: (dbg) daemon: [out/minikube-linux-arm64 dashboard --url --port 36195 -p functional-479614 --alsologtostderr -v=1]
functional_test.go:906: (dbg) stopping [out/minikube-linux-arm64 dashboard --url --port 36195 -p functional-479614 --alsologtostderr -v=1] ...
helpers_test.go:508: unable to kill pid 1257311: os: process already finished
--- PASS: TestFunctional/parallel/DashboardCmd (10.33s)

                                                
                                    
x
+
TestFunctional/parallel/DryRun (0.67s)

                                                
                                                
=== RUN   TestFunctional/parallel/DryRun
=== PAUSE TestFunctional/parallel/DryRun

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/DryRun
functional_test.go:970: (dbg) Run:  out/minikube-linux-arm64 start -p functional-479614 --dry-run --memory 250MB --alsologtostderr --driver=docker  --container-runtime=containerd
functional_test.go:970: (dbg) Non-zero exit: out/minikube-linux-arm64 start -p functional-479614 --dry-run --memory 250MB --alsologtostderr --driver=docker  --container-runtime=containerd: exit status 23 (247.260495ms)

                                                
                                                
-- stdout --
	* [functional-479614] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	  - MINIKUBE_LOCATION=17114
	  - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	  - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	  - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	  - MINIKUBE_BIN=out/minikube-linux-arm64
	  - MINIKUBE_FORCE_SYSTEMD=
	* Using the docker driver based on existing profile
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	I0830 23:03:23.075449 1256748 out.go:296] Setting OutFile to fd 1 ...
	I0830 23:03:23.075680 1256748 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:03:23.075706 1256748 out.go:309] Setting ErrFile to fd 2...
	I0830 23:03:23.075726 1256748 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:03:23.076030 1256748 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
	I0830 23:03:23.076442 1256748 out.go:303] Setting JSON to false
	I0830 23:03:23.077781 1256748 start.go:128] hostinfo: {"hostname":"ip-172-31-31-251","uptime":27938,"bootTime":1693408666,"procs":404,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1043-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"982e3628-3742-4b3e-bb63-ac1b07660ec7"}
	I0830 23:03:23.077907 1256748 start.go:138] virtualization:  
	I0830 23:03:23.080534 1256748 out.go:177] * [functional-479614] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	I0830 23:03:23.082981 1256748 out.go:177]   - MINIKUBE_LOCATION=17114
	I0830 23:03:23.083121 1256748 notify.go:220] Checking for updates...
	I0830 23:03:23.085549 1256748 out.go:177]   - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0830 23:03:23.087382 1256748 out.go:177]   - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	I0830 23:03:23.089087 1256748 out.go:177]   - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	I0830 23:03:23.091198 1256748 out.go:177]   - MINIKUBE_BIN=out/minikube-linux-arm64
	I0830 23:03:23.092872 1256748 out.go:177]   - MINIKUBE_FORCE_SYSTEMD=
	I0830 23:03:23.095402 1256748 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
	I0830 23:03:23.095941 1256748 driver.go:373] Setting default libvirt URI to qemu:///system
	I0830 23:03:23.127844 1256748 docker.go:121] docker version: linux-24.0.5:Docker Engine - Community
	I0830 23:03:23.127937 1256748 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 23:03:23.245352 1256748 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:1 ContainersRunning:1 ContainersPaused:0 ContainersStopped:0 Images:4 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:31 OomKillDisable:true NGoroutines:46 SystemTime:2023-08-30 23:03:23.232145147 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 23:03:23.245504 1256748 docker.go:294] overlay module found
	I0830 23:03:23.247445 1256748 out.go:177] * Using the docker driver based on existing profile
	I0830 23:03:23.249229 1256748 start.go:298] selected driver: docker
	I0830 23:03:23.249257 1256748 start.go:902] validating driver "docker" against &{Name:functional-479614 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:4000 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.28.1 ClusterName:functional-479614 Namespace:default APIServerName:miniku
beCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[{Component:apiserver Key:enable-admission-plugins Value:NamespaceAutoProvision}] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8441 NodeName:} Nodes:[{Name: IP:192.168.49.2 Port:8441 KubernetesVersion:v1.28.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersi
on:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0}
	I0830 23:03:23.249389 1256748 start.go:913] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
	I0830 23:03:23.251676 1256748 out.go:177] 
	W0830 23:03:23.253522 1256748 out.go:239] X Exiting due to RSRC_INSUFFICIENT_REQ_MEMORY: Requested memory allocation 250MiB is less than the usable minimum of 1800MB
	X Exiting due to RSRC_INSUFFICIENT_REQ_MEMORY: Requested memory allocation 250MiB is less than the usable minimum of 1800MB
	I0830 23:03:23.255326 1256748 out.go:177] 

                                                
                                                
** /stderr **
functional_test.go:987: (dbg) Run:  out/minikube-linux-arm64 start -p functional-479614 --dry-run --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd
--- PASS: TestFunctional/parallel/DryRun (0.67s)

                                                
                                    
x
+
TestFunctional/parallel/InternationalLanguage (0.3s)

                                                
                                                
=== RUN   TestFunctional/parallel/InternationalLanguage
=== PAUSE TestFunctional/parallel/InternationalLanguage

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/InternationalLanguage
functional_test.go:1016: (dbg) Run:  out/minikube-linux-arm64 start -p functional-479614 --dry-run --memory 250MB --alsologtostderr --driver=docker  --container-runtime=containerd
functional_test.go:1016: (dbg) Non-zero exit: out/minikube-linux-arm64 start -p functional-479614 --dry-run --memory 250MB --alsologtostderr --driver=docker  --container-runtime=containerd: exit status 23 (302.453815ms)

                                                
                                                
-- stdout --
	* [functional-479614] minikube v1.31.2 sur Ubuntu 20.04 (arm64)
	  - MINIKUBE_LOCATION=17114
	  - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	  - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	  - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	  - MINIKUBE_BIN=out/minikube-linux-arm64
	  - MINIKUBE_FORCE_SYSTEMD=
	* Utilisation du pilote docker basé sur le profil existant
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	I0830 23:03:23.773830 1256904 out.go:296] Setting OutFile to fd 1 ...
	I0830 23:03:23.774007 1256904 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:03:23.774018 1256904 out.go:309] Setting ErrFile to fd 2...
	I0830 23:03:23.774024 1256904 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:03:23.774394 1256904 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
	I0830 23:03:23.774929 1256904 out.go:303] Setting JSON to false
	I0830 23:03:23.776132 1256904 start.go:128] hostinfo: {"hostname":"ip-172-31-31-251","uptime":27938,"bootTime":1693408666,"procs":407,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1043-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"982e3628-3742-4b3e-bb63-ac1b07660ec7"}
	I0830 23:03:23.776210 1256904 start.go:138] virtualization:  
	I0830 23:03:23.778488 1256904 out.go:177] * [functional-479614] minikube v1.31.2 sur Ubuntu 20.04 (arm64)
	I0830 23:03:23.780879 1256904 out.go:177]   - MINIKUBE_LOCATION=17114
	I0830 23:03:23.782557 1256904 out.go:177]   - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0830 23:03:23.781093 1256904 notify.go:220] Checking for updates...
	I0830 23:03:23.786665 1256904 out.go:177]   - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	I0830 23:03:23.789324 1256904 out.go:177]   - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	I0830 23:03:23.791340 1256904 out.go:177]   - MINIKUBE_BIN=out/minikube-linux-arm64
	I0830 23:03:23.793444 1256904 out.go:177]   - MINIKUBE_FORCE_SYSTEMD=
	I0830 23:03:23.796054 1256904 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
	I0830 23:03:23.796667 1256904 driver.go:373] Setting default libvirt URI to qemu:///system
	I0830 23:03:23.848520 1256904 docker.go:121] docker version: linux-24.0.5:Docker Engine - Community
	I0830 23:03:23.848615 1256904 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 23:03:23.964727 1256904 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:1 ContainersRunning:1 ContainersPaused:0 ContainersStopped:0 Images:4 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:31 OomKillDisable:true NGoroutines:46 SystemTime:2023-08-30 23:03:23.952609709 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 23:03:23.964891 1256904 docker.go:294] overlay module found
	I0830 23:03:23.966945 1256904 out.go:177] * Utilisation du pilote docker basé sur le profil existant
	I0830 23:03:23.968865 1256904 start.go:298] selected driver: docker
	I0830 23:03:23.968888 1256904 start.go:902] validating driver "docker" against &{Name:functional-479614 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.40-1692872184-17120@sha256:42602f0d347faca66d9347bdc33243fe5f4d6b3fff3ba53f3b2fc2d5fe63e9ec Memory:4000 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.28.1 ClusterName:functional-479614 Namespace:default APIServerName:miniku
beCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[{Component:apiserver Key:enable-admission-plugins Value:NamespaceAutoProvision}] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8441 NodeName:} Nodes:[{Name: IP:192.168.49.2 Port:8441 KubernetesVersion:v1.28.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersi
on:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0}
	I0830 23:03:23.969010 1256904 start.go:913] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
	I0830 23:03:23.971676 1256904 out.go:177] 
	W0830 23:03:23.973697 1256904 out.go:239] X Fermeture en raison de RSRC_INSUFFICIENT_REQ_MEMORY : L'allocation de mémoire demandée 250 Mio est inférieure au minimum utilisable de 1800 Mo
	X Fermeture en raison de RSRC_INSUFFICIENT_REQ_MEMORY : L'allocation de mémoire demandée 250 Mio est inférieure au minimum utilisable de 1800 Mo
	I0830 23:03:23.975681 1256904 out.go:177] 

                                                
                                                
** /stderr **
--- PASS: TestFunctional/parallel/InternationalLanguage (0.30s)

                                                
                                    
x
+
TestFunctional/parallel/StatusCmd (1.55s)

                                                
                                                
=== RUN   TestFunctional/parallel/StatusCmd
=== PAUSE TestFunctional/parallel/StatusCmd

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/StatusCmd
functional_test.go:850: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 status
functional_test.go:856: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 status -f host:{{.Host}},kublet:{{.Kubelet}},apiserver:{{.APIServer}},kubeconfig:{{.Kubeconfig}}
functional_test.go:868: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 status -o json
--- PASS: TestFunctional/parallel/StatusCmd (1.55s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmdConnect (6.76s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmdConnect
=== PAUSE TestFunctional/parallel/ServiceCmdConnect

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ServiceCmdConnect
functional_test.go:1626: (dbg) Run:  kubectl --context functional-479614 create deployment hello-node-connect --image=registry.k8s.io/echoserver-arm:1.8
functional_test.go:1634: (dbg) Run:  kubectl --context functional-479614 expose deployment hello-node-connect --type=NodePort --port=8080
functional_test.go:1639: (dbg) TestFunctional/parallel/ServiceCmdConnect: waiting 10m0s for pods matching "app=hello-node-connect" in namespace "default" ...
helpers_test.go:344: "hello-node-connect-7799dfb7c6-4m85v" [2d8d0c2b-d5d9-4e4c-8c94-8bf90428a52b] Pending / Ready:ContainersNotReady (containers with unready status: [echoserver-arm]) / ContainersReady:ContainersNotReady (containers with unready status: [echoserver-arm])
helpers_test.go:344: "hello-node-connect-7799dfb7c6-4m85v" [2d8d0c2b-d5d9-4e4c-8c94-8bf90428a52b] Running
functional_test.go:1639: (dbg) TestFunctional/parallel/ServiceCmdConnect: app=hello-node-connect healthy within 6.014674053s
functional_test.go:1648: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 service hello-node-connect --url
functional_test.go:1654: found endpoint for hello-node-connect: http://192.168.49.2:30267
functional_test.go:1674: http://192.168.49.2:30267: success! body:

                                                
                                                

                                                
                                                
Hostname: hello-node-connect-7799dfb7c6-4m85v

                                                
                                                
Pod Information:
	-no pod information available-

                                                
                                                
Server values:
	server_version=nginx: 1.13.3 - lua: 10008

                                                
                                                
Request Information:
	client_address=10.244.0.1
	method=GET
	real path=/
	query=
	request_version=1.1
	request_uri=http://192.168.49.2:8080/

                                                
                                                
Request Headers:
	accept-encoding=gzip
	host=192.168.49.2:30267
	user-agent=Go-http-client/1.1

                                                
                                                
Request Body:
	-no body in request-

                                                
                                                
--- PASS: TestFunctional/parallel/ServiceCmdConnect (6.76s)

                                                
                                    
x
+
TestFunctional/parallel/AddonsCmd (0.22s)

                                                
                                                
=== RUN   TestFunctional/parallel/AddonsCmd
=== PAUSE TestFunctional/parallel/AddonsCmd

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/AddonsCmd
functional_test.go:1689: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 addons list
functional_test.go:1701: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 addons list -o json
--- PASS: TestFunctional/parallel/AddonsCmd (0.22s)

                                                
                                    
x
+
TestFunctional/parallel/PersistentVolumeClaim (24.88s)

                                                
                                                
=== RUN   TestFunctional/parallel/PersistentVolumeClaim
=== PAUSE TestFunctional/parallel/PersistentVolumeClaim

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/PersistentVolumeClaim
functional_test_pvc_test.go:44: (dbg) TestFunctional/parallel/PersistentVolumeClaim: waiting 4m0s for pods matching "integration-test=storage-provisioner" in namespace "kube-system" ...
helpers_test.go:344: "storage-provisioner" [edc04be5-5acc-4bd3-b8ff-5bac7e86389d] Running
functional_test_pvc_test.go:44: (dbg) TestFunctional/parallel/PersistentVolumeClaim: integration-test=storage-provisioner healthy within 5.013117298s
functional_test_pvc_test.go:49: (dbg) Run:  kubectl --context functional-479614 get storageclass -o=json
functional_test_pvc_test.go:69: (dbg) Run:  kubectl --context functional-479614 apply -f testdata/storage-provisioner/pvc.yaml
functional_test_pvc_test.go:76: (dbg) Run:  kubectl --context functional-479614 get pvc myclaim -o=json
functional_test_pvc_test.go:125: (dbg) Run:  kubectl --context functional-479614 apply -f testdata/storage-provisioner/pod.yaml
functional_test_pvc_test.go:130: (dbg) TestFunctional/parallel/PersistentVolumeClaim: waiting 3m0s for pods matching "test=storage-provisioner" in namespace "default" ...
helpers_test.go:344: "sp-pod" [9ab84a8d-ba9d-4200-be3e-f44ef40c0ef4] Pending
helpers_test.go:344: "sp-pod" [9ab84a8d-ba9d-4200-be3e-f44ef40c0ef4] Pending / Ready:ContainersNotReady (containers with unready status: [myfrontend]) / ContainersReady:ContainersNotReady (containers with unready status: [myfrontend])
helpers_test.go:344: "sp-pod" [9ab84a8d-ba9d-4200-be3e-f44ef40c0ef4] Running
functional_test_pvc_test.go:130: (dbg) TestFunctional/parallel/PersistentVolumeClaim: test=storage-provisioner healthy within 10.027748844s
functional_test_pvc_test.go:100: (dbg) Run:  kubectl --context functional-479614 exec sp-pod -- touch /tmp/mount/foo
functional_test_pvc_test.go:106: (dbg) Run:  kubectl --context functional-479614 delete -f testdata/storage-provisioner/pod.yaml
functional_test_pvc_test.go:106: (dbg) Done: kubectl --context functional-479614 delete -f testdata/storage-provisioner/pod.yaml: (1.311287189s)
functional_test_pvc_test.go:125: (dbg) Run:  kubectl --context functional-479614 apply -f testdata/storage-provisioner/pod.yaml
functional_test_pvc_test.go:130: (dbg) TestFunctional/parallel/PersistentVolumeClaim: waiting 3m0s for pods matching "test=storage-provisioner" in namespace "default" ...
helpers_test.go:344: "sp-pod" [ab38ef7f-6426-44f2-b2e1-d14a91b2a563] Pending
helpers_test.go:344: "sp-pod" [ab38ef7f-6426-44f2-b2e1-d14a91b2a563] Pending / Ready:ContainersNotReady (containers with unready status: [myfrontend]) / ContainersReady:ContainersNotReady (containers with unready status: [myfrontend])
helpers_test.go:344: "sp-pod" [ab38ef7f-6426-44f2-b2e1-d14a91b2a563] Running
functional_test_pvc_test.go:130: (dbg) TestFunctional/parallel/PersistentVolumeClaim: test=storage-provisioner healthy within 7.020714994s
functional_test_pvc_test.go:114: (dbg) Run:  kubectl --context functional-479614 exec sp-pod -- ls /tmp/mount
--- PASS: TestFunctional/parallel/PersistentVolumeClaim (24.88s)

                                                
                                    
x
+
TestFunctional/parallel/SSHCmd (0.85s)

                                                
                                                
=== RUN   TestFunctional/parallel/SSHCmd
=== PAUSE TestFunctional/parallel/SSHCmd

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/SSHCmd
functional_test.go:1724: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "echo hello"
functional_test.go:1741: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "cat /etc/hostname"
--- PASS: TestFunctional/parallel/SSHCmd (0.85s)

                                                
                                    
x
+
TestFunctional/parallel/CpCmd (1.61s)

                                                
                                                
=== RUN   TestFunctional/parallel/CpCmd
=== PAUSE TestFunctional/parallel/CpCmd

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/CpCmd
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 cp testdata/cp-test.txt /home/docker/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh -n functional-479614 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 cp functional-479614:/home/docker/cp-test.txt /tmp/TestFunctionalparallelCpCmd271046065/001/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh -n functional-479614 "sudo cat /home/docker/cp-test.txt"
--- PASS: TestFunctional/parallel/CpCmd (1.61s)

                                                
                                    
x
+
TestFunctional/parallel/FileSync (0.4s)

                                                
                                                
=== RUN   TestFunctional/parallel/FileSync
=== PAUSE TestFunctional/parallel/FileSync

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/FileSync
functional_test.go:1925: Checking for existence of /etc/test/nested/copy/1225375/hosts within VM
functional_test.go:1927: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "sudo cat /etc/test/nested/copy/1225375/hosts"
functional_test.go:1932: file sync test content: Test file for checking file sync process
--- PASS: TestFunctional/parallel/FileSync (0.40s)

                                                
                                    
x
+
TestFunctional/parallel/CertSync (2.42s)

                                                
                                                
=== RUN   TestFunctional/parallel/CertSync
=== PAUSE TestFunctional/parallel/CertSync

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/CertSync
functional_test.go:1968: Checking for existence of /etc/ssl/certs/1225375.pem within VM
functional_test.go:1969: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "sudo cat /etc/ssl/certs/1225375.pem"
functional_test.go:1968: Checking for existence of /usr/share/ca-certificates/1225375.pem within VM
functional_test.go:1969: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "sudo cat /usr/share/ca-certificates/1225375.pem"
functional_test.go:1968: Checking for existence of /etc/ssl/certs/51391683.0 within VM
functional_test.go:1969: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "sudo cat /etc/ssl/certs/51391683.0"
functional_test.go:1995: Checking for existence of /etc/ssl/certs/12253752.pem within VM
functional_test.go:1996: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "sudo cat /etc/ssl/certs/12253752.pem"
functional_test.go:1995: Checking for existence of /usr/share/ca-certificates/12253752.pem within VM
functional_test.go:1996: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "sudo cat /usr/share/ca-certificates/12253752.pem"
functional_test.go:1995: Checking for existence of /etc/ssl/certs/3ec20f2e.0 within VM
functional_test.go:1996: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "sudo cat /etc/ssl/certs/3ec20f2e.0"
--- PASS: TestFunctional/parallel/CertSync (2.42s)

                                                
                                    
x
+
TestFunctional/parallel/NodeLabels (0.11s)

                                                
                                                
=== RUN   TestFunctional/parallel/NodeLabels
=== PAUSE TestFunctional/parallel/NodeLabels

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/NodeLabels
functional_test.go:218: (dbg) Run:  kubectl --context functional-479614 get nodes --output=go-template "--template='{{range $k, $v := (index .items 0).metadata.labels}}{{$k}} {{end}}'"
--- PASS: TestFunctional/parallel/NodeLabels (0.11s)

                                                
                                    
x
+
TestFunctional/parallel/NonActiveRuntimeDisabled (0.82s)

                                                
                                                
=== RUN   TestFunctional/parallel/NonActiveRuntimeDisabled
=== PAUSE TestFunctional/parallel/NonActiveRuntimeDisabled

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/NonActiveRuntimeDisabled
functional_test.go:2023: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "sudo systemctl is-active docker"
functional_test.go:2023: (dbg) Non-zero exit: out/minikube-linux-arm64 -p functional-479614 ssh "sudo systemctl is-active docker": exit status 1 (451.013498ms)

                                                
                                                
-- stdout --
	inactive

                                                
                                                
-- /stdout --
** stderr ** 
	ssh: Process exited with status 3

                                                
                                                
** /stderr **
functional_test.go:2023: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "sudo systemctl is-active crio"
functional_test.go:2023: (dbg) Non-zero exit: out/minikube-linux-arm64 -p functional-479614 ssh "sudo systemctl is-active crio": exit status 1 (368.342258ms)

                                                
                                                
-- stdout --
	inactive

                                                
                                                
-- /stdout --
** stderr ** 
	ssh: Process exited with status 3

                                                
                                                
** /stderr **
--- PASS: TestFunctional/parallel/NonActiveRuntimeDisabled (0.82s)

                                                
                                    
x
+
TestFunctional/parallel/License (0.43s)

                                                
                                                
=== RUN   TestFunctional/parallel/License
=== PAUSE TestFunctional/parallel/License

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/License
functional_test.go:2284: (dbg) Run:  out/minikube-linux-arm64 license
--- PASS: TestFunctional/parallel/License (0.43s)

                                                
                                    
x
+
TestFunctional/parallel/Version/short (0.09s)

                                                
                                                
=== RUN   TestFunctional/parallel/Version/short
=== PAUSE TestFunctional/parallel/Version/short

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/Version/short
functional_test.go:2252: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 version --short
--- PASS: TestFunctional/parallel/Version/short (0.09s)

                                                
                                    
x
+
TestFunctional/parallel/Version/components (1.24s)

                                                
                                                
=== RUN   TestFunctional/parallel/Version/components
=== PAUSE TestFunctional/parallel/Version/components

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/Version/components
functional_test.go:2266: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 version -o=json --components
functional_test.go:2266: (dbg) Done: out/minikube-linux-arm64 -p functional-479614 version -o=json --components: (1.241420207s)
--- PASS: TestFunctional/parallel/Version/components (1.24s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageListShort (0.42s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageListShort
=== PAUSE TestFunctional/parallel/ImageCommands/ImageListShort

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ImageCommands/ImageListShort
functional_test.go:260: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image ls --format short --alsologtostderr
functional_test.go:265: (dbg) Stdout: out/minikube-linux-arm64 -p functional-479614 image ls --format short --alsologtostderr:
registry.k8s.io/pause:latest
registry.k8s.io/pause:3.9
registry.k8s.io/pause:3.3
registry.k8s.io/pause:3.1
registry.k8s.io/kube-scheduler:v1.28.1
registry.k8s.io/kube-proxy:v1.28.1
registry.k8s.io/kube-controller-manager:v1.28.1
registry.k8s.io/kube-apiserver:v1.28.1
registry.k8s.io/etcd:3.5.9-0
registry.k8s.io/echoserver-arm:1.8
registry.k8s.io/coredns/coredns:v1.10.1
gcr.io/k8s-minikube/storage-provisioner:v5
gcr.io/k8s-minikube/busybox:1.28.4-glibc
docker.io/library/nginx:latest
docker.io/library/nginx:alpine
docker.io/library/minikube-local-cache-test:functional-479614
docker.io/kindest/kindnetd:v20230511-dc714da8
functional_test.go:268: (dbg) Stderr: out/minikube-linux-arm64 -p functional-479614 image ls --format short --alsologtostderr:
I0830 23:03:32.012348 1258186 out.go:296] Setting OutFile to fd 1 ...
I0830 23:03:32.012539 1258186 out.go:343] TERM=,COLORTERM=, which probably does not support color
I0830 23:03:32.012565 1258186 out.go:309] Setting ErrFile to fd 2...
I0830 23:03:32.012583 1258186 out.go:343] TERM=,COLORTERM=, which probably does not support color
I0830 23:03:32.012875 1258186 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
I0830 23:03:32.013651 1258186 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
I0830 23:03:32.013864 1258186 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
I0830 23:03:32.014363 1258186 cli_runner.go:164] Run: docker container inspect functional-479614 --format={{.State.Status}}
I0830 23:03:32.057000 1258186 ssh_runner.go:195] Run: systemctl --version
I0830 23:03:32.057061 1258186 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" functional-479614
I0830 23:03:32.092721 1258186 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34329 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/functional-479614/id_rsa Username:docker}
I0830 23:03:32.223152 1258186 ssh_runner.go:195] Run: sudo crictl images --output json
--- PASS: TestFunctional/parallel/ImageCommands/ImageListShort (0.42s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageListTable (0.34s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageListTable
=== PAUSE TestFunctional/parallel/ImageCommands/ImageListTable

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ImageCommands/ImageListTable
functional_test.go:260: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image ls --format table --alsologtostderr
functional_test.go:265: (dbg) Stdout: out/minikube-linux-arm64 -p functional-479614 image ls --format table --alsologtostderr:
|---------------------------------------------|--------------------|---------------|--------|
|                    Image                    |        Tag         |   Image ID    |  Size  |
|---------------------------------------------|--------------------|---------------|--------|
| docker.io/library/nginx                     | latest             | sha256:ab73c7 | 67.2MB |
| registry.k8s.io/etcd                        | 3.5.9-0            | sha256:9cdd64 | 86.5MB |
| registry.k8s.io/kube-controller-manager     | v1.28.1            | sha256:8b6e19 | 30.3MB |
| registry.k8s.io/kube-scheduler              | v1.28.1            | sha256:b4a5a5 | 17.1MB |
| docker.io/library/minikube-local-cache-test | functional-479614  | sha256:5474d3 | 1.01kB |
| gcr.io/k8s-minikube/busybox                 | 1.28.4-glibc       | sha256:1611cd | 1.94MB |
| registry.k8s.io/echoserver-arm              | 1.8                | sha256:72565b | 45.3MB |
| registry.k8s.io/kube-apiserver              | v1.28.1            | sha256:b29fb6 | 31.5MB |
| registry.k8s.io/pause                       | 3.3                | sha256:3d1873 | 249kB  |
| registry.k8s.io/pause                       | 3.9                | sha256:829e9d | 268kB  |
| registry.k8s.io/coredns/coredns             | v1.10.1            | sha256:97e046 | 14.6MB |
| registry.k8s.io/pause                       | 3.1                | sha256:8057e0 | 262kB  |
| docker.io/kindest/kindnetd                  | v20230511-dc714da8 | sha256:b18bf7 | 25.3MB |
| docker.io/library/nginx                     | alpine             | sha256:fa0c6b | 17.6MB |
| gcr.io/k8s-minikube/storage-provisioner     | v5                 | sha256:ba04bb | 8.03MB |
| registry.k8s.io/kube-proxy                  | v1.28.1            | sha256:812f52 | 22MB   |
| registry.k8s.io/pause                       | latest             | sha256:8cb209 | 71.3kB |
|---------------------------------------------|--------------------|---------------|--------|
functional_test.go:268: (dbg) Stderr: out/minikube-linux-arm64 -p functional-479614 image ls --format table --alsologtostderr:
I0830 23:03:35.071992 1258502 out.go:296] Setting OutFile to fd 1 ...
I0830 23:03:35.072161 1258502 out.go:343] TERM=,COLORTERM=, which probably does not support color
I0830 23:03:35.072171 1258502 out.go:309] Setting ErrFile to fd 2...
I0830 23:03:35.072176 1258502 out.go:343] TERM=,COLORTERM=, which probably does not support color
I0830 23:03:35.072481 1258502 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
I0830 23:03:35.073262 1258502 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
I0830 23:03:35.073426 1258502 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
I0830 23:03:35.073905 1258502 cli_runner.go:164] Run: docker container inspect functional-479614 --format={{.State.Status}}
I0830 23:03:35.106600 1258502 ssh_runner.go:195] Run: systemctl --version
I0830 23:03:35.106687 1258502 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" functional-479614
I0830 23:03:35.126573 1258502 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34329 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/functional-479614/id_rsa Username:docker}
I0830 23:03:35.229445 1258502 ssh_runner.go:195] Run: sudo crictl images --output json
--- PASS: TestFunctional/parallel/ImageCommands/ImageListTable (0.34s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageListJson (0.33s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageListJson
=== PAUSE TestFunctional/parallel/ImageCommands/ImageListJson

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ImageCommands/ImageListJson
functional_test.go:260: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image ls --format json --alsologtostderr
functional_test.go:265: (dbg) Stdout: out/minikube-linux-arm64 -p functional-479614 image ls --format json --alsologtostderr:
[{"id":"sha256:fa0c6bb795403f8762e5cbf7b9f395aa036e7bd61c707485c1968b79bb3da9f1","repoDigests":["docker.io/library/nginx@sha256:16164a43b5faec40adb521e98272edc528e74f31c1352719132b8f7e53418d70"],"repoTags":["docker.io/library/nginx:alpine"],"size":"17568097"},{"id":"sha256:3d18732f8686cc3c878055d99a05fa80289502fa496b36b6a0fe0f77206a7300","repoDigests":[],"repoTags":["registry.k8s.io/pause:3.3"],"size":"249461"},{"id":"sha256:8cb2091f603e75187e2f6226c5901d12e00b1d1f778c6471ae4578e8a1c4724a","repoDigests":[],"repoTags":["registry.k8s.io/pause:latest"],"size":"71300"},{"id":"sha256:ab73c7fd672341e41ec600081253d0b99ea31d0c1acdfb46a1485004472da7ac","repoDigests":["docker.io/library/nginx@sha256:104c7c5c54f2685f0f46f3be607ce60da7085da3eaa5ad22d3d9f01594295e9c"],"repoTags":["docker.io/library/nginx:latest"],"size":"67190345"},{"id":"sha256:ba04bb24b95753201135cbc420b233c1b0b9fa2e1fd21d28319c348c33fbcde6","repoDigests":["gcr.io/k8s-minikube/storage-provisioner@sha256:18eb69d1418e854ad5a19e399310e52808a8321e4c
441c1dddad8977a0d7a944"],"repoTags":["gcr.io/k8s-minikube/storage-provisioner:v5"],"size":"8034419"},{"id":"sha256:72565bf5bbedfb62e9d21afa2b1221b2c7a5e05b746dae33430bc550d3f87beb","repoDigests":["registry.k8s.io/echoserver-arm@sha256:b33d4cdf6ed097f4e9b77b135d83a596ab73c6268b0342648818eb85f5edfdb5"],"repoTags":["registry.k8s.io/echoserver-arm:1.8"],"size":"45324675"},{"id":"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e","repoDigests":["registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097"],"repoTags":["registry.k8s.io/pause:3.9"],"size":"268051"},{"id":"sha256:b18bf71b941bae2e12db1c07e567ad14e4febbc778310a0fc64487f1ac877d79","repoDigests":["docker.io/kindest/kindnetd@sha256:6c00e28db008c2afa67d9ee085c86184ec9ae5281d5ae1bd15006746fb9a1974"],"repoTags":["docker.io/kindest/kindnetd:v20230511-dc714da8"],"size":"25334607"},{"id":"sha256:8057e0500773a37cde2cff041eb13ebd68c748419a2fbfd1dfb5bf38696cc8e5","repoDigests":[],"repoTags":["registry.k8s.io/
pause:3.1"],"size":"262191"},{"id":"sha256:a422e0e982356f6c1cf0e5bb7b733363caae3992a07c99951fbcc73e58ed656a","repoDigests":["docker.io/kubernetesui/metrics-scraper@sha256:76049887f07a0476dc93efc2d3569b9529bf982b22d29f356092ce206e98765c"],"repoTags":[],"size":"18306114"},{"id":"sha256:1611cd07b61d57dbbfebe6db242513fd51e1c02d20ba08af17a45837d86a8a8c","repoDigests":["gcr.io/k8s-minikube/busybox@sha256:2d03e6ceeb99250061dd110530b0ece7998cd84121f952adef120ea7c5a6f00e"],"repoTags":["gcr.io/k8s-minikube/busybox:1.28.4-glibc"],"size":"1935750"},{"id":"sha256:97e04611ad43405a2e5863ae17c6f1bc9181bdefdaa78627c432ef754a4eb108","repoDigests":["registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e"],"repoTags":["registry.k8s.io/coredns/coredns:v1.10.1"],"size":"14557471"},{"id":"sha256:9cdd6470f48c8b127530b7ce6ea4b3524137984481e48bcde619735890840ace","repoDigests":["registry.k8s.io/etcd@sha256:e013d0d5e4e25d00c61a7ff839927a1f36479678f11e49502b53a5e0b14f10c3"],"repoTags":["r
egistry.k8s.io/etcd:3.5.9-0"],"size":"86464836"},{"id":"sha256:8b6e1980b7584ebf92ee961322982c26a525c4e4e2181e037b8854697be71965","repoDigests":["registry.k8s.io/kube-controller-manager@sha256:dda6dba8a55203ed1595efcda865a526b9282c2d9b959e9ed0a88f54a7a91195"],"repoTags":["registry.k8s.io/kube-controller-manager:v1.28.1"],"size":"30330541"},{"id":"sha256:20b332c9a70d8516d849d1ac23eff5800cbb2f263d379f0ec11ee908db6b25a8","repoDigests":["docker.io/kubernetesui/dashboard@sha256:2e500d29e9d5f4a086b908eb8dfe7ecac57d2ab09d65b24f588b1d449841ef93"],"repoTags":[],"size":"74084559"},{"id":"sha256:5474d3a9094a7d1d738dba1eaa91ae5f06a9eb414c0f54f4e3d4bcd3ebf977a4","repoDigests":[],"repoTags":["docker.io/library/minikube-local-cache-test:functional-479614"],"size":"1006"},{"id":"sha256:b29fb62480892633ac479243b9841b88f9ae30865773fd76b97522541cd5633a","repoDigests":["registry.k8s.io/kube-apiserver@sha256:f517207d13adeb50c63f9bdac2824e0e7512817eca47ac0540685771243742b2"],"repoTags":["registry.k8s.io/kube-apiserver:v1.28.1"],"si
ze":"31519813"},{"id":"sha256:812f5241df7fd64adb98d461bd6259a825a371fb3b2d5258752579380bc39c26","repoDigests":["registry.k8s.io/kube-proxy@sha256:30096ad233e7bfe72662180c5ac4497f732346d6d25b7c1f1c0c7cb1a1e7e41c"],"repoTags":["registry.k8s.io/kube-proxy:v1.28.1"],"size":"21974303"},{"id":"sha256:b4a5a57e994924bffc4556da6c6c39d27ebaf593155983163d0b2367037bcb87","repoDigests":["registry.k8s.io/kube-scheduler@sha256:3c9249a1f7623007a8db3522eac203f94cbb3910501879b792d95ea8470cc3d4"],"repoTags":["registry.k8s.io/kube-scheduler:v1.28.1"],"size":"17052956"}]
functional_test.go:268: (dbg) Stderr: out/minikube-linux-arm64 -p functional-479614 image ls --format json --alsologtostderr:
I0830 23:03:34.730881 1258408 out.go:296] Setting OutFile to fd 1 ...
I0830 23:03:34.731031 1258408 out.go:343] TERM=,COLORTERM=, which probably does not support color
I0830 23:03:34.731038 1258408 out.go:309] Setting ErrFile to fd 2...
I0830 23:03:34.731043 1258408 out.go:343] TERM=,COLORTERM=, which probably does not support color
I0830 23:03:34.731333 1258408 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
I0830 23:03:34.732049 1258408 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
I0830 23:03:34.732170 1258408 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
I0830 23:03:34.732712 1258408 cli_runner.go:164] Run: docker container inspect functional-479614 --format={{.State.Status}}
I0830 23:03:34.753263 1258408 ssh_runner.go:195] Run: systemctl --version
I0830 23:03:34.753312 1258408 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" functional-479614
I0830 23:03:34.799820 1258408 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34329 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/functional-479614/id_rsa Username:docker}
I0830 23:03:34.915186 1258408 ssh_runner.go:195] Run: sudo crictl images --output json
--- PASS: TestFunctional/parallel/ImageCommands/ImageListJson (0.33s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageListYaml (0.33s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageListYaml
=== PAUSE TestFunctional/parallel/ImageCommands/ImageListYaml

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ImageCommands/ImageListYaml
functional_test.go:260: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image ls --format yaml --alsologtostderr
functional_test.go:265: (dbg) Stdout: out/minikube-linux-arm64 -p functional-479614 image ls --format yaml --alsologtostderr:
- id: sha256:72565bf5bbedfb62e9d21afa2b1221b2c7a5e05b746dae33430bc550d3f87beb
repoDigests:
- registry.k8s.io/echoserver-arm@sha256:b33d4cdf6ed097f4e9b77b135d83a596ab73c6268b0342648818eb85f5edfdb5
repoTags:
- registry.k8s.io/echoserver-arm:1.8
size: "45324675"
- id: sha256:812f5241df7fd64adb98d461bd6259a825a371fb3b2d5258752579380bc39c26
repoDigests:
- registry.k8s.io/kube-proxy@sha256:30096ad233e7bfe72662180c5ac4497f732346d6d25b7c1f1c0c7cb1a1e7e41c
repoTags:
- registry.k8s.io/kube-proxy:v1.28.1
size: "21974303"
- id: sha256:20b332c9a70d8516d849d1ac23eff5800cbb2f263d379f0ec11ee908db6b25a8
repoDigests:
- docker.io/kubernetesui/dashboard@sha256:2e500d29e9d5f4a086b908eb8dfe7ecac57d2ab09d65b24f588b1d449841ef93
repoTags: []
size: "74084559"
- id: sha256:ab73c7fd672341e41ec600081253d0b99ea31d0c1acdfb46a1485004472da7ac
repoDigests:
- docker.io/library/nginx@sha256:104c7c5c54f2685f0f46f3be607ce60da7085da3eaa5ad22d3d9f01594295e9c
repoTags:
- docker.io/library/nginx:latest
size: "67190345"
- id: sha256:ba04bb24b95753201135cbc420b233c1b0b9fa2e1fd21d28319c348c33fbcde6
repoDigests:
- gcr.io/k8s-minikube/storage-provisioner@sha256:18eb69d1418e854ad5a19e399310e52808a8321e4c441c1dddad8977a0d7a944
repoTags:
- gcr.io/k8s-minikube/storage-provisioner:v5
size: "8034419"
- id: sha256:97e04611ad43405a2e5863ae17c6f1bc9181bdefdaa78627c432ef754a4eb108
repoDigests:
- registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
repoTags:
- registry.k8s.io/coredns/coredns:v1.10.1
size: "14557471"
- id: sha256:8b6e1980b7584ebf92ee961322982c26a525c4e4e2181e037b8854697be71965
repoDigests:
- registry.k8s.io/kube-controller-manager@sha256:dda6dba8a55203ed1595efcda865a526b9282c2d9b959e9ed0a88f54a7a91195
repoTags:
- registry.k8s.io/kube-controller-manager:v1.28.1
size: "30330541"
- id: sha256:b4a5a57e994924bffc4556da6c6c39d27ebaf593155983163d0b2367037bcb87
repoDigests:
- registry.k8s.io/kube-scheduler@sha256:3c9249a1f7623007a8db3522eac203f94cbb3910501879b792d95ea8470cc3d4
repoTags:
- registry.k8s.io/kube-scheduler:v1.28.1
size: "17052956"
- id: sha256:b29fb62480892633ac479243b9841b88f9ae30865773fd76b97522541cd5633a
repoDigests:
- registry.k8s.io/kube-apiserver@sha256:f517207d13adeb50c63f9bdac2824e0e7512817eca47ac0540685771243742b2
repoTags:
- registry.k8s.io/kube-apiserver:v1.28.1
size: "31519813"
- id: sha256:8057e0500773a37cde2cff041eb13ebd68c748419a2fbfd1dfb5bf38696cc8e5
repoDigests: []
repoTags:
- registry.k8s.io/pause:3.1
size: "262191"
- id: sha256:b18bf71b941bae2e12db1c07e567ad14e4febbc778310a0fc64487f1ac877d79
repoDigests:
- docker.io/kindest/kindnetd@sha256:6c00e28db008c2afa67d9ee085c86184ec9ae5281d5ae1bd15006746fb9a1974
repoTags:
- docker.io/kindest/kindnetd:v20230511-dc714da8
size: "25334607"
- id: sha256:fa0c6bb795403f8762e5cbf7b9f395aa036e7bd61c707485c1968b79bb3da9f1
repoDigests:
- docker.io/library/nginx@sha256:16164a43b5faec40adb521e98272edc528e74f31c1352719132b8f7e53418d70
repoTags:
- docker.io/library/nginx:alpine
size: "17568097"
- id: sha256:9cdd6470f48c8b127530b7ce6ea4b3524137984481e48bcde619735890840ace
repoDigests:
- registry.k8s.io/etcd@sha256:e013d0d5e4e25d00c61a7ff839927a1f36479678f11e49502b53a5e0b14f10c3
repoTags:
- registry.k8s.io/etcd:3.5.9-0
size: "86464836"
- id: sha256:3d18732f8686cc3c878055d99a05fa80289502fa496b36b6a0fe0f77206a7300
repoDigests: []
repoTags:
- registry.k8s.io/pause:3.3
size: "249461"
- id: sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e
repoDigests:
- registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097
repoTags:
- registry.k8s.io/pause:3.9
size: "268051"
- id: sha256:8cb2091f603e75187e2f6226c5901d12e00b1d1f778c6471ae4578e8a1c4724a
repoDigests: []
repoTags:
- registry.k8s.io/pause:latest
size: "71300"
- id: sha256:5474d3a9094a7d1d738dba1eaa91ae5f06a9eb414c0f54f4e3d4bcd3ebf977a4
repoDigests: []
repoTags:
- docker.io/library/minikube-local-cache-test:functional-479614
size: "1006"
- id: sha256:a422e0e982356f6c1cf0e5bb7b733363caae3992a07c99951fbcc73e58ed656a
repoDigests:
- docker.io/kubernetesui/metrics-scraper@sha256:76049887f07a0476dc93efc2d3569b9529bf982b22d29f356092ce206e98765c
repoTags: []
size: "18306114"
- id: sha256:1611cd07b61d57dbbfebe6db242513fd51e1c02d20ba08af17a45837d86a8a8c
repoDigests:
- gcr.io/k8s-minikube/busybox@sha256:2d03e6ceeb99250061dd110530b0ece7998cd84121f952adef120ea7c5a6f00e
repoTags:
- gcr.io/k8s-minikube/busybox:1.28.4-glibc
size: "1935750"

                                                
                                                
functional_test.go:268: (dbg) Stderr: out/minikube-linux-arm64 -p functional-479614 image ls --format yaml --alsologtostderr:
I0830 23:03:34.401668 1258374 out.go:296] Setting OutFile to fd 1 ...
I0830 23:03:34.401932 1258374 out.go:343] TERM=,COLORTERM=, which probably does not support color
I0830 23:03:34.401946 1258374 out.go:309] Setting ErrFile to fd 2...
I0830 23:03:34.401952 1258374 out.go:343] TERM=,COLORTERM=, which probably does not support color
I0830 23:03:34.402281 1258374 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
I0830 23:03:34.403087 1258374 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
I0830 23:03:34.403266 1258374 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
I0830 23:03:34.404706 1258374 cli_runner.go:164] Run: docker container inspect functional-479614 --format={{.State.Status}}
I0830 23:03:34.434962 1258374 ssh_runner.go:195] Run: systemctl --version
I0830 23:03:34.435021 1258374 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" functional-479614
I0830 23:03:34.467444 1258374 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34329 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/functional-479614/id_rsa Username:docker}
I0830 23:03:34.580599 1258374 ssh_runner.go:195] Run: sudo crictl images --output json
--- PASS: TestFunctional/parallel/ImageCommands/ImageListYaml (0.33s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageBuild (3.6s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageBuild
=== PAUSE TestFunctional/parallel/ImageCommands/ImageBuild

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ImageCommands/ImageBuild
functional_test.go:307: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh pgrep buildkitd
functional_test.go:307: (dbg) Non-zero exit: out/minikube-linux-arm64 -p functional-479614 ssh pgrep buildkitd: exit status 1 (454.643799ms)

                                                
                                                
** stderr ** 
	ssh: Process exited with status 1

                                                
                                                
** /stderr **
functional_test.go:314: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image build -t localhost/my-image:functional-479614 testdata/build --alsologtostderr
2023/08/30 23:03:34 [DEBUG] GET http://127.0.0.1:36195/api/v1/namespaces/kubernetes-dashboard/services/http:kubernetes-dashboard:/proxy/
functional_test.go:314: (dbg) Done: out/minikube-linux-arm64 -p functional-479614 image build -t localhost/my-image:functional-479614 testdata/build --alsologtostderr: (2.878876757s)
functional_test.go:322: (dbg) Stderr: out/minikube-linux-arm64 -p functional-479614 image build -t localhost/my-image:functional-479614 testdata/build --alsologtostderr:
I0830 23:03:32.867456 1258267 out.go:296] Setting OutFile to fd 1 ...
I0830 23:03:32.871705 1258267 out.go:343] TERM=,COLORTERM=, which probably does not support color
I0830 23:03:32.871731 1258267 out.go:309] Setting ErrFile to fd 2...
I0830 23:03:32.871739 1258267 out.go:343] TERM=,COLORTERM=, which probably does not support color
I0830 23:03:32.872053 1258267 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
I0830 23:03:32.872705 1258267 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
I0830 23:03:32.874489 1258267 config.go:182] Loaded profile config "functional-479614": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
I0830 23:03:32.874987 1258267 cli_runner.go:164] Run: docker container inspect functional-479614 --format={{.State.Status}}
I0830 23:03:32.897888 1258267 ssh_runner.go:195] Run: systemctl --version
I0830 23:03:32.897944 1258267 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" functional-479614
I0830 23:03:32.925772 1258267 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34329 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/functional-479614/id_rsa Username:docker}
I0830 23:03:33.047366 1258267 build_images.go:151] Building image from path: /tmp/build.487408690.tar
I0830 23:03:33.047444 1258267 ssh_runner.go:195] Run: sudo mkdir -p /var/lib/minikube/build
I0830 23:03:33.059955 1258267 ssh_runner.go:195] Run: stat -c "%s %y" /var/lib/minikube/build/build.487408690.tar
I0830 23:03:33.065731 1258267 ssh_runner.go:352] existence check for /var/lib/minikube/build/build.487408690.tar: stat -c "%s %y" /var/lib/minikube/build/build.487408690.tar: Process exited with status 1
stdout:

                                                
                                                
stderr:
stat: cannot statx '/var/lib/minikube/build/build.487408690.tar': No such file or directory
I0830 23:03:33.065770 1258267 ssh_runner.go:362] scp /tmp/build.487408690.tar --> /var/lib/minikube/build/build.487408690.tar (3072 bytes)
I0830 23:03:33.101827 1258267 ssh_runner.go:195] Run: sudo mkdir -p /var/lib/minikube/build/build.487408690
I0830 23:03:33.117608 1258267 ssh_runner.go:195] Run: sudo tar -C /var/lib/minikube/build/build.487408690 -xf /var/lib/minikube/build/build.487408690.tar
I0830 23:03:33.133623 1258267 containerd.go:378] Building image: /var/lib/minikube/build/build.487408690
I0830 23:03:33.133713 1258267 ssh_runner.go:195] Run: sudo buildctl build --frontend dockerfile.v0 --local context=/var/lib/minikube/build/build.487408690 --local dockerfile=/var/lib/minikube/build/build.487408690 --output type=image,name=localhost/my-image:functional-479614
#1 [internal] load .dockerignore
#1 transferring context: 2B done
#1 DONE 0.0s

                                                
                                                
#2 [internal] load build definition from Dockerfile
#2 transferring dockerfile: 97B done
#2 DONE 0.1s

                                                
                                                
#3 [internal] load metadata for gcr.io/k8s-minikube/busybox:latest
#3 DONE 0.8s

                                                
                                                
#4 [internal] load build context
#4 transferring context: 62B done
#4 DONE 0.1s

                                                
                                                
#5 [1/3] FROM gcr.io/k8s-minikube/busybox@sha256:ca5ae90100d50772da31f3b5016209e25ad61972404e2ccd83d44f10dee7e79b
#5 resolve gcr.io/k8s-minikube/busybox@sha256:ca5ae90100d50772da31f3b5016209e25ad61972404e2ccd83d44f10dee7e79b 0.0s done
#5 sha256:a01966dde7f8d5ba10b6d87e776c7c8fb5a5f6bfa678874bd28b33b1fc6dba34 0B / 828.50kB 0.2s
#5 sha256:a01966dde7f8d5ba10b6d87e776c7c8fb5a5f6bfa678874bd28b33b1fc6dba34 828.50kB / 828.50kB 0.2s done
#5 extracting sha256:a01966dde7f8d5ba10b6d87e776c7c8fb5a5f6bfa678874bd28b33b1fc6dba34
#5 extracting sha256:a01966dde7f8d5ba10b6d87e776c7c8fb5a5f6bfa678874bd28b33b1fc6dba34 0.2s done
#5 DONE 0.5s

                                                
                                                
#6 [2/3] RUN true
#6 DONE 0.8s

                                                
                                                
#7 [3/3] ADD content.txt /
#7 DONE 0.0s

                                                
                                                
#8 exporting to image
#8 exporting layers 0.1s done
#8 exporting manifest sha256:1a0aa11723a19192f19a88bbf036bc7b7a85c97dfaaa4275cdc748acd576ff2d 0.0s done
#8 exporting config sha256:0dce78baf52e44500e7af0c096bad12fad45b1f4c2f621694ca7d4dcfe1c6743
#8 exporting config sha256:0dce78baf52e44500e7af0c096bad12fad45b1f4c2f621694ca7d4dcfe1c6743 0.0s done
#8 naming to localhost/my-image:functional-479614 done
#8 DONE 0.1s
I0830 23:03:35.626997 1258267 ssh_runner.go:235] Completed: sudo buildctl build --frontend dockerfile.v0 --local context=/var/lib/minikube/build/build.487408690 --local dockerfile=/var/lib/minikube/build/build.487408690 --output type=image,name=localhost/my-image:functional-479614: (2.493247941s)
I0830 23:03:35.627071 1258267 ssh_runner.go:195] Run: sudo rm -rf /var/lib/minikube/build/build.487408690
I0830 23:03:35.638867 1258267 ssh_runner.go:195] Run: sudo rm -f /var/lib/minikube/build/build.487408690.tar
I0830 23:03:35.650686 1258267 build_images.go:207] Built localhost/my-image:functional-479614 from /tmp/build.487408690.tar
I0830 23:03:35.650719 1258267 build_images.go:123] succeeded building to: functional-479614
I0830 23:03:35.650725 1258267 build_images.go:124] failed building to: 
functional_test.go:447: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image ls
--- PASS: TestFunctional/parallel/ImageCommands/ImageBuild (3.60s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/Setup (2.11s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/Setup
functional_test.go:341: (dbg) Run:  docker pull gcr.io/google-containers/addon-resizer:1.8.8
functional_test.go:341: (dbg) Done: docker pull gcr.io/google-containers/addon-resizer:1.8.8: (2.081583651s)
functional_test.go:346: (dbg) Run:  docker tag gcr.io/google-containers/addon-resizer:1.8.8 gcr.io/google-containers/addon-resizer:functional-479614
--- PASS: TestFunctional/parallel/ImageCommands/Setup (2.11s)

                                                
                                    
x
+
TestFunctional/parallel/UpdateContextCmd/no_changes (0.25s)

                                                
                                                
=== RUN   TestFunctional/parallel/UpdateContextCmd/no_changes
=== PAUSE TestFunctional/parallel/UpdateContextCmd/no_changes

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/UpdateContextCmd/no_changes
functional_test.go:2115: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 update-context --alsologtostderr -v=2
--- PASS: TestFunctional/parallel/UpdateContextCmd/no_changes (0.25s)

                                                
                                    
x
+
TestFunctional/parallel/UpdateContextCmd/no_minikube_cluster (0.33s)

                                                
                                                
=== RUN   TestFunctional/parallel/UpdateContextCmd/no_minikube_cluster
=== PAUSE TestFunctional/parallel/UpdateContextCmd/no_minikube_cluster

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/UpdateContextCmd/no_minikube_cluster
functional_test.go:2115: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 update-context --alsologtostderr -v=2
--- PASS: TestFunctional/parallel/UpdateContextCmd/no_minikube_cluster (0.33s)

                                                
                                    
x
+
TestFunctional/parallel/UpdateContextCmd/no_clusters (0.26s)

                                                
                                                
=== RUN   TestFunctional/parallel/UpdateContextCmd/no_clusters
=== PAUSE TestFunctional/parallel/UpdateContextCmd/no_clusters

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/UpdateContextCmd/no_clusters
functional_test.go:2115: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 update-context --alsologtostderr -v=2
--- PASS: TestFunctional/parallel/UpdateContextCmd/no_clusters (0.26s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmd/DeployApp (10.35s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmd/DeployApp
functional_test.go:1436: (dbg) Run:  kubectl --context functional-479614 create deployment hello-node --image=registry.k8s.io/echoserver-arm:1.8
functional_test.go:1444: (dbg) Run:  kubectl --context functional-479614 expose deployment hello-node --type=NodePort --port=8080
functional_test.go:1449: (dbg) TestFunctional/parallel/ServiceCmd/DeployApp: waiting 10m0s for pods matching "app=hello-node" in namespace "default" ...
helpers_test.go:344: "hello-node-759d89bdcc-q2d4t" [c8a40d6e-071e-4896-81a6-43aeb4963102] Pending / Ready:ContainersNotReady (containers with unready status: [echoserver-arm]) / ContainersReady:ContainersNotReady (containers with unready status: [echoserver-arm])
helpers_test.go:344: "hello-node-759d89bdcc-q2d4t" [c8a40d6e-071e-4896-81a6-43aeb4963102] Running
functional_test.go:1449: (dbg) TestFunctional/parallel/ServiceCmd/DeployApp: app=hello-node healthy within 10.031215347s
--- PASS: TestFunctional/parallel/ServiceCmd/DeployApp (10.35s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmd/List (0.44s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmd/List
functional_test.go:1458: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 service list
--- PASS: TestFunctional/parallel/ServiceCmd/List (0.44s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmd/JSONOutput (0.55s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmd/JSONOutput
functional_test.go:1488: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 service list -o json
functional_test.go:1493: Took "547.406182ms" to run "out/minikube-linux-arm64 -p functional-479614 service list -o json"
--- PASS: TestFunctional/parallel/ServiceCmd/JSONOutput (0.55s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmd/HTTPS (0.54s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmd/HTTPS
functional_test.go:1508: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 service --namespace=default --https --url hello-node
functional_test.go:1521: found endpoint: https://192.168.49.2:30804
--- PASS: TestFunctional/parallel/ServiceCmd/HTTPS (0.54s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmd/Format (0.53s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmd/Format
functional_test.go:1539: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 service hello-node --url --format={{.IP}}
--- PASS: TestFunctional/parallel/ServiceCmd/Format (0.53s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmd/URL (0.49s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmd/URL
functional_test.go:1558: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 service hello-node --url
functional_test.go:1564: found endpoint for hello-node: http://192.168.49.2:30804
--- PASS: TestFunctional/parallel/ServiceCmd/URL (0.49s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageRemove (0.68s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageRemove
functional_test.go:391: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image rm gcr.io/google-containers/addon-resizer:functional-479614 --alsologtostderr
functional_test.go:447: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image ls
--- PASS: TestFunctional/parallel/ImageCommands/ImageRemove (0.68s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageSaveDaemon (0.75s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageSaveDaemon
functional_test.go:418: (dbg) Run:  docker rmi gcr.io/google-containers/addon-resizer:functional-479614
functional_test.go:423: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 image save --daemon gcr.io/google-containers/addon-resizer:functional-479614 --alsologtostderr
functional_test.go:428: (dbg) Run:  docker image inspect gcr.io/google-containers/addon-resizer:functional-479614
--- PASS: TestFunctional/parallel/ImageCommands/ImageSaveDaemon (0.75s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/RunSecondTunnel (0.78s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/RunSecondTunnel
functional_test_tunnel_test.go:154: (dbg) daemon: [out/minikube-linux-arm64 -p functional-479614 tunnel --alsologtostderr]
functional_test_tunnel_test.go:154: (dbg) daemon: [out/minikube-linux-arm64 -p functional-479614 tunnel --alsologtostderr]
functional_test_tunnel_test.go:194: (dbg) stopping [out/minikube-linux-arm64 -p functional-479614 tunnel --alsologtostderr] ...
helpers_test.go:490: unable to find parent, assuming dead: process does not exist
functional_test_tunnel_test.go:194: (dbg) stopping [out/minikube-linux-arm64 -p functional-479614 tunnel --alsologtostderr] ...
helpers_test.go:508: unable to kill pid 1255023: os: process already finished
helpers_test.go:502: unable to terminate pid 1254892: os: process already finished
--- PASS: TestFunctional/parallel/TunnelCmd/serial/RunSecondTunnel (0.78s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/StartTunnel (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/StartTunnel
functional_test_tunnel_test.go:129: (dbg) daemon: [out/minikube-linux-arm64 -p functional-479614 tunnel --alsologtostderr]
--- PASS: TestFunctional/parallel/TunnelCmd/serial/StartTunnel (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/WaitService/Setup (8.44s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/WaitService/Setup
functional_test_tunnel_test.go:212: (dbg) Run:  kubectl --context functional-479614 apply -f testdata/testsvc.yaml
functional_test_tunnel_test.go:216: (dbg) TestFunctional/parallel/TunnelCmd/serial/WaitService/Setup: waiting 4m0s for pods matching "run=nginx-svc" in namespace "default" ...
helpers_test.go:344: "nginx-svc" [0fde6aa0-41b1-4894-8913-636ce18ad7a1] Pending / Ready:ContainersNotReady (containers with unready status: [nginx]) / ContainersReady:ContainersNotReady (containers with unready status: [nginx])
helpers_test.go:344: "nginx-svc" [0fde6aa0-41b1-4894-8913-636ce18ad7a1] Running
functional_test_tunnel_test.go:216: (dbg) TestFunctional/parallel/TunnelCmd/serial/WaitService/Setup: run=nginx-svc healthy within 8.017518933s
--- PASS: TestFunctional/parallel/TunnelCmd/serial/WaitService/Setup (8.44s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/WaitService/IngressIP (0.14s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/WaitService/IngressIP
functional_test_tunnel_test.go:234: (dbg) Run:  kubectl --context functional-479614 get svc nginx-svc -o jsonpath={.status.loadBalancer.ingress[0].ip}
--- PASS: TestFunctional/parallel/TunnelCmd/serial/WaitService/IngressIP (0.14s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/AccessDirect (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/AccessDirect
functional_test_tunnel_test.go:299: tunnel at http://10.96.134.44 is working!
--- PASS: TestFunctional/parallel/TunnelCmd/serial/AccessDirect (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/DeleteTunnel (0.11s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/DeleteTunnel
functional_test_tunnel_test.go:434: (dbg) stopping [out/minikube-linux-arm64 -p functional-479614 tunnel --alsologtostderr] ...
--- PASS: TestFunctional/parallel/TunnelCmd/serial/DeleteTunnel (0.11s)

                                                
                                    
x
+
TestFunctional/parallel/ProfileCmd/profile_not_create (0.58s)

                                                
                                                
=== RUN   TestFunctional/parallel/ProfileCmd/profile_not_create
functional_test.go:1269: (dbg) Run:  out/minikube-linux-arm64 profile lis
functional_test.go:1274: (dbg) Run:  out/minikube-linux-arm64 profile list --output json
--- PASS: TestFunctional/parallel/ProfileCmd/profile_not_create (0.58s)

                                                
                                    
x
+
TestFunctional/parallel/ProfileCmd/profile_list (0.52s)

                                                
                                                
=== RUN   TestFunctional/parallel/ProfileCmd/profile_list
functional_test.go:1309: (dbg) Run:  out/minikube-linux-arm64 profile list
functional_test.go:1314: Took "431.366569ms" to run "out/minikube-linux-arm64 profile list"
functional_test.go:1323: (dbg) Run:  out/minikube-linux-arm64 profile list -l
functional_test.go:1328: Took "83.974733ms" to run "out/minikube-linux-arm64 profile list -l"
--- PASS: TestFunctional/parallel/ProfileCmd/profile_list (0.52s)

                                                
                                    
x
+
TestFunctional/parallel/ProfileCmd/profile_json_output (0.47s)

                                                
                                                
=== RUN   TestFunctional/parallel/ProfileCmd/profile_json_output
functional_test.go:1360: (dbg) Run:  out/minikube-linux-arm64 profile list -o json
functional_test.go:1365: Took "393.982878ms" to run "out/minikube-linux-arm64 profile list -o json"
functional_test.go:1373: (dbg) Run:  out/minikube-linux-arm64 profile list -o json --light
functional_test.go:1378: Took "78.697959ms" to run "out/minikube-linux-arm64 profile list -o json --light"
--- PASS: TestFunctional/parallel/ProfileCmd/profile_json_output (0.47s)

                                                
                                    
x
+
TestFunctional/parallel/MountCmd/any-port (7.74s)

                                                
                                                
=== RUN   TestFunctional/parallel/MountCmd/any-port
functional_test_mount_test.go:73: (dbg) daemon: [out/minikube-linux-arm64 mount -p functional-479614 /tmp/TestFunctionalparallelMountCmdany-port1621359176/001:/mount-9p --alsologtostderr -v=1]
functional_test_mount_test.go:107: wrote "test-1693436597068207956" to /tmp/TestFunctionalparallelMountCmdany-port1621359176/001/created-by-test
functional_test_mount_test.go:107: wrote "test-1693436597068207956" to /tmp/TestFunctionalparallelMountCmdany-port1621359176/001/created-by-test-removed-by-pod
functional_test_mount_test.go:107: wrote "test-1693436597068207956" to /tmp/TestFunctionalparallelMountCmdany-port1621359176/001/test-1693436597068207956
functional_test_mount_test.go:115: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "findmnt -T /mount-9p | grep 9p"
functional_test_mount_test.go:115: (dbg) Non-zero exit: out/minikube-linux-arm64 -p functional-479614 ssh "findmnt -T /mount-9p | grep 9p": exit status 1 (470.6464ms)

                                                
                                                
** stderr ** 
	ssh: Process exited with status 1

                                                
                                                
** /stderr **
functional_test_mount_test.go:115: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "findmnt -T /mount-9p | grep 9p"
functional_test_mount_test.go:129: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh -- ls -la /mount-9p
functional_test_mount_test.go:133: guest mount directory contents
total 2
-rw-r--r-- 1 docker docker 24 Aug 30 23:03 created-by-test
-rw-r--r-- 1 docker docker 24 Aug 30 23:03 created-by-test-removed-by-pod
-rw-r--r-- 1 docker docker 24 Aug 30 23:03 test-1693436597068207956
functional_test_mount_test.go:137: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh cat /mount-9p/test-1693436597068207956
functional_test_mount_test.go:148: (dbg) Run:  kubectl --context functional-479614 replace --force -f testdata/busybox-mount-test.yaml
functional_test_mount_test.go:153: (dbg) TestFunctional/parallel/MountCmd/any-port: waiting 4m0s for pods matching "integration-test=busybox-mount" in namespace "default" ...
helpers_test.go:344: "busybox-mount" [0ef53f7c-6ab6-44ed-8175-245171b7fefe] Pending
helpers_test.go:344: "busybox-mount" [0ef53f7c-6ab6-44ed-8175-245171b7fefe] Pending / Ready:ContainersNotReady (containers with unready status: [mount-munger]) / ContainersReady:ContainersNotReady (containers with unready status: [mount-munger])
helpers_test.go:344: "busybox-mount" [0ef53f7c-6ab6-44ed-8175-245171b7fefe] Pending: Initialized:PodCompleted / Ready:PodCompleted / ContainersReady:PodCompleted
helpers_test.go:344: "busybox-mount" [0ef53f7c-6ab6-44ed-8175-245171b7fefe] Succeeded: Initialized:PodCompleted / Ready:PodCompleted / ContainersReady:PodCompleted
functional_test_mount_test.go:153: (dbg) TestFunctional/parallel/MountCmd/any-port: integration-test=busybox-mount healthy within 4.011706154s
functional_test_mount_test.go:169: (dbg) Run:  kubectl --context functional-479614 logs busybox-mount
functional_test_mount_test.go:181: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh stat /mount-9p/created-by-test
functional_test_mount_test.go:181: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh stat /mount-9p/created-by-pod
functional_test_mount_test.go:90: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "sudo umount -f /mount-9p"
functional_test_mount_test.go:94: (dbg) stopping [out/minikube-linux-arm64 mount -p functional-479614 /tmp/TestFunctionalparallelMountCmdany-port1621359176/001:/mount-9p --alsologtostderr -v=1] ...
--- PASS: TestFunctional/parallel/MountCmd/any-port (7.74s)

                                                
                                    
x
+
TestFunctional/parallel/MountCmd/specific-port (2.76s)

                                                
                                                
=== RUN   TestFunctional/parallel/MountCmd/specific-port
functional_test_mount_test.go:213: (dbg) daemon: [out/minikube-linux-arm64 mount -p functional-479614 /tmp/TestFunctionalparallelMountCmdspecific-port2624238856/001:/mount-9p --alsologtostderr -v=1 --port 46464]
functional_test_mount_test.go:243: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "findmnt -T /mount-9p | grep 9p"
functional_test_mount_test.go:243: (dbg) Non-zero exit: out/minikube-linux-arm64 -p functional-479614 ssh "findmnt -T /mount-9p | grep 9p": exit status 1 (684.081245ms)

                                                
                                                
** stderr ** 
	ssh: Process exited with status 1

                                                
                                                
** /stderr **
functional_test_mount_test.go:243: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "findmnt -T /mount-9p | grep 9p"
functional_test_mount_test.go:257: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh -- ls -la /mount-9p
functional_test_mount_test.go:261: guest mount directory contents
total 0
functional_test_mount_test.go:263: (dbg) stopping [out/minikube-linux-arm64 mount -p functional-479614 /tmp/TestFunctionalparallelMountCmdspecific-port2624238856/001:/mount-9p --alsologtostderr -v=1 --port 46464] ...
functional_test_mount_test.go:264: reading mount text
functional_test_mount_test.go:278: done reading mount text
functional_test_mount_test.go:230: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "sudo umount -f /mount-9p"
functional_test_mount_test.go:230: (dbg) Non-zero exit: out/minikube-linux-arm64 -p functional-479614 ssh "sudo umount -f /mount-9p": exit status 1 (429.054103ms)

                                                
                                                
-- stdout --
	umount: /mount-9p: not mounted.

                                                
                                                
-- /stdout --
** stderr ** 
	ssh: Process exited with status 32

                                                
                                                
** /stderr **
functional_test_mount_test.go:232: "out/minikube-linux-arm64 -p functional-479614 ssh \"sudo umount -f /mount-9p\"": exit status 1
functional_test_mount_test.go:234: (dbg) stopping [out/minikube-linux-arm64 mount -p functional-479614 /tmp/TestFunctionalparallelMountCmdspecific-port2624238856/001:/mount-9p --alsologtostderr -v=1 --port 46464] ...
--- PASS: TestFunctional/parallel/MountCmd/specific-port (2.76s)

                                                
                                    
x
+
TestFunctional/parallel/MountCmd/VerifyCleanup (2.18s)

                                                
                                                
=== RUN   TestFunctional/parallel/MountCmd/VerifyCleanup
functional_test_mount_test.go:298: (dbg) daemon: [out/minikube-linux-arm64 mount -p functional-479614 /tmp/TestFunctionalparallelMountCmdVerifyCleanup4213640551/001:/mount1 --alsologtostderr -v=1]
functional_test_mount_test.go:298: (dbg) daemon: [out/minikube-linux-arm64 mount -p functional-479614 /tmp/TestFunctionalparallelMountCmdVerifyCleanup4213640551/001:/mount2 --alsologtostderr -v=1]
functional_test_mount_test.go:298: (dbg) daemon: [out/minikube-linux-arm64 mount -p functional-479614 /tmp/TestFunctionalparallelMountCmdVerifyCleanup4213640551/001:/mount3 --alsologtostderr -v=1]
functional_test_mount_test.go:325: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "findmnt -T" /mount1
E0830 23:03:28.069281 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
functional_test_mount_test.go:325: (dbg) Done: out/minikube-linux-arm64 -p functional-479614 ssh "findmnt -T" /mount1: (1.205496766s)
functional_test_mount_test.go:325: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "findmnt -T" /mount2
functional_test_mount_test.go:325: (dbg) Run:  out/minikube-linux-arm64 -p functional-479614 ssh "findmnt -T" /mount3
functional_test_mount_test.go:370: (dbg) Run:  out/minikube-linux-arm64 mount -p functional-479614 --kill=true
functional_test_mount_test.go:313: (dbg) stopping [out/minikube-linux-arm64 mount -p functional-479614 /tmp/TestFunctionalparallelMountCmdVerifyCleanup4213640551/001:/mount1 --alsologtostderr -v=1] ...
helpers_test.go:490: unable to find parent, assuming dead: process does not exist
functional_test_mount_test.go:313: (dbg) stopping [out/minikube-linux-arm64 mount -p functional-479614 /tmp/TestFunctionalparallelMountCmdVerifyCleanup4213640551/001:/mount2 --alsologtostderr -v=1] ...
helpers_test.go:490: unable to find parent, assuming dead: process does not exist
functional_test_mount_test.go:313: (dbg) stopping [out/minikube-linux-arm64 mount -p functional-479614 /tmp/TestFunctionalparallelMountCmdVerifyCleanup4213640551/001:/mount3 --alsologtostderr -v=1] ...
helpers_test.go:490: unable to find parent, assuming dead: process does not exist
--- PASS: TestFunctional/parallel/MountCmd/VerifyCleanup (2.18s)

                                                
                                    
x
+
TestFunctional/delete_addon-resizer_images (0.08s)

                                                
                                                
=== RUN   TestFunctional/delete_addon-resizer_images
functional_test.go:189: (dbg) Run:  docker rmi -f gcr.io/google-containers/addon-resizer:1.8.8
functional_test.go:189: (dbg) Run:  docker rmi -f gcr.io/google-containers/addon-resizer:functional-479614
--- PASS: TestFunctional/delete_addon-resizer_images (0.08s)

                                                
                                    
x
+
TestFunctional/delete_my-image_image (0.02s)

                                                
                                                
=== RUN   TestFunctional/delete_my-image_image
functional_test.go:197: (dbg) Run:  docker rmi -f localhost/my-image:functional-479614
--- PASS: TestFunctional/delete_my-image_image (0.02s)

                                                
                                    
x
+
TestFunctional/delete_minikube_cached_images (0.02s)

                                                
                                                
=== RUN   TestFunctional/delete_minikube_cached_images
functional_test.go:205: (dbg) Run:  docker rmi -f minikube-local-cache-test:functional-479614
--- PASS: TestFunctional/delete_minikube_cached_images (0.02s)

                                                
                                    
x
+
TestIngressAddonLegacy/StartLegacyK8sCluster (94.83s)

                                                
                                                
=== RUN   TestIngressAddonLegacy/StartLegacyK8sCluster
ingress_addon_legacy_test.go:39: (dbg) Run:  out/minikube-linux-arm64 start -p ingress-addon-legacy-235029 --kubernetes-version=v1.18.20 --memory=4096 --wait=true --alsologtostderr -v=5 --driver=docker  --container-runtime=containerd
E0830 23:04:49.990285 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
ingress_addon_legacy_test.go:39: (dbg) Done: out/minikube-linux-arm64 start -p ingress-addon-legacy-235029 --kubernetes-version=v1.18.20 --memory=4096 --wait=true --alsologtostderr -v=5 --driver=docker  --container-runtime=containerd: (1m34.833116972s)
--- PASS: TestIngressAddonLegacy/StartLegacyK8sCluster (94.83s)

                                                
                                    
x
+
TestIngressAddonLegacy/serial/ValidateIngressAddonActivation (10s)

                                                
                                                
=== RUN   TestIngressAddonLegacy/serial/ValidateIngressAddonActivation
ingress_addon_legacy_test.go:70: (dbg) Run:  out/minikube-linux-arm64 -p ingress-addon-legacy-235029 addons enable ingress --alsologtostderr -v=5
ingress_addon_legacy_test.go:70: (dbg) Done: out/minikube-linux-arm64 -p ingress-addon-legacy-235029 addons enable ingress --alsologtostderr -v=5: (10.002504903s)
--- PASS: TestIngressAddonLegacy/serial/ValidateIngressAddonActivation (10.00s)

                                                
                                    
x
+
TestIngressAddonLegacy/serial/ValidateIngressDNSAddonActivation (0.67s)

                                                
                                                
=== RUN   TestIngressAddonLegacy/serial/ValidateIngressDNSAddonActivation
ingress_addon_legacy_test.go:79: (dbg) Run:  out/minikube-linux-arm64 -p ingress-addon-legacy-235029 addons enable ingress-dns --alsologtostderr -v=5
--- PASS: TestIngressAddonLegacy/serial/ValidateIngressDNSAddonActivation (0.67s)

                                                
                                    
x
+
TestJSONOutput/start/Command (85.64s)

                                                
                                                
=== RUN   TestJSONOutput/start/Command
json_output_test.go:63: (dbg) Run:  out/minikube-linux-arm64 start -p json-output-622675 --output=json --user=testUser --memory=2200 --wait=true --driver=docker  --container-runtime=containerd
E0830 23:07:06.147554 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:07:33.830526 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:07:44.061593 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:07:44.066890 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:07:44.077220 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:07:44.097469 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:07:44.137735 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:07:44.217975 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:07:44.378466 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:07:44.699037 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:07:45.340092 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:07:46.620867 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
json_output_test.go:63: (dbg) Done: out/minikube-linux-arm64 start -p json-output-622675 --output=json --user=testUser --memory=2200 --wait=true --driver=docker  --container-runtime=containerd: (1m25.634610098s)
--- PASS: TestJSONOutput/start/Command (85.64s)

                                                
                                    
x
+
TestJSONOutput/start/Audit (0s)

                                                
                                                
=== RUN   TestJSONOutput/start/Audit
--- PASS: TestJSONOutput/start/Audit (0.00s)

                                                
                                    
x
+
TestJSONOutput/start/parallel/DistinctCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/start/parallel/DistinctCurrentSteps
=== PAUSE TestJSONOutput/start/parallel/DistinctCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/start/parallel/DistinctCurrentSteps
--- PASS: TestJSONOutput/start/parallel/DistinctCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/start/parallel/IncreasingCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/start/parallel/IncreasingCurrentSteps
=== PAUSE TestJSONOutput/start/parallel/IncreasingCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/start/parallel/IncreasingCurrentSteps
--- PASS: TestJSONOutput/start/parallel/IncreasingCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/pause/Command (0.85s)

                                                
                                                
=== RUN   TestJSONOutput/pause/Command
json_output_test.go:63: (dbg) Run:  out/minikube-linux-arm64 pause -p json-output-622675 --output=json --user=testUser
--- PASS: TestJSONOutput/pause/Command (0.85s)

                                                
                                    
x
+
TestJSONOutput/pause/Audit (0s)

                                                
                                                
=== RUN   TestJSONOutput/pause/Audit
--- PASS: TestJSONOutput/pause/Audit (0.00s)

                                                
                                    
x
+
TestJSONOutput/pause/parallel/DistinctCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/pause/parallel/DistinctCurrentSteps
=== PAUSE TestJSONOutput/pause/parallel/DistinctCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/pause/parallel/DistinctCurrentSteps
--- PASS: TestJSONOutput/pause/parallel/DistinctCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/pause/parallel/IncreasingCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/pause/parallel/IncreasingCurrentSteps
=== PAUSE TestJSONOutput/pause/parallel/IncreasingCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/pause/parallel/IncreasingCurrentSteps
--- PASS: TestJSONOutput/pause/parallel/IncreasingCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/unpause/Command (0.74s)

                                                
                                                
=== RUN   TestJSONOutput/unpause/Command
json_output_test.go:63: (dbg) Run:  out/minikube-linux-arm64 unpause -p json-output-622675 --output=json --user=testUser
E0830 23:07:49.181927 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
--- PASS: TestJSONOutput/unpause/Command (0.74s)

                                                
                                    
x
+
TestJSONOutput/unpause/Audit (0s)

                                                
                                                
=== RUN   TestJSONOutput/unpause/Audit
--- PASS: TestJSONOutput/unpause/Audit (0.00s)

                                                
                                    
x
+
TestJSONOutput/unpause/parallel/DistinctCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/unpause/parallel/DistinctCurrentSteps
=== PAUSE TestJSONOutput/unpause/parallel/DistinctCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/unpause/parallel/DistinctCurrentSteps
--- PASS: TestJSONOutput/unpause/parallel/DistinctCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/unpause/parallel/IncreasingCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/unpause/parallel/IncreasingCurrentSteps
=== PAUSE TestJSONOutput/unpause/parallel/IncreasingCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/unpause/parallel/IncreasingCurrentSteps
--- PASS: TestJSONOutput/unpause/parallel/IncreasingCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/stop/Command (1.32s)

                                                
                                                
=== RUN   TestJSONOutput/stop/Command
json_output_test.go:63: (dbg) Run:  out/minikube-linux-arm64 stop -p json-output-622675 --output=json --user=testUser
json_output_test.go:63: (dbg) Done: out/minikube-linux-arm64 stop -p json-output-622675 --output=json --user=testUser: (1.318977995s)
--- PASS: TestJSONOutput/stop/Command (1.32s)

                                                
                                    
x
+
TestJSONOutput/stop/Audit (0s)

                                                
                                                
=== RUN   TestJSONOutput/stop/Audit
--- PASS: TestJSONOutput/stop/Audit (0.00s)

                                                
                                    
x
+
TestJSONOutput/stop/parallel/DistinctCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/stop/parallel/DistinctCurrentSteps
=== PAUSE TestJSONOutput/stop/parallel/DistinctCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/stop/parallel/DistinctCurrentSteps
--- PASS: TestJSONOutput/stop/parallel/DistinctCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/stop/parallel/IncreasingCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/stop/parallel/IncreasingCurrentSteps
=== PAUSE TestJSONOutput/stop/parallel/IncreasingCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/stop/parallel/IncreasingCurrentSteps
--- PASS: TestJSONOutput/stop/parallel/IncreasingCurrentSteps (0.00s)

                                                
                                    
x
+
TestErrorJSONOutput (0.26s)

                                                
                                                
=== RUN   TestErrorJSONOutput
json_output_test.go:160: (dbg) Run:  out/minikube-linux-arm64 start -p json-output-error-672109 --memory=2200 --output=json --wait=true --driver=fail
json_output_test.go:160: (dbg) Non-zero exit: out/minikube-linux-arm64 start -p json-output-error-672109 --memory=2200 --output=json --wait=true --driver=fail: exit status 56 (101.693998ms)

                                                
                                                
-- stdout --
	{"specversion":"1.0","id":"3bd0eeec-b945-4339-b579-907e52594b31","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.step","datacontenttype":"application/json","data":{"currentstep":"0","message":"[json-output-error-672109] minikube v1.31.2 on Ubuntu 20.04 (arm64)","name":"Initial Minikube Setup","totalsteps":"19"}}
	{"specversion":"1.0","id":"b30d3901-8b97-4370-a6a4-112788805ac1","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_LOCATION=17114"}}
	{"specversion":"1.0","id":"7fbbe0f9-b997-428a-820b-04e1a4a07c67","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true"}}
	{"specversion":"1.0","id":"8fa00eb4-1e3a-4573-a4c5-7b4003f2f18b","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig"}}
	{"specversion":"1.0","id":"a6408cf9-22ca-42e2-970b-7a635a97397b","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube"}}
	{"specversion":"1.0","id":"4e534c32-0c28-4c25-9083-0a334b2041a7","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_BIN=out/minikube-linux-arm64"}}
	{"specversion":"1.0","id":"c0a089f2-b0c7-429b-a76f-20252ba3b519","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_FORCE_SYSTEMD="}}
	{"specversion":"1.0","id":"3b6015c4-5e4a-4805-ae67-d12a928ccc36","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.error","datacontenttype":"application/json","data":{"advice":"","exitcode":"56","issues":"","message":"The driver 'fail' is not supported on linux/arm64","name":"DRV_UNSUPPORTED_OS","url":""}}

                                                
                                                
-- /stdout --
helpers_test.go:175: Cleaning up "json-output-error-672109" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p json-output-error-672109
--- PASS: TestErrorJSONOutput (0.26s)

                                                
                                    
x
+
TestKicCustomNetwork/create_custom_network (44.9s)

                                                
                                                
=== RUN   TestKicCustomNetwork/create_custom_network
kic_custom_network_test.go:57: (dbg) Run:  out/minikube-linux-arm64 start -p docker-network-007343 --network=
E0830 23:08:04.542403 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:08:25.022659 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
kic_custom_network_test.go:57: (dbg) Done: out/minikube-linux-arm64 start -p docker-network-007343 --network=: (42.65785038s)
kic_custom_network_test.go:150: (dbg) Run:  docker network ls --format {{.Name}}
helpers_test.go:175: Cleaning up "docker-network-007343" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p docker-network-007343
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p docker-network-007343: (2.214856514s)
--- PASS: TestKicCustomNetwork/create_custom_network (44.90s)

                                                
                                    
x
+
TestKicCustomNetwork/use_default_bridge_network (33.12s)

                                                
                                                
=== RUN   TestKicCustomNetwork/use_default_bridge_network
kic_custom_network_test.go:57: (dbg) Run:  out/minikube-linux-arm64 start -p docker-network-314442 --network=bridge
E0830 23:09:05.983392 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
kic_custom_network_test.go:57: (dbg) Done: out/minikube-linux-arm64 start -p docker-network-314442 --network=bridge: (31.080979652s)
kic_custom_network_test.go:150: (dbg) Run:  docker network ls --format {{.Name}}
helpers_test.go:175: Cleaning up "docker-network-314442" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p docker-network-314442
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p docker-network-314442: (2.016520753s)
--- PASS: TestKicCustomNetwork/use_default_bridge_network (33.12s)

                                                
                                    
x
+
TestKicExistingNetwork (32.97s)

                                                
                                                
=== RUN   TestKicExistingNetwork
kic_custom_network_test.go:150: (dbg) Run:  docker network ls --format {{.Name}}
kic_custom_network_test.go:93: (dbg) Run:  out/minikube-linux-arm64 start -p existing-network-013440 --network=existing-network
kic_custom_network_test.go:93: (dbg) Done: out/minikube-linux-arm64 start -p existing-network-013440 --network=existing-network: (30.86254389s)
helpers_test.go:175: Cleaning up "existing-network-013440" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p existing-network-013440
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p existing-network-013440: (1.952443221s)
--- PASS: TestKicExistingNetwork (32.97s)

                                                
                                    
x
+
TestKicCustomSubnet (35.39s)

                                                
                                                
=== RUN   TestKicCustomSubnet
kic_custom_network_test.go:112: (dbg) Run:  out/minikube-linux-arm64 start -p custom-subnet-057665 --subnet=192.168.60.0/24
kic_custom_network_test.go:112: (dbg) Done: out/minikube-linux-arm64 start -p custom-subnet-057665 --subnet=192.168.60.0/24: (33.271597876s)
kic_custom_network_test.go:161: (dbg) Run:  docker network inspect custom-subnet-057665 --format "{{(index .IPAM.Config 0).Subnet}}"
helpers_test.go:175: Cleaning up "custom-subnet-057665" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p custom-subnet-057665
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p custom-subnet-057665: (2.089814347s)
--- PASS: TestKicCustomSubnet (35.39s)

                                                
                                    
x
+
TestKicStaticIP (36.74s)

                                                
                                                
=== RUN   TestKicStaticIP
kic_custom_network_test.go:132: (dbg) Run:  out/minikube-linux-arm64 start -p static-ip-241388 --static-ip=192.168.200.200
E0830 23:10:24.312310 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:10:24.317562 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:10:24.327832 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:10:24.348079 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:10:24.388331 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:10:24.469005 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:10:24.629358 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:10:24.949919 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:10:25.590811 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:10:26.871029 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:10:27.903771 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:10:29.431259 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:10:34.551820 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:10:44.792553 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
kic_custom_network_test.go:132: (dbg) Done: out/minikube-linux-arm64 start -p static-ip-241388 --static-ip=192.168.200.200: (34.410531962s)
kic_custom_network_test.go:138: (dbg) Run:  out/minikube-linux-arm64 -p static-ip-241388 ip
helpers_test.go:175: Cleaning up "static-ip-241388" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p static-ip-241388
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p static-ip-241388: (2.154565199s)
--- PASS: TestKicStaticIP (36.74s)

                                                
                                    
x
+
TestMainNoArgs (0.07s)

                                                
                                                
=== RUN   TestMainNoArgs
main_test.go:68: (dbg) Run:  out/minikube-linux-arm64
--- PASS: TestMainNoArgs (0.07s)

                                                
                                    
x
+
TestMinikubeProfile (73.55s)

                                                
                                                
=== RUN   TestMinikubeProfile
minikube_profile_test.go:44: (dbg) Run:  out/minikube-linux-arm64 start -p first-099126 --driver=docker  --container-runtime=containerd
E0830 23:11:05.273242 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
minikube_profile_test.go:44: (dbg) Done: out/minikube-linux-arm64 start -p first-099126 --driver=docker  --container-runtime=containerd: (33.164359545s)
minikube_profile_test.go:44: (dbg) Run:  out/minikube-linux-arm64 start -p second-101850 --driver=docker  --container-runtime=containerd
E0830 23:11:46.234363 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:12:06.147681 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
minikube_profile_test.go:44: (dbg) Done: out/minikube-linux-arm64 start -p second-101850 --driver=docker  --container-runtime=containerd: (34.767243759s)
minikube_profile_test.go:51: (dbg) Run:  out/minikube-linux-arm64 profile first-099126
minikube_profile_test.go:55: (dbg) Run:  out/minikube-linux-arm64 profile list -ojson
minikube_profile_test.go:51: (dbg) Run:  out/minikube-linux-arm64 profile second-101850
minikube_profile_test.go:55: (dbg) Run:  out/minikube-linux-arm64 profile list -ojson
helpers_test.go:175: Cleaning up "second-101850" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p second-101850
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p second-101850: (2.009839739s)
helpers_test.go:175: Cleaning up "first-099126" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p first-099126
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p first-099126: (2.251551461s)
--- PASS: TestMinikubeProfile (73.55s)

                                                
                                    
x
+
TestMountStart/serial/StartWithMountFirst (7.03s)

                                                
                                                
=== RUN   TestMountStart/serial/StartWithMountFirst
mount_start_test.go:98: (dbg) Run:  out/minikube-linux-arm64 start -p mount-start-1-199343 --memory=2048 --mount --mount-gid 0 --mount-msize 6543 --mount-port 46464 --mount-uid 0 --no-kubernetes --driver=docker  --container-runtime=containerd
mount_start_test.go:98: (dbg) Done: out/minikube-linux-arm64 start -p mount-start-1-199343 --memory=2048 --mount --mount-gid 0 --mount-msize 6543 --mount-port 46464 --mount-uid 0 --no-kubernetes --driver=docker  --container-runtime=containerd: (6.029943987s)
--- PASS: TestMountStart/serial/StartWithMountFirst (7.03s)

                                                
                                    
x
+
TestMountStart/serial/VerifyMountFirst (0.29s)

                                                
                                                
=== RUN   TestMountStart/serial/VerifyMountFirst
mount_start_test.go:114: (dbg) Run:  out/minikube-linux-arm64 -p mount-start-1-199343 ssh -- ls /minikube-host
--- PASS: TestMountStart/serial/VerifyMountFirst (0.29s)

                                                
                                    
x
+
TestMountStart/serial/StartWithMountSecond (9.68s)

                                                
                                                
=== RUN   TestMountStart/serial/StartWithMountSecond
mount_start_test.go:98: (dbg) Run:  out/minikube-linux-arm64 start -p mount-start-2-201460 --memory=2048 --mount --mount-gid 0 --mount-msize 6543 --mount-port 46465 --mount-uid 0 --no-kubernetes --driver=docker  --container-runtime=containerd
mount_start_test.go:98: (dbg) Done: out/minikube-linux-arm64 start -p mount-start-2-201460 --memory=2048 --mount --mount-gid 0 --mount-msize 6543 --mount-port 46465 --mount-uid 0 --no-kubernetes --driver=docker  --container-runtime=containerd: (8.679926884s)
--- PASS: TestMountStart/serial/StartWithMountSecond (9.68s)

                                                
                                    
x
+
TestMountStart/serial/VerifyMountSecond (0.31s)

                                                
                                                
=== RUN   TestMountStart/serial/VerifyMountSecond
mount_start_test.go:114: (dbg) Run:  out/minikube-linux-arm64 -p mount-start-2-201460 ssh -- ls /minikube-host
--- PASS: TestMountStart/serial/VerifyMountSecond (0.31s)

                                                
                                    
x
+
TestMountStart/serial/DeleteFirst (1.7s)

                                                
                                                
=== RUN   TestMountStart/serial/DeleteFirst
pause_test.go:132: (dbg) Run:  out/minikube-linux-arm64 delete -p mount-start-1-199343 --alsologtostderr -v=5
pause_test.go:132: (dbg) Done: out/minikube-linux-arm64 delete -p mount-start-1-199343 --alsologtostderr -v=5: (1.702742008s)
--- PASS: TestMountStart/serial/DeleteFirst (1.70s)

                                                
                                    
x
+
TestMountStart/serial/VerifyMountPostDelete (0.29s)

                                                
                                                
=== RUN   TestMountStart/serial/VerifyMountPostDelete
mount_start_test.go:114: (dbg) Run:  out/minikube-linux-arm64 -p mount-start-2-201460 ssh -- ls /minikube-host
--- PASS: TestMountStart/serial/VerifyMountPostDelete (0.29s)

                                                
                                    
x
+
TestMountStart/serial/Stop (1.24s)

                                                
                                                
=== RUN   TestMountStart/serial/Stop
mount_start_test.go:155: (dbg) Run:  out/minikube-linux-arm64 stop -p mount-start-2-201460
mount_start_test.go:155: (dbg) Done: out/minikube-linux-arm64 stop -p mount-start-2-201460: (1.238930795s)
--- PASS: TestMountStart/serial/Stop (1.24s)

                                                
                                    
x
+
TestMountStart/serial/RestartStopped (7.66s)

                                                
                                                
=== RUN   TestMountStart/serial/RestartStopped
mount_start_test.go:166: (dbg) Run:  out/minikube-linux-arm64 start -p mount-start-2-201460
mount_start_test.go:166: (dbg) Done: out/minikube-linux-arm64 start -p mount-start-2-201460: (6.659147983s)
--- PASS: TestMountStart/serial/RestartStopped (7.66s)

                                                
                                    
x
+
TestMountStart/serial/VerifyMountPostStop (0.3s)

                                                
                                                
=== RUN   TestMountStart/serial/VerifyMountPostStop
mount_start_test.go:114: (dbg) Run:  out/minikube-linux-arm64 -p mount-start-2-201460 ssh -- ls /minikube-host
--- PASS: TestMountStart/serial/VerifyMountPostStop (0.30s)

                                                
                                    
x
+
TestMultiNode/serial/FreshStart2Nodes (77.75s)

                                                
                                                
=== RUN   TestMultiNode/serial/FreshStart2Nodes
multinode_test.go:85: (dbg) Run:  out/minikube-linux-arm64 start -p multinode-961490 --wait=true --memory=2200 --nodes=2 -v=8 --alsologtostderr --driver=docker  --container-runtime=containerd
E0830 23:12:44.061447 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:13:08.154572 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:13:11.743984 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
multinode_test.go:85: (dbg) Done: out/minikube-linux-arm64 start -p multinode-961490 --wait=true --memory=2200 --nodes=2 -v=8 --alsologtostderr --driver=docker  --container-runtime=containerd: (1m16.996000485s)
multinode_test.go:91: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 status --alsologtostderr
--- PASS: TestMultiNode/serial/FreshStart2Nodes (77.75s)

                                                
                                    
x
+
TestMultiNode/serial/DeployApp2Nodes (4.92s)

                                                
                                                
=== RUN   TestMultiNode/serial/DeployApp2Nodes
multinode_test.go:481: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- apply -f ./testdata/multinodes/multinode-pod-dns-test.yaml
multinode_test.go:486: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- rollout status deployment/busybox
multinode_test.go:486: (dbg) Done: out/minikube-linux-arm64 kubectl -p multinode-961490 -- rollout status deployment/busybox: (2.635249509s)
multinode_test.go:493: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- get pods -o jsonpath='{.items[*].status.podIP}'
multinode_test.go:516: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- get pods -o jsonpath='{.items[*].metadata.name}'
multinode_test.go:524: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- exec busybox-5bc68d56bd-b6z78 -- nslookup kubernetes.io
multinode_test.go:524: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- exec busybox-5bc68d56bd-lqbhb -- nslookup kubernetes.io
multinode_test.go:534: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- exec busybox-5bc68d56bd-b6z78 -- nslookup kubernetes.default
multinode_test.go:534: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- exec busybox-5bc68d56bd-lqbhb -- nslookup kubernetes.default
multinode_test.go:542: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- exec busybox-5bc68d56bd-b6z78 -- nslookup kubernetes.default.svc.cluster.local
multinode_test.go:542: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- exec busybox-5bc68d56bd-lqbhb -- nslookup kubernetes.default.svc.cluster.local
--- PASS: TestMultiNode/serial/DeployApp2Nodes (4.92s)

                                                
                                    
x
+
TestMultiNode/serial/PingHostFrom2Pods (1.17s)

                                                
                                                
=== RUN   TestMultiNode/serial/PingHostFrom2Pods
multinode_test.go:552: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- get pods -o jsonpath='{.items[*].metadata.name}'
multinode_test.go:560: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- exec busybox-5bc68d56bd-b6z78 -- sh -c "nslookup host.minikube.internal | awk 'NR==5' | cut -d' ' -f3"
multinode_test.go:571: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- exec busybox-5bc68d56bd-b6z78 -- sh -c "ping -c 1 192.168.58.1"
multinode_test.go:560: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- exec busybox-5bc68d56bd-lqbhb -- sh -c "nslookup host.minikube.internal | awk 'NR==5' | cut -d' ' -f3"
multinode_test.go:571: (dbg) Run:  out/minikube-linux-arm64 kubectl -p multinode-961490 -- exec busybox-5bc68d56bd-lqbhb -- sh -c "ping -c 1 192.168.58.1"
--- PASS: TestMultiNode/serial/PingHostFrom2Pods (1.17s)

                                                
                                    
x
+
TestMultiNode/serial/AddNode (21.69s)

                                                
                                                
=== RUN   TestMultiNode/serial/AddNode
multinode_test.go:110: (dbg) Run:  out/minikube-linux-arm64 node add -p multinode-961490 -v 3 --alsologtostderr
multinode_test.go:110: (dbg) Done: out/minikube-linux-arm64 node add -p multinode-961490 -v 3 --alsologtostderr: (20.927780181s)
multinode_test.go:116: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 status --alsologtostderr
--- PASS: TestMultiNode/serial/AddNode (21.69s)

                                                
                                    
x
+
TestMultiNode/serial/ProfileList (0.37s)

                                                
                                                
=== RUN   TestMultiNode/serial/ProfileList
multinode_test.go:132: (dbg) Run:  out/minikube-linux-arm64 profile list --output json
--- PASS: TestMultiNode/serial/ProfileList (0.37s)

                                                
                                    
x
+
TestMultiNode/serial/CopyFile (11.52s)

                                                
                                                
=== RUN   TestMultiNode/serial/CopyFile
multinode_test.go:173: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 status --output json --alsologtostderr
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 cp testdata/cp-test.txt multinode-961490:/home/docker/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 cp multinode-961490:/home/docker/cp-test.txt /tmp/TestMultiNodeserialCopyFile2771223651/001/cp-test_multinode-961490.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 cp multinode-961490:/home/docker/cp-test.txt multinode-961490-m02:/home/docker/cp-test_multinode-961490_multinode-961490-m02.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490-m02 "sudo cat /home/docker/cp-test_multinode-961490_multinode-961490-m02.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 cp multinode-961490:/home/docker/cp-test.txt multinode-961490-m03:/home/docker/cp-test_multinode-961490_multinode-961490-m03.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490-m03 "sudo cat /home/docker/cp-test_multinode-961490_multinode-961490-m03.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 cp testdata/cp-test.txt multinode-961490-m02:/home/docker/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490-m02 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 cp multinode-961490-m02:/home/docker/cp-test.txt /tmp/TestMultiNodeserialCopyFile2771223651/001/cp-test_multinode-961490-m02.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490-m02 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 cp multinode-961490-m02:/home/docker/cp-test.txt multinode-961490:/home/docker/cp-test_multinode-961490-m02_multinode-961490.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490-m02 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490 "sudo cat /home/docker/cp-test_multinode-961490-m02_multinode-961490.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 cp multinode-961490-m02:/home/docker/cp-test.txt multinode-961490-m03:/home/docker/cp-test_multinode-961490-m02_multinode-961490-m03.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490-m02 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490-m03 "sudo cat /home/docker/cp-test_multinode-961490-m02_multinode-961490-m03.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 cp testdata/cp-test.txt multinode-961490-m03:/home/docker/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490-m03 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 cp multinode-961490-m03:/home/docker/cp-test.txt /tmp/TestMultiNodeserialCopyFile2771223651/001/cp-test_multinode-961490-m03.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490-m03 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 cp multinode-961490-m03:/home/docker/cp-test.txt multinode-961490:/home/docker/cp-test_multinode-961490-m03_multinode-961490.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490-m03 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490 "sudo cat /home/docker/cp-test_multinode-961490-m03_multinode-961490.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 cp multinode-961490-m03:/home/docker/cp-test.txt multinode-961490-m02:/home/docker/cp-test_multinode-961490-m03_multinode-961490-m02.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490-m03 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 ssh -n multinode-961490-m02 "sudo cat /home/docker/cp-test_multinode-961490-m03_multinode-961490-m02.txt"
--- PASS: TestMultiNode/serial/CopyFile (11.52s)

                                                
                                    
x
+
TestMultiNode/serial/StopNode (2.46s)

                                                
                                                
=== RUN   TestMultiNode/serial/StopNode
multinode_test.go:210: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 node stop m03
multinode_test.go:210: (dbg) Done: out/minikube-linux-arm64 -p multinode-961490 node stop m03: (1.275270209s)
multinode_test.go:216: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 status
multinode_test.go:216: (dbg) Non-zero exit: out/minikube-linux-arm64 -p multinode-961490 status: exit status 7 (565.812304ms)

                                                
                                                
-- stdout --
	multinode-961490
	type: Control Plane
	host: Running
	kubelet: Running
	apiserver: Running
	kubeconfig: Configured
	
	multinode-961490-m02
	type: Worker
	host: Running
	kubelet: Running
	
	multinode-961490-m03
	type: Worker
	host: Stopped
	kubelet: Stopped
	

                                                
                                                
-- /stdout --
multinode_test.go:223: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 status --alsologtostderr
multinode_test.go:223: (dbg) Non-zero exit: out/minikube-linux-arm64 -p multinode-961490 status --alsologtostderr: exit status 7 (614.687247ms)

                                                
                                                
-- stdout --
	multinode-961490
	type: Control Plane
	host: Running
	kubelet: Running
	apiserver: Running
	kubeconfig: Configured
	
	multinode-961490-m02
	type: Worker
	host: Running
	kubelet: Running
	
	multinode-961490-m03
	type: Worker
	host: Stopped
	kubelet: Stopped
	

                                                
                                                
-- /stdout --
** stderr ** 
	I0830 23:14:41.985821 1305486 out.go:296] Setting OutFile to fd 1 ...
	I0830 23:14:41.986094 1305486 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:14:41.986117 1305486 out.go:309] Setting ErrFile to fd 2...
	I0830 23:14:41.986136 1305486 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:14:41.986505 1305486 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
	I0830 23:14:41.986759 1305486 out.go:303] Setting JSON to false
	I0830 23:14:41.986861 1305486 mustload.go:65] Loading cluster: multinode-961490
	I0830 23:14:41.987025 1305486 notify.go:220] Checking for updates...
	I0830 23:14:41.987366 1305486 config.go:182] Loaded profile config "multinode-961490": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
	I0830 23:14:41.987468 1305486 status.go:255] checking status of multinode-961490 ...
	I0830 23:14:41.988073 1305486 cli_runner.go:164] Run: docker container inspect multinode-961490 --format={{.State.Status}}
	I0830 23:14:42.008657 1305486 status.go:330] multinode-961490 host status = "Running" (err=<nil>)
	I0830 23:14:42.008682 1305486 host.go:66] Checking if "multinode-961490" exists ...
	I0830 23:14:42.008993 1305486 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" multinode-961490
	I0830 23:14:42.027718 1305486 host.go:66] Checking if "multinode-961490" exists ...
	I0830 23:14:42.028119 1305486 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0830 23:14:42.028179 1305486 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" multinode-961490
	I0830 23:14:42.057502 1305486 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34394 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/multinode-961490/id_rsa Username:docker}
	I0830 23:14:42.160831 1305486 ssh_runner.go:195] Run: systemctl --version
	I0830 23:14:42.167536 1305486 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0830 23:14:42.183697 1305486 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 23:14:42.275001 1305486 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:3 ContainersRunning:2 ContainersPaused:0 ContainersStopped:1 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:40 OomKillDisable:true NGoroutines:55 SystemTime:2023-08-30 23:14:42.258328443 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 23:14:42.275614 1305486 kubeconfig.go:92] found "multinode-961490" server: "https://192.168.58.2:8443"
	I0830 23:14:42.275632 1305486 api_server.go:166] Checking apiserver status ...
	I0830 23:14:42.275674 1305486 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0830 23:14:42.293394 1305486 ssh_runner.go:195] Run: sudo egrep ^[0-9]+:freezer: /proc/1255/cgroup
	I0830 23:14:42.306506 1305486 api_server.go:182] apiserver freezer: "8:freezer:/docker/c7448df3b36f5b1dda6dfd7f22addb95a542a29f53c599979eb1444df4f73cf8/kubepods/burstable/pod2317eabe6d9da84e912275656cafa471/f179df86f9798c4b35e40f21202ce3a51d07f4f7ee2100cec27a557bc9116903"
	I0830 23:14:42.306580 1305486 ssh_runner.go:195] Run: sudo cat /sys/fs/cgroup/freezer/docker/c7448df3b36f5b1dda6dfd7f22addb95a542a29f53c599979eb1444df4f73cf8/kubepods/burstable/pod2317eabe6d9da84e912275656cafa471/f179df86f9798c4b35e40f21202ce3a51d07f4f7ee2100cec27a557bc9116903/freezer.state
	I0830 23:14:42.318944 1305486 api_server.go:204] freezer state: "THAWED"
	I0830 23:14:42.318976 1305486 api_server.go:253] Checking apiserver healthz at https://192.168.58.2:8443/healthz ...
	I0830 23:14:42.328359 1305486 api_server.go:279] https://192.168.58.2:8443/healthz returned 200:
	ok
	I0830 23:14:42.328387 1305486 status.go:421] multinode-961490 apiserver status = Running (err=<nil>)
	I0830 23:14:42.328407 1305486 status.go:257] multinode-961490 status: &{Name:multinode-961490 Host:Running Kubelet:Running APIServer:Running Kubeconfig:Configured Worker:false TimeToStop: DockerEnv: PodManEnv:}
	I0830 23:14:42.328431 1305486 status.go:255] checking status of multinode-961490-m02 ...
	I0830 23:14:42.328753 1305486 cli_runner.go:164] Run: docker container inspect multinode-961490-m02 --format={{.State.Status}}
	I0830 23:14:42.347226 1305486 status.go:330] multinode-961490-m02 host status = "Running" (err=<nil>)
	I0830 23:14:42.347253 1305486 host.go:66] Checking if "multinode-961490-m02" exists ...
	I0830 23:14:42.347560 1305486 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" multinode-961490-m02
	I0830 23:14:42.366033 1305486 host.go:66] Checking if "multinode-961490-m02" exists ...
	I0830 23:14:42.366370 1305486 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0830 23:14:42.366426 1305486 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" multinode-961490-m02
	I0830 23:14:42.389415 1305486 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34399 SSHKeyPath:/home/jenkins/minikube-integration/17114-1219981/.minikube/machines/multinode-961490-m02/id_rsa Username:docker}
	I0830 23:14:42.487801 1305486 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0830 23:14:42.501717 1305486 status.go:257] multinode-961490-m02 status: &{Name:multinode-961490-m02 Host:Running Kubelet:Running APIServer:Irrelevant Kubeconfig:Irrelevant Worker:true TimeToStop: DockerEnv: PodManEnv:}
	I0830 23:14:42.501753 1305486 status.go:255] checking status of multinode-961490-m03 ...
	I0830 23:14:42.502058 1305486 cli_runner.go:164] Run: docker container inspect multinode-961490-m03 --format={{.State.Status}}
	I0830 23:14:42.524879 1305486 status.go:330] multinode-961490-m03 host status = "Stopped" (err=<nil>)
	I0830 23:14:42.524901 1305486 status.go:343] host is not running, skipping remaining checks
	I0830 23:14:42.524908 1305486 status.go:257] multinode-961490-m03 status: &{Name:multinode-961490-m03 Host:Stopped Kubelet:Stopped APIServer:Stopped Kubeconfig:Stopped Worker:true TimeToStop: DockerEnv: PodManEnv:}

                                                
                                                
** /stderr **
--- PASS: TestMultiNode/serial/StopNode (2.46s)

                                                
                                    
x
+
TestMultiNode/serial/StartAfterStop (13.2s)

                                                
                                                
=== RUN   TestMultiNode/serial/StartAfterStop
multinode_test.go:244: (dbg) Run:  docker version -f {{.Server.Version}}
multinode_test.go:254: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 node start m03 --alsologtostderr
multinode_test.go:254: (dbg) Done: out/minikube-linux-arm64 -p multinode-961490 node start m03 --alsologtostderr: (12.259848194s)
multinode_test.go:261: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 status
multinode_test.go:275: (dbg) Run:  kubectl get nodes
--- PASS: TestMultiNode/serial/StartAfterStop (13.20s)

                                                
                                    
x
+
TestMultiNode/serial/RestartKeepsNodes (124.83s)

                                                
                                                
=== RUN   TestMultiNode/serial/RestartKeepsNodes
multinode_test.go:283: (dbg) Run:  out/minikube-linux-arm64 node list -p multinode-961490
multinode_test.go:290: (dbg) Run:  out/minikube-linux-arm64 stop -p multinode-961490
multinode_test.go:290: (dbg) Done: out/minikube-linux-arm64 stop -p multinode-961490: (25.148329525s)
multinode_test.go:295: (dbg) Run:  out/minikube-linux-arm64 start -p multinode-961490 --wait=true -v=8 --alsologtostderr
E0830 23:15:24.312330 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:15:51.995359 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
multinode_test.go:295: (dbg) Done: out/minikube-linux-arm64 start -p multinode-961490 --wait=true -v=8 --alsologtostderr: (1m39.495524616s)
multinode_test.go:300: (dbg) Run:  out/minikube-linux-arm64 node list -p multinode-961490
--- PASS: TestMultiNode/serial/RestartKeepsNodes (124.83s)

                                                
                                    
x
+
TestMultiNode/serial/DeleteNode (5.48s)

                                                
                                                
=== RUN   TestMultiNode/serial/DeleteNode
multinode_test.go:394: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 node delete m03
multinode_test.go:394: (dbg) Done: out/minikube-linux-arm64 -p multinode-961490 node delete m03: (4.556862928s)
multinode_test.go:400: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 status --alsologtostderr
multinode_test.go:414: (dbg) Run:  docker volume ls
multinode_test.go:424: (dbg) Run:  kubectl get nodes
multinode_test.go:432: (dbg) Run:  kubectl get nodes -o "go-template='{{range .items}}{{range .status.conditions}}{{if eq .type "Ready"}} {{.status}}{{"\n"}}{{end}}{{end}}{{end}}'"
--- PASS: TestMultiNode/serial/DeleteNode (5.48s)

                                                
                                    
x
+
TestMultiNode/serial/StopMultiNode (24.28s)

                                                
                                                
=== RUN   TestMultiNode/serial/StopMultiNode
multinode_test.go:314: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 stop
E0830 23:17:06.148233 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
multinode_test.go:314: (dbg) Done: out/minikube-linux-arm64 -p multinode-961490 stop: (24.041006004s)
multinode_test.go:320: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 status
multinode_test.go:320: (dbg) Non-zero exit: out/minikube-linux-arm64 -p multinode-961490 status: exit status 7 (119.014795ms)

                                                
                                                
-- stdout --
	multinode-961490
	type: Control Plane
	host: Stopped
	kubelet: Stopped
	apiserver: Stopped
	kubeconfig: Stopped
	
	multinode-961490-m02
	type: Worker
	host: Stopped
	kubelet: Stopped
	

                                                
                                                
-- /stdout --
multinode_test.go:327: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 status --alsologtostderr
multinode_test.go:327: (dbg) Non-zero exit: out/minikube-linux-arm64 -p multinode-961490 status --alsologtostderr: exit status 7 (121.720397ms)

                                                
                                                
-- stdout --
	multinode-961490
	type: Control Plane
	host: Stopped
	kubelet: Stopped
	apiserver: Stopped
	kubeconfig: Stopped
	
	multinode-961490-m02
	type: Worker
	host: Stopped
	kubelet: Stopped
	

                                                
                                                
-- /stdout --
** stderr ** 
	I0830 23:17:30.258813 1314115 out.go:296] Setting OutFile to fd 1 ...
	I0830 23:17:30.259075 1314115 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:17:30.259108 1314115 out.go:309] Setting ErrFile to fd 2...
	I0830 23:17:30.259133 1314115 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:17:30.259467 1314115 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
	I0830 23:17:30.259692 1314115 out.go:303] Setting JSON to false
	I0830 23:17:30.259781 1314115 mustload.go:65] Loading cluster: multinode-961490
	I0830 23:17:30.259873 1314115 notify.go:220] Checking for updates...
	I0830 23:17:30.260286 1314115 config.go:182] Loaded profile config "multinode-961490": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
	I0830 23:17:30.260325 1314115 status.go:255] checking status of multinode-961490 ...
	I0830 23:17:30.262206 1314115 cli_runner.go:164] Run: docker container inspect multinode-961490 --format={{.State.Status}}
	I0830 23:17:30.282863 1314115 status.go:330] multinode-961490 host status = "Stopped" (err=<nil>)
	I0830 23:17:30.282883 1314115 status.go:343] host is not running, skipping remaining checks
	I0830 23:17:30.282890 1314115 status.go:257] multinode-961490 status: &{Name:multinode-961490 Host:Stopped Kubelet:Stopped APIServer:Stopped Kubeconfig:Stopped Worker:false TimeToStop: DockerEnv: PodManEnv:}
	I0830 23:17:30.282920 1314115 status.go:255] checking status of multinode-961490-m02 ...
	I0830 23:17:30.283263 1314115 cli_runner.go:164] Run: docker container inspect multinode-961490-m02 --format={{.State.Status}}
	I0830 23:17:30.314574 1314115 status.go:330] multinode-961490-m02 host status = "Stopped" (err=<nil>)
	I0830 23:17:30.314596 1314115 status.go:343] host is not running, skipping remaining checks
	I0830 23:17:30.314603 1314115 status.go:257] multinode-961490-m02 status: &{Name:multinode-961490-m02 Host:Stopped Kubelet:Stopped APIServer:Stopped Kubeconfig:Stopped Worker:true TimeToStop: DockerEnv: PodManEnv:}

                                                
                                                
** /stderr **
--- PASS: TestMultiNode/serial/StopMultiNode (24.28s)

                                                
                                    
x
+
TestMultiNode/serial/RestartMultiNode (78.55s)

                                                
                                                
=== RUN   TestMultiNode/serial/RestartMultiNode
multinode_test.go:344: (dbg) Run:  docker version -f {{.Server.Version}}
multinode_test.go:354: (dbg) Run:  out/minikube-linux-arm64 start -p multinode-961490 --wait=true -v=8 --alsologtostderr --driver=docker  --container-runtime=containerd
E0830 23:17:44.062064 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:18:29.191401 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
multinode_test.go:354: (dbg) Done: out/minikube-linux-arm64 start -p multinode-961490 --wait=true -v=8 --alsologtostderr --driver=docker  --container-runtime=containerd: (1m17.74109476s)
multinode_test.go:360: (dbg) Run:  out/minikube-linux-arm64 -p multinode-961490 status --alsologtostderr
multinode_test.go:374: (dbg) Run:  kubectl get nodes
multinode_test.go:382: (dbg) Run:  kubectl get nodes -o "go-template='{{range .items}}{{range .status.conditions}}{{if eq .type "Ready"}} {{.status}}{{"\n"}}{{end}}{{end}}{{end}}'"
--- PASS: TestMultiNode/serial/RestartMultiNode (78.55s)

                                                
                                    
x
+
TestMultiNode/serial/ValidateNameConflict (39.73s)

                                                
                                                
=== RUN   TestMultiNode/serial/ValidateNameConflict
multinode_test.go:443: (dbg) Run:  out/minikube-linux-arm64 node list -p multinode-961490
multinode_test.go:452: (dbg) Run:  out/minikube-linux-arm64 start -p multinode-961490-m02 --driver=docker  --container-runtime=containerd
multinode_test.go:452: (dbg) Non-zero exit: out/minikube-linux-arm64 start -p multinode-961490-m02 --driver=docker  --container-runtime=containerd: exit status 14 (97.273891ms)

                                                
                                                
-- stdout --
	* [multinode-961490-m02] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	  - MINIKUBE_LOCATION=17114
	  - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	  - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	  - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	  - MINIKUBE_BIN=out/minikube-linux-arm64
	  - MINIKUBE_FORCE_SYSTEMD=
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	! Profile name 'multinode-961490-m02' is duplicated with machine name 'multinode-961490-m02' in profile 'multinode-961490'
	X Exiting due to MK_USAGE: Profile name should be unique

                                                
                                                
** /stderr **
multinode_test.go:460: (dbg) Run:  out/minikube-linux-arm64 start -p multinode-961490-m03 --driver=docker  --container-runtime=containerd
multinode_test.go:460: (dbg) Done: out/minikube-linux-arm64 start -p multinode-961490-m03 --driver=docker  --container-runtime=containerd: (37.193801717s)
multinode_test.go:467: (dbg) Run:  out/minikube-linux-arm64 node add -p multinode-961490
multinode_test.go:467: (dbg) Non-zero exit: out/minikube-linux-arm64 node add -p multinode-961490: exit status 80 (376.595517ms)

                                                
                                                
-- stdout --
	* Adding node m03 to cluster multinode-961490
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	X Exiting due to GUEST_NODE_ADD: failed to add node: Node multinode-961490-m03 already exists in multinode-961490-m03 profile
	* 
	╭─────────────────────────────────────────────────────────────────────────────────────────────╮
	│                                                                                             │
	│    * If the above advice does not help, please let us know:                                 │
	│      https://github.com/kubernetes/minikube/issues/new/choose                               │
	│                                                                                             │
	│    * Please run `minikube logs --file=logs.txt` and attach logs.txt to the GitHub issue.    │
	│    * Please also attach the following file to the GitHub issue:                             │
	│    * - /tmp/minikube_node_040ea7097fd6ed71e65be9a474587f81f0ccd21d_0.log                    │
	│                                                                                             │
	╰─────────────────────────────────────────────────────────────────────────────────────────────╯

                                                
                                                
** /stderr **
multinode_test.go:472: (dbg) Run:  out/minikube-linux-arm64 delete -p multinode-961490-m03
multinode_test.go:472: (dbg) Done: out/minikube-linux-arm64 delete -p multinode-961490-m03: (1.994896822s)
--- PASS: TestMultiNode/serial/ValidateNameConflict (39.73s)

                                                
                                    
x
+
TestPreload (150s)

                                                
                                                
=== RUN   TestPreload
preload_test.go:44: (dbg) Run:  out/minikube-linux-arm64 start -p test-preload-311148 --memory=2200 --alsologtostderr --wait=true --preload=false --driver=docker  --container-runtime=containerd --kubernetes-version=v1.24.4
E0830 23:20:24.311720 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
preload_test.go:44: (dbg) Done: out/minikube-linux-arm64 start -p test-preload-311148 --memory=2200 --alsologtostderr --wait=true --preload=false --driver=docker  --container-runtime=containerd --kubernetes-version=v1.24.4: (1m28.424547253s)
preload_test.go:52: (dbg) Run:  out/minikube-linux-arm64 -p test-preload-311148 image pull gcr.io/k8s-minikube/busybox
preload_test.go:52: (dbg) Done: out/minikube-linux-arm64 -p test-preload-311148 image pull gcr.io/k8s-minikube/busybox: (1.39932797s)
preload_test.go:58: (dbg) Run:  out/minikube-linux-arm64 stop -p test-preload-311148
preload_test.go:58: (dbg) Done: out/minikube-linux-arm64 stop -p test-preload-311148: (5.819337584s)
preload_test.go:66: (dbg) Run:  out/minikube-linux-arm64 start -p test-preload-311148 --memory=2200 --alsologtostderr -v=1 --wait=true --driver=docker  --container-runtime=containerd
preload_test.go:66: (dbg) Done: out/minikube-linux-arm64 start -p test-preload-311148 --memory=2200 --alsologtostderr -v=1 --wait=true --driver=docker  --container-runtime=containerd: (51.5223633s)
preload_test.go:71: (dbg) Run:  out/minikube-linux-arm64 -p test-preload-311148 image list
helpers_test.go:175: Cleaning up "test-preload-311148" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p test-preload-311148
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p test-preload-311148: (2.425281557s)
--- PASS: TestPreload (150.00s)

                                                
                                    
x
+
TestScheduledStopUnix (107.5s)

                                                
                                                
=== RUN   TestScheduledStopUnix
scheduled_stop_test.go:128: (dbg) Run:  out/minikube-linux-arm64 start -p scheduled-stop-717243 --memory=2048 --driver=docker  --container-runtime=containerd
E0830 23:22:06.147800 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
scheduled_stop_test.go:128: (dbg) Done: out/minikube-linux-arm64 start -p scheduled-stop-717243 --memory=2048 --driver=docker  --container-runtime=containerd: (30.354151752s)
scheduled_stop_test.go:137: (dbg) Run:  out/minikube-linux-arm64 stop -p scheduled-stop-717243 --schedule 5m
scheduled_stop_test.go:191: (dbg) Run:  out/minikube-linux-arm64 status --format={{.TimeToStop}} -p scheduled-stop-717243 -n scheduled-stop-717243
scheduled_stop_test.go:169: signal error was:  <nil>
scheduled_stop_test.go:137: (dbg) Run:  out/minikube-linux-arm64 stop -p scheduled-stop-717243 --schedule 15s
scheduled_stop_test.go:169: signal error was:  os: process already finished
scheduled_stop_test.go:137: (dbg) Run:  out/minikube-linux-arm64 stop -p scheduled-stop-717243 --cancel-scheduled
E0830 23:22:44.061320 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
scheduled_stop_test.go:176: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p scheduled-stop-717243 -n scheduled-stop-717243
scheduled_stop_test.go:205: (dbg) Run:  out/minikube-linux-arm64 status -p scheduled-stop-717243
scheduled_stop_test.go:137: (dbg) Run:  out/minikube-linux-arm64 stop -p scheduled-stop-717243 --schedule 15s
scheduled_stop_test.go:169: signal error was:  os: process already finished
scheduled_stop_test.go:205: (dbg) Run:  out/minikube-linux-arm64 status -p scheduled-stop-717243
scheduled_stop_test.go:205: (dbg) Non-zero exit: out/minikube-linux-arm64 status -p scheduled-stop-717243: exit status 7 (86.103023ms)

                                                
                                                
-- stdout --
	scheduled-stop-717243
	type: Control Plane
	host: Stopped
	kubelet: Stopped
	apiserver: Stopped
	kubeconfig: Stopped
	

                                                
                                                
-- /stdout --
scheduled_stop_test.go:176: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p scheduled-stop-717243 -n scheduled-stop-717243
scheduled_stop_test.go:176: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.Host}} -p scheduled-stop-717243 -n scheduled-stop-717243: exit status 7 (79.140653ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
scheduled_stop_test.go:176: status error: exit status 7 (may be ok)
helpers_test.go:175: Cleaning up "scheduled-stop-717243" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p scheduled-stop-717243
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p scheduled-stop-717243: (5.230514293s)
--- PASS: TestScheduledStopUnix (107.50s)

                                                
                                    
x
+
TestInsufficientStorage (14.56s)

                                                
                                                
=== RUN   TestInsufficientStorage
status_test.go:50: (dbg) Run:  out/minikube-linux-arm64 start -p insufficient-storage-844407 --memory=2048 --output=json --wait=true --driver=docker  --container-runtime=containerd
status_test.go:50: (dbg) Non-zero exit: out/minikube-linux-arm64 start -p insufficient-storage-844407 --memory=2048 --output=json --wait=true --driver=docker  --container-runtime=containerd: exit status 26 (11.963225337s)

                                                
                                                
-- stdout --
	{"specversion":"1.0","id":"e5f075cc-22d1-4814-b041-3b680bcbacd0","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.step","datacontenttype":"application/json","data":{"currentstep":"0","message":"[insufficient-storage-844407] minikube v1.31.2 on Ubuntu 20.04 (arm64)","name":"Initial Minikube Setup","totalsteps":"19"}}
	{"specversion":"1.0","id":"22c8ab16-4322-42b1-b8ba-9b8930106708","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_LOCATION=17114"}}
	{"specversion":"1.0","id":"ff41ee0d-b0c7-41ef-a1b0-e0cc4ecab553","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true"}}
	{"specversion":"1.0","id":"80350183-abe6-4209-85c1-a9faf3befc4b","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig"}}
	{"specversion":"1.0","id":"65486e70-e8e2-4f63-8eab-40c3666fe22c","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube"}}
	{"specversion":"1.0","id":"c7d2cc13-68a7-45b9-8c09-b11a81615c3f","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_BIN=out/minikube-linux-arm64"}}
	{"specversion":"1.0","id":"d622e268-8b24-4d3f-b770-d2f0e489255e","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_FORCE_SYSTEMD="}}
	{"specversion":"1.0","id":"38ab878d-f96e-4fe4-89aa-368f6d995cc0","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_TEST_STORAGE_CAPACITY=100"}}
	{"specversion":"1.0","id":"4baea441-6ad7-4b77-97c2-479555d57d74","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_TEST_AVAILABLE_STORAGE=19"}}
	{"specversion":"1.0","id":"cccdfb78-ecde-46a7-9c5a-0c5db8a86e57","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.step","datacontenttype":"application/json","data":{"currentstep":"1","message":"Using the docker driver based on user configuration","name":"Selecting Driver","totalsteps":"19"}}
	{"specversion":"1.0","id":"db091254-a9f2-4165-9d46-39f04dc7fc62","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"Using Docker driver with root privileges"}}
	{"specversion":"1.0","id":"0e14f546-ee1e-4f0e-b022-709c38ffab81","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.step","datacontenttype":"application/json","data":{"currentstep":"3","message":"Starting control plane node insufficient-storage-844407 in cluster insufficient-storage-844407","name":"Starting Node","totalsteps":"19"}}
	{"specversion":"1.0","id":"bec4f3dd-4ef8-422a-a6df-28b78e01b204","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.step","datacontenttype":"application/json","data":{"currentstep":"5","message":"Pulling base image ...","name":"Pulling Base Image","totalsteps":"19"}}
	{"specversion":"1.0","id":"85b3a817-f6a3-4706-9072-72412a11ac7b","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.step","datacontenttype":"application/json","data":{"currentstep":"8","message":"Creating docker container (CPUs=2, Memory=2048MB) ...","name":"Creating Container","totalsteps":"19"}}
	{"specversion":"1.0","id":"3786ccc8-585c-4ada-9501-d9599bdc32be","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.error","datacontenttype":"application/json","data":{"advice":"Try one or more of the following to free up space on the device:\n\t\n\t\t\t1. Run \"docker system prune\" to remove unused Docker data (optionally with \"-a\")\n\t\t\t2. Increase the storage allocated to Docker for Desktop by clicking on:\n\t\t\t\tDocker icon \u003e Preferences \u003e Resources \u003e Disk Image Size\n\t\t\t3. Run \"minikube ssh -- docker system prune\" if using the Docker container runtime","exitcode":"26","issues":"https://github.com/kubernetes/minikube/issues/9024","message":"Docker is out of disk space! (/var is at 100%% of capacity). You can pass '--force' to skip this check.","name":"RSRC_DOCKER_STORAGE","url":""}}

                                                
                                                
-- /stdout --
status_test.go:76: (dbg) Run:  out/minikube-linux-arm64 status -p insufficient-storage-844407 --output=json --layout=cluster
status_test.go:76: (dbg) Non-zero exit: out/minikube-linux-arm64 status -p insufficient-storage-844407 --output=json --layout=cluster: exit status 7 (336.945005ms)

                                                
                                                
-- stdout --
	{"Name":"insufficient-storage-844407","StatusCode":507,"StatusName":"InsufficientStorage","StatusDetail":"/var is almost out of disk space","Step":"Creating Container","StepDetail":"Creating docker container (CPUs=2, Memory=2048MB) ...","BinaryVersion":"v1.31.2","Components":{"kubeconfig":{"Name":"kubeconfig","StatusCode":500,"StatusName":"Error"}},"Nodes":[{"Name":"insufficient-storage-844407","StatusCode":507,"StatusName":"InsufficientStorage","Components":{"apiserver":{"Name":"apiserver","StatusCode":405,"StatusName":"Stopped"},"kubelet":{"Name":"kubelet","StatusCode":405,"StatusName":"Stopped"}}}]}

                                                
                                                
-- /stdout --
** stderr ** 
	E0830 23:24:02.399412 1331593 status.go:415] kubeconfig endpoint: extract IP: "insufficient-storage-844407" does not appear in /home/jenkins/minikube-integration/17114-1219981/kubeconfig

                                                
                                                
** /stderr **
status_test.go:76: (dbg) Run:  out/minikube-linux-arm64 status -p insufficient-storage-844407 --output=json --layout=cluster
status_test.go:76: (dbg) Non-zero exit: out/minikube-linux-arm64 status -p insufficient-storage-844407 --output=json --layout=cluster: exit status 7 (337.601983ms)

                                                
                                                
-- stdout --
	{"Name":"insufficient-storage-844407","StatusCode":507,"StatusName":"InsufficientStorage","StatusDetail":"/var is almost out of disk space","BinaryVersion":"v1.31.2","Components":{"kubeconfig":{"Name":"kubeconfig","StatusCode":500,"StatusName":"Error"}},"Nodes":[{"Name":"insufficient-storage-844407","StatusCode":507,"StatusName":"InsufficientStorage","Components":{"apiserver":{"Name":"apiserver","StatusCode":405,"StatusName":"Stopped"},"kubelet":{"Name":"kubelet","StatusCode":405,"StatusName":"Stopped"}}}]}

                                                
                                                
-- /stdout --
** stderr ** 
	E0830 23:24:02.738392 1331646 status.go:415] kubeconfig endpoint: extract IP: "insufficient-storage-844407" does not appear in /home/jenkins/minikube-integration/17114-1219981/kubeconfig
	E0830 23:24:02.751190 1331646 status.go:559] unable to read event log: stat: stat /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/insufficient-storage-844407/events.json: no such file or directory

                                                
                                                
** /stderr **
helpers_test.go:175: Cleaning up "insufficient-storage-844407" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p insufficient-storage-844407
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p insufficient-storage-844407: (1.916803591s)
--- PASS: TestInsufficientStorage (14.56s)

                                                
                                    
x
+
TestRunningBinaryUpgrade (119.3s)

                                                
                                                
=== RUN   TestRunningBinaryUpgrade
=== PAUSE TestRunningBinaryUpgrade

                                                
                                                

                                                
                                                
=== CONT  TestRunningBinaryUpgrade
version_upgrade_test.go:132: (dbg) Run:  /tmp/minikube-v1.22.0.2188236460.exe start -p running-upgrade-179344 --memory=2200 --vm-driver=docker  --container-runtime=containerd
version_upgrade_test.go:132: (dbg) Done: /tmp/minikube-v1.22.0.2188236460.exe start -p running-upgrade-179344 --memory=2200 --vm-driver=docker  --container-runtime=containerd: (1m21.532669866s)
version_upgrade_test.go:142: (dbg) Run:  out/minikube-linux-arm64 start -p running-upgrade-179344 --memory=2200 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd
version_upgrade_test.go:142: (dbg) Done: out/minikube-linux-arm64 start -p running-upgrade-179344 --memory=2200 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd: (33.740933613s)
helpers_test.go:175: Cleaning up "running-upgrade-179344" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p running-upgrade-179344
E0830 23:35:09.192256 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p running-upgrade-179344: (2.749670504s)
--- PASS: TestRunningBinaryUpgrade (119.30s)

                                                
                                    
x
+
TestKubernetesUpgrade (422.67s)

                                                
                                                
=== RUN   TestKubernetesUpgrade
=== PAUSE TestKubernetesUpgrade

                                                
                                                

                                                
                                                
=== CONT  TestKubernetesUpgrade
version_upgrade_test.go:234: (dbg) Run:  out/minikube-linux-arm64 start -p kubernetes-upgrade-164934 --memory=2200 --kubernetes-version=v1.16.0 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd
E0830 23:30:24.311925 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
version_upgrade_test.go:234: (dbg) Done: out/minikube-linux-arm64 start -p kubernetes-upgrade-164934 --memory=2200 --kubernetes-version=v1.16.0 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd: (1m11.633333088s)
version_upgrade_test.go:239: (dbg) Run:  out/minikube-linux-arm64 stop -p kubernetes-upgrade-164934
version_upgrade_test.go:239: (dbg) Done: out/minikube-linux-arm64 stop -p kubernetes-upgrade-164934: (1.394704285s)
version_upgrade_test.go:244: (dbg) Run:  out/minikube-linux-arm64 -p kubernetes-upgrade-164934 status --format={{.Host}}
version_upgrade_test.go:244: (dbg) Non-zero exit: out/minikube-linux-arm64 -p kubernetes-upgrade-164934 status --format={{.Host}}: exit status 7 (84.048706ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
version_upgrade_test.go:246: status error: exit status 7 (may be ok)
version_upgrade_test.go:255: (dbg) Run:  out/minikube-linux-arm64 start -p kubernetes-upgrade-164934 --memory=2200 --kubernetes-version=v1.28.1 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd
version_upgrade_test.go:255: (dbg) Done: out/minikube-linux-arm64 start -p kubernetes-upgrade-164934 --memory=2200 --kubernetes-version=v1.28.1 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd: (5m17.968700083s)
version_upgrade_test.go:260: (dbg) Run:  kubectl --context kubernetes-upgrade-164934 version --output=json
version_upgrade_test.go:279: Attempting to downgrade Kubernetes (should fail)
version_upgrade_test.go:281: (dbg) Run:  out/minikube-linux-arm64 start -p kubernetes-upgrade-164934 --memory=2200 --kubernetes-version=v1.16.0 --driver=docker  --container-runtime=containerd
version_upgrade_test.go:281: (dbg) Non-zero exit: out/minikube-linux-arm64 start -p kubernetes-upgrade-164934 --memory=2200 --kubernetes-version=v1.16.0 --driver=docker  --container-runtime=containerd: exit status 106 (105.988866ms)

                                                
                                                
-- stdout --
	* [kubernetes-upgrade-164934] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	  - MINIKUBE_LOCATION=17114
	  - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	  - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	  - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	  - MINIKUBE_BIN=out/minikube-linux-arm64
	  - MINIKUBE_FORCE_SYSTEMD=
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	X Exiting due to K8S_DOWNGRADE_UNSUPPORTED: Unable to safely downgrade existing Kubernetes v1.28.1 cluster to v1.16.0
	* Suggestion: 
	
	    1) Recreate the cluster with Kubernetes 1.16.0, by running:
	    
	    minikube delete -p kubernetes-upgrade-164934
	    minikube start -p kubernetes-upgrade-164934 --kubernetes-version=v1.16.0
	    
	    2) Create a second cluster with Kubernetes 1.16.0, by running:
	    
	    minikube start -p kubernetes-upgrade-1649342 --kubernetes-version=v1.16.0
	    
	    3) Use the existing cluster at version Kubernetes 1.28.1, by running:
	    
	    minikube start -p kubernetes-upgrade-164934 --kubernetes-version=v1.28.1
	    

                                                
                                                
** /stderr **
version_upgrade_test.go:285: Attempting restart after unsuccessful downgrade
version_upgrade_test.go:287: (dbg) Run:  out/minikube-linux-arm64 start -p kubernetes-upgrade-164934 --memory=2200 --kubernetes-version=v1.28.1 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd
version_upgrade_test.go:287: (dbg) Done: out/minikube-linux-arm64 start -p kubernetes-upgrade-164934 --memory=2200 --kubernetes-version=v1.28.1 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd: (28.375862066s)
helpers_test.go:175: Cleaning up "kubernetes-upgrade-164934" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p kubernetes-upgrade-164934
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p kubernetes-upgrade-164934: (3.020075166s)
--- PASS: TestKubernetesUpgrade (422.67s)

                                                
                                    
x
+
TestNoKubernetes/serial/StartNoK8sWithVersion (0.1s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/StartNoK8sWithVersion
no_kubernetes_test.go:83: (dbg) Run:  out/minikube-linux-arm64 start -p NoKubernetes-862547 --no-kubernetes --kubernetes-version=1.20 --driver=docker  --container-runtime=containerd
no_kubernetes_test.go:83: (dbg) Non-zero exit: out/minikube-linux-arm64 start -p NoKubernetes-862547 --no-kubernetes --kubernetes-version=1.20 --driver=docker  --container-runtime=containerd: exit status 14 (100.795624ms)

                                                
                                                
-- stdout --
	* [NoKubernetes-862547] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	  - MINIKUBE_LOCATION=17114
	  - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	  - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	  - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	  - MINIKUBE_BIN=out/minikube-linux-arm64
	  - MINIKUBE_FORCE_SYSTEMD=
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	X Exiting due to MK_USAGE: cannot specify --kubernetes-version with --no-kubernetes,
	to unset a global config run:
	
	$ minikube config unset kubernetes-version

                                                
                                                
** /stderr **
--- PASS: TestNoKubernetes/serial/StartNoK8sWithVersion (0.10s)

                                                
                                    
x
+
TestPause/serial/Start (97.39s)

                                                
                                                
=== RUN   TestPause/serial/Start
pause_test.go:80: (dbg) Run:  out/minikube-linux-arm64 start -p pause-498674 --memory=2048 --install-addons=false --wait=all --driver=docker  --container-runtime=containerd
pause_test.go:80: (dbg) Done: out/minikube-linux-arm64 start -p pause-498674 --memory=2048 --install-addons=false --wait=all --driver=docker  --container-runtime=containerd: (1m37.391356962s)
--- PASS: TestPause/serial/Start (97.39s)

                                                
                                    
x
+
TestNoKubernetes/serial/StartWithK8s (44.3s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/StartWithK8s
no_kubernetes_test.go:95: (dbg) Run:  out/minikube-linux-arm64 start -p NoKubernetes-862547 --driver=docker  --container-runtime=containerd
E0830 23:24:07.104196 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
no_kubernetes_test.go:95: (dbg) Done: out/minikube-linux-arm64 start -p NoKubernetes-862547 --driver=docker  --container-runtime=containerd: (43.88009758s)
no_kubernetes_test.go:200: (dbg) Run:  out/minikube-linux-arm64 -p NoKubernetes-862547 status -o json
--- PASS: TestNoKubernetes/serial/StartWithK8s (44.30s)

                                                
                                    
x
+
TestNoKubernetes/serial/StartWithStopK8s (16.38s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/StartWithStopK8s
no_kubernetes_test.go:112: (dbg) Run:  out/minikube-linux-arm64 start -p NoKubernetes-862547 --no-kubernetes --driver=docker  --container-runtime=containerd
no_kubernetes_test.go:112: (dbg) Done: out/minikube-linux-arm64 start -p NoKubernetes-862547 --no-kubernetes --driver=docker  --container-runtime=containerd: (14.038823205s)
no_kubernetes_test.go:200: (dbg) Run:  out/minikube-linux-arm64 -p NoKubernetes-862547 status -o json
no_kubernetes_test.go:200: (dbg) Non-zero exit: out/minikube-linux-arm64 -p NoKubernetes-862547 status -o json: exit status 2 (373.742316ms)

                                                
                                                
-- stdout --
	{"Name":"NoKubernetes-862547","Host":"Running","Kubelet":"Stopped","APIServer":"Stopped","Kubeconfig":"Configured","Worker":false}

                                                
                                                
-- /stdout --
no_kubernetes_test.go:124: (dbg) Run:  out/minikube-linux-arm64 delete -p NoKubernetes-862547
no_kubernetes_test.go:124: (dbg) Done: out/minikube-linux-arm64 delete -p NoKubernetes-862547: (1.966176128s)
--- PASS: TestNoKubernetes/serial/StartWithStopK8s (16.38s)

                                                
                                    
x
+
TestNoKubernetes/serial/Start (6s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/Start
no_kubernetes_test.go:136: (dbg) Run:  out/minikube-linux-arm64 start -p NoKubernetes-862547 --no-kubernetes --driver=docker  --container-runtime=containerd
no_kubernetes_test.go:136: (dbg) Done: out/minikube-linux-arm64 start -p NoKubernetes-862547 --no-kubernetes --driver=docker  --container-runtime=containerd: (5.997844728s)
--- PASS: TestNoKubernetes/serial/Start (6.00s)

                                                
                                    
x
+
TestNoKubernetes/serial/VerifyK8sNotRunning (0.33s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/VerifyK8sNotRunning
no_kubernetes_test.go:147: (dbg) Run:  out/minikube-linux-arm64 ssh -p NoKubernetes-862547 "sudo systemctl is-active --quiet service kubelet"
no_kubernetes_test.go:147: (dbg) Non-zero exit: out/minikube-linux-arm64 ssh -p NoKubernetes-862547 "sudo systemctl is-active --quiet service kubelet": exit status 1 (327.825969ms)

                                                
                                                
** stderr ** 
	ssh: Process exited with status 3

                                                
                                                
** /stderr **
--- PASS: TestNoKubernetes/serial/VerifyK8sNotRunning (0.33s)

                                                
                                    
x
+
TestNoKubernetes/serial/ProfileList (1.03s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/ProfileList
no_kubernetes_test.go:169: (dbg) Run:  out/minikube-linux-arm64 profile list
no_kubernetes_test.go:179: (dbg) Run:  out/minikube-linux-arm64 profile list --output=json
--- PASS: TestNoKubernetes/serial/ProfileList (1.03s)

                                                
                                    
x
+
TestNoKubernetes/serial/Stop (1.26s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/Stop
no_kubernetes_test.go:158: (dbg) Run:  out/minikube-linux-arm64 stop -p NoKubernetes-862547
no_kubernetes_test.go:158: (dbg) Done: out/minikube-linux-arm64 stop -p NoKubernetes-862547: (1.256340776s)
--- PASS: TestNoKubernetes/serial/Stop (1.26s)

                                                
                                    
x
+
TestNoKubernetes/serial/StartNoArgs (6.57s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/StartNoArgs
no_kubernetes_test.go:191: (dbg) Run:  out/minikube-linux-arm64 start -p NoKubernetes-862547 --driver=docker  --container-runtime=containerd
no_kubernetes_test.go:191: (dbg) Done: out/minikube-linux-arm64 start -p NoKubernetes-862547 --driver=docker  --container-runtime=containerd: (6.568792498s)
--- PASS: TestNoKubernetes/serial/StartNoArgs (6.57s)

                                                
                                    
x
+
TestNoKubernetes/serial/VerifyK8sNotRunningSecond (0.32s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/VerifyK8sNotRunningSecond
no_kubernetes_test.go:147: (dbg) Run:  out/minikube-linux-arm64 ssh -p NoKubernetes-862547 "sudo systemctl is-active --quiet service kubelet"
no_kubernetes_test.go:147: (dbg) Non-zero exit: out/minikube-linux-arm64 ssh -p NoKubernetes-862547 "sudo systemctl is-active --quiet service kubelet": exit status 1 (315.377247ms)

                                                
                                                
** stderr ** 
	ssh: Process exited with status 3

                                                
                                                
** /stderr **
--- PASS: TestNoKubernetes/serial/VerifyK8sNotRunningSecond (0.32s)

                                                
                                    
x
+
TestNetworkPlugins/group/false (4.26s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/false
net_test.go:246: (dbg) Run:  out/minikube-linux-arm64 start -p false-506633 --memory=2048 --alsologtostderr --cni=false --driver=docker  --container-runtime=containerd
net_test.go:246: (dbg) Non-zero exit: out/minikube-linux-arm64 start -p false-506633 --memory=2048 --alsologtostderr --cni=false --driver=docker  --container-runtime=containerd: exit status 14 (228.893585ms)

                                                
                                                
-- stdout --
	* [false-506633] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	  - MINIKUBE_LOCATION=17114
	  - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	  - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	  - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	  - MINIKUBE_BIN=out/minikube-linux-arm64
	  - MINIKUBE_FORCE_SYSTEMD=
	* Using the docker driver based on user configuration
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	I0830 23:25:26.760694 1340525 out.go:296] Setting OutFile to fd 1 ...
	I0830 23:25:26.760900 1340525 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:25:26.760928 1340525 out.go:309] Setting ErrFile to fd 2...
	I0830 23:25:26.760948 1340525 out.go:343] TERM=,COLORTERM=, which probably does not support color
	I0830 23:25:26.761272 1340525 root.go:338] Updating PATH: /home/jenkins/minikube-integration/17114-1219981/.minikube/bin
	I0830 23:25:26.761720 1340525 out.go:303] Setting JSON to false
	I0830 23:25:26.762932 1340525 start.go:128] hostinfo: {"hostname":"ip-172-31-31-251","uptime":29261,"bootTime":1693408666,"procs":456,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1043-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"982e3628-3742-4b3e-bb63-ac1b07660ec7"}
	I0830 23:25:26.763033 1340525 start.go:138] virtualization:  
	I0830 23:25:26.767584 1340525 out.go:177] * [false-506633] minikube v1.31.2 on Ubuntu 20.04 (arm64)
	I0830 23:25:26.769883 1340525 out.go:177]   - MINIKUBE_LOCATION=17114
	I0830 23:25:26.769986 1340525 notify.go:220] Checking for updates...
	I0830 23:25:26.774435 1340525 out.go:177]   - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0830 23:25:26.776711 1340525 out.go:177]   - KUBECONFIG=/home/jenkins/minikube-integration/17114-1219981/kubeconfig
	I0830 23:25:26.778645 1340525 out.go:177]   - MINIKUBE_HOME=/home/jenkins/minikube-integration/17114-1219981/.minikube
	I0830 23:25:26.780840 1340525 out.go:177]   - MINIKUBE_BIN=out/minikube-linux-arm64
	I0830 23:25:26.783225 1340525 out.go:177]   - MINIKUBE_FORCE_SYSTEMD=
	I0830 23:25:26.785494 1340525 config.go:182] Loaded profile config "pause-498674": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.1
	I0830 23:25:26.785592 1340525 driver.go:373] Setting default libvirt URI to qemu:///system
	I0830 23:25:26.810848 1340525 docker.go:121] docker version: linux-24.0.5:Docker Engine - Community
	I0830 23:25:26.810949 1340525 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0830 23:25:26.913936 1340525 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:1 ContainersRunning:1 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:33 OomKillDisable:true NGoroutines:45 SystemTime:2023-08-30 23:25:26.902075392 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1043-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Archi
tecture:aarch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8215113728 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:24.0.5 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8165feabfdfe38c65b599c4993d227328c231fca Expected:8165feabfdfe38c65b599c4993d227328c231fca} RuncCommit:{ID:v1.1.8-0-g82f18fe Expected:v1.1.8-0-g82f18fe} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> S
erverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.11.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.20.2]] Warnings:<nil>}}
	I0830 23:25:26.914051 1340525 docker.go:294] overlay module found
	I0830 23:25:26.916025 1340525 out.go:177] * Using the docker driver based on user configuration
	I0830 23:25:26.917701 1340525 start.go:298] selected driver: docker
	I0830 23:25:26.917723 1340525 start.go:902] validating driver "docker" against <nil>
	I0830 23:25:26.917736 1340525 start.go:913] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
	I0830 23:25:26.920165 1340525 out.go:177] 
	W0830 23:25:26.921825 1340525 out.go:239] X Exiting due to MK_USAGE: The "containerd" container runtime requires CNI
	X Exiting due to MK_USAGE: The "containerd" container runtime requires CNI
	I0830 23:25:26.923740 1340525 out.go:177] 

                                                
                                                
** /stderr **
net_test.go:88: 
----------------------- debugLogs start: false-506633 [pass: true] --------------------------------
>>> netcat: nslookup kubernetes.default:
Error in configuration: context was not found for specified context: false-506633

                                                
                                                

                                                
                                                
>>> netcat: nslookup debug kubernetes.default a-records:
Error in configuration: context was not found for specified context: false-506633

                                                
                                                

                                                
                                                
>>> netcat: dig search kubernetes.default:
Error in configuration: context was not found for specified context: false-506633

                                                
                                                

                                                
                                                
>>> netcat: dig @10.96.0.10 kubernetes.default.svc.cluster.local udp/53:
Error in configuration: context was not found for specified context: false-506633

                                                
                                                

                                                
                                                
>>> netcat: dig @10.96.0.10 kubernetes.default.svc.cluster.local tcp/53:
Error in configuration: context was not found for specified context: false-506633

                                                
                                                

                                                
                                                
>>> netcat: nc 10.96.0.10 udp/53:
Error in configuration: context was not found for specified context: false-506633

                                                
                                                

                                                
                                                
>>> netcat: nc 10.96.0.10 tcp/53:
Error in configuration: context was not found for specified context: false-506633

                                                
                                                

                                                
                                                
>>> netcat: /etc/nsswitch.conf:
Error in configuration: context was not found for specified context: false-506633

                                                
                                                

                                                
                                                
>>> netcat: /etc/hosts:
Error in configuration: context was not found for specified context: false-506633

                                                
                                                

                                                
                                                
>>> netcat: /etc/resolv.conf:
Error in configuration: context was not found for specified context: false-506633

                                                
                                                

                                                
                                                
>>> host: /etc/nsswitch.conf:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/hosts:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/resolv.conf:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> k8s: nodes, services, endpoints, daemon sets, deployments and pods, :
Error in configuration: context was not found for specified context: false-506633

                                                
                                                

                                                
                                                
>>> host: crictl pods:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: crictl containers:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> k8s: describe netcat deployment:
error: context "false-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe netcat pod(s):
error: context "false-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: netcat logs:
error: context "false-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe coredns deployment:
error: context "false-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe coredns pods:
error: context "false-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: coredns logs:
error: context "false-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe api server pod(s):
error: context "false-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: api server logs:
error: context "false-506633" does not exist

                                                
                                                

                                                
                                                
>>> host: /etc/cni:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: ip a s:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: ip r s:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: iptables-save:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: iptables table nat:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> k8s: describe kube-proxy daemon set:
error: context "false-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe kube-proxy pod(s):
error: context "false-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: kube-proxy logs:
error: context "false-506633" does not exist

                                                
                                                

                                                
                                                
>>> host: kubelet daemon status:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: kubelet daemon config:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> k8s: kubelet logs:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/kubernetes/kubelet.conf:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: /var/lib/kubelet/config.yaml:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> k8s: kubectl config:
apiVersion: v1
clusters:
- cluster:
certificate-authority: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt
extensions:
- extension:
last-update: Wed, 30 Aug 2023 23:24:58 UTC
provider: minikube.sigs.k8s.io
version: v1.31.2
name: cluster_info
server: https://192.168.67.2:8443
name: pause-498674
contexts:
- context:
cluster: pause-498674
extensions:
- extension:
last-update: Wed, 30 Aug 2023 23:24:58 UTC
provider: minikube.sigs.k8s.io
version: v1.31.2
name: context_info
namespace: default
user: pause-498674
name: pause-498674
current-context: pause-498674
kind: Config
preferences: {}
users:
- name: pause-498674
user:
client-certificate: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/pause-498674/client.crt
client-key: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/pause-498674/client.key

                                                
                                                

                                                
                                                
>>> k8s: cms:
Error in configuration: context was not found for specified context: false-506633

                                                
                                                

                                                
                                                
>>> host: docker daemon status:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: docker daemon config:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/docker/daemon.json:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: docker system info:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: cri-docker daemon status:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: cri-docker daemon config:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/systemd/system/cri-docker.service.d/10-cni.conf:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: /usr/lib/systemd/system/cri-docker.service:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: cri-dockerd version:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: containerd daemon status:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: containerd daemon config:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: /lib/systemd/system/containerd.service:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/containerd/config.toml:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: containerd config dump:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: crio daemon status:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: crio daemon config:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/crio:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                

                                                
                                                
>>> host: crio config:
* Profile "false-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p false-506633"

                                                
                                                
----------------------- debugLogs end: false-506633 [took: 3.847330325s] --------------------------------
helpers_test.go:175: Cleaning up "false-506633" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p false-506633
--- PASS: TestNetworkPlugins/group/false (4.26s)

                                                
                                    
x
+
TestPause/serial/SecondStartNoReconfiguration (6.53s)

                                                
                                                
=== RUN   TestPause/serial/SecondStartNoReconfiguration
pause_test.go:92: (dbg) Run:  out/minikube-linux-arm64 start -p pause-498674 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd
pause_test.go:92: (dbg) Done: out/minikube-linux-arm64 start -p pause-498674 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd: (6.521582558s)
--- PASS: TestPause/serial/SecondStartNoReconfiguration (6.53s)

                                                
                                    
x
+
TestPause/serial/Pause (0.93s)

                                                
                                                
=== RUN   TestPause/serial/Pause
pause_test.go:110: (dbg) Run:  out/minikube-linux-arm64 pause -p pause-498674 --alsologtostderr -v=5
--- PASS: TestPause/serial/Pause (0.93s)

                                                
                                    
x
+
TestPause/serial/VerifyStatus (0.46s)

                                                
                                                
=== RUN   TestPause/serial/VerifyStatus
status_test.go:76: (dbg) Run:  out/minikube-linux-arm64 status -p pause-498674 --output=json --layout=cluster
status_test.go:76: (dbg) Non-zero exit: out/minikube-linux-arm64 status -p pause-498674 --output=json --layout=cluster: exit status 2 (457.580884ms)

                                                
                                                
-- stdout --
	{"Name":"pause-498674","StatusCode":418,"StatusName":"Paused","Step":"Done","StepDetail":"* Paused 7 containers in: kube-system, kubernetes-dashboard, storage-gluster, istio-operator","BinaryVersion":"v1.31.2","Components":{"kubeconfig":{"Name":"kubeconfig","StatusCode":200,"StatusName":"OK"}},"Nodes":[{"Name":"pause-498674","StatusCode":200,"StatusName":"OK","Components":{"apiserver":{"Name":"apiserver","StatusCode":418,"StatusName":"Paused"},"kubelet":{"Name":"kubelet","StatusCode":405,"StatusName":"Stopped"}}}]}

                                                
                                                
-- /stdout --
--- PASS: TestPause/serial/VerifyStatus (0.46s)

                                                
                                    
x
+
TestPause/serial/Unpause (0.91s)

                                                
                                                
=== RUN   TestPause/serial/Unpause
pause_test.go:121: (dbg) Run:  out/minikube-linux-arm64 unpause -p pause-498674 --alsologtostderr -v=5
--- PASS: TestPause/serial/Unpause (0.91s)

                                                
                                    
x
+
TestPause/serial/PauseAgain (1.27s)

                                                
                                                
=== RUN   TestPause/serial/PauseAgain
pause_test.go:110: (dbg) Run:  out/minikube-linux-arm64 pause -p pause-498674 --alsologtostderr -v=5
pause_test.go:110: (dbg) Done: out/minikube-linux-arm64 pause -p pause-498674 --alsologtostderr -v=5: (1.269578906s)
--- PASS: TestPause/serial/PauseAgain (1.27s)

                                                
                                    
x
+
TestPause/serial/DeletePaused (3.01s)

                                                
                                                
=== RUN   TestPause/serial/DeletePaused
pause_test.go:132: (dbg) Run:  out/minikube-linux-arm64 delete -p pause-498674 --alsologtostderr -v=5
pause_test.go:132: (dbg) Done: out/minikube-linux-arm64 delete -p pause-498674 --alsologtostderr -v=5: (3.014608887s)
--- PASS: TestPause/serial/DeletePaused (3.01s)

                                                
                                    
x
+
TestPause/serial/VerifyDeletedResources (0.47s)

                                                
                                                
=== RUN   TestPause/serial/VerifyDeletedResources
pause_test.go:142: (dbg) Run:  out/minikube-linux-arm64 profile list --output json
pause_test.go:168: (dbg) Run:  docker ps -a
pause_test.go:173: (dbg) Run:  docker volume inspect pause-498674
pause_test.go:173: (dbg) Non-zero exit: docker volume inspect pause-498674: exit status 1 (19.073345ms)

                                                
                                                
-- stdout --
	[]

                                                
                                                
-- /stdout --
** stderr ** 
	Error response from daemon: get pause-498674: no such volume

                                                
                                                
** /stderr **
pause_test.go:178: (dbg) Run:  docker network ls
--- PASS: TestPause/serial/VerifyDeletedResources (0.47s)

                                                
                                    
x
+
TestStoppedBinaryUpgrade/Setup (1.21s)

                                                
                                                
=== RUN   TestStoppedBinaryUpgrade/Setup
--- PASS: TestStoppedBinaryUpgrade/Setup (1.21s)

                                                
                                    
x
+
TestStoppedBinaryUpgrade/Upgrade (178.44s)

                                                
                                                
=== RUN   TestStoppedBinaryUpgrade/Upgrade
version_upgrade_test.go:195: (dbg) Run:  /tmp/minikube-v1.22.0.68057774.exe start -p stopped-upgrade-636929 --memory=2200 --vm-driver=docker  --container-runtime=containerd
E0830 23:27:44.061299 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
version_upgrade_test.go:195: (dbg) Done: /tmp/minikube-v1.22.0.68057774.exe start -p stopped-upgrade-636929 --memory=2200 --vm-driver=docker  --container-runtime=containerd: (1m34.404058733s)
version_upgrade_test.go:204: (dbg) Run:  /tmp/minikube-v1.22.0.68057774.exe -p stopped-upgrade-636929 stop
version_upgrade_test.go:204: (dbg) Done: /tmp/minikube-v1.22.0.68057774.exe -p stopped-upgrade-636929 stop: (12.358746825s)
version_upgrade_test.go:210: (dbg) Run:  out/minikube-linux-arm64 start -p stopped-upgrade-636929 --memory=2200 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd
version_upgrade_test.go:210: (dbg) Done: out/minikube-linux-arm64 start -p stopped-upgrade-636929 --memory=2200 --alsologtostderr -v=1 --driver=docker  --container-runtime=containerd: (1m11.676120211s)
--- PASS: TestStoppedBinaryUpgrade/Upgrade (178.44s)

                                                
                                    
x
+
TestStoppedBinaryUpgrade/MinikubeLogs (1.73s)

                                                
                                                
=== RUN   TestStoppedBinaryUpgrade/MinikubeLogs
version_upgrade_test.go:218: (dbg) Run:  out/minikube-linux-arm64 logs -p stopped-upgrade-636929
version_upgrade_test.go:218: (dbg) Done: out/minikube-linux-arm64 logs -p stopped-upgrade-636929: (1.732078212s)
--- PASS: TestStoppedBinaryUpgrade/MinikubeLogs (1.73s)

                                                
                                    
x
+
TestNetworkPlugins/group/auto/Start (95.42s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/auto/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-arm64 start -p auto-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --driver=docker  --container-runtime=containerd
E0830 23:35:24.311689 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
net_test.go:112: (dbg) Done: out/minikube-linux-arm64 start -p auto-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --driver=docker  --container-runtime=containerd: (1m35.416652634s)
--- PASS: TestNetworkPlugins/group/auto/Start (95.42s)

                                                
                                    
x
+
TestNetworkPlugins/group/auto/KubeletFlags (0.47s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/auto/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-arm64 ssh -p auto-506633 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/auto/KubeletFlags (0.47s)

                                                
                                    
x
+
TestNetworkPlugins/group/auto/NetCatPod (9.61s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/auto/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context auto-506633 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/auto/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-56589dfd74-wcjxx" [2962f798-30e5-436f-8b21-888b6bc961f8] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
helpers_test.go:344: "netcat-56589dfd74-wcjxx" [2962f798-30e5-436f-8b21-888b6bc961f8] Running
net_test.go:163: (dbg) TestNetworkPlugins/group/auto/NetCatPod: app=netcat healthy within 9.015215868s
--- PASS: TestNetworkPlugins/group/auto/NetCatPod (9.61s)

                                                
                                    
x
+
TestNetworkPlugins/group/auto/DNS (0.29s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/auto/DNS
net_test.go:175: (dbg) Run:  kubectl --context auto-506633 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/auto/DNS (0.29s)

                                                
                                    
x
+
TestNetworkPlugins/group/auto/Localhost (0.25s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/auto/Localhost
net_test.go:194: (dbg) Run:  kubectl --context auto-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/auto/Localhost (0.25s)

                                                
                                    
x
+
TestNetworkPlugins/group/auto/HairPin (0.2s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/auto/HairPin
net_test.go:264: (dbg) Run:  kubectl --context auto-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/auto/HairPin (0.20s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/Start (91.98s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-arm64 start -p kindnet-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=kindnet --driver=docker  --container-runtime=containerd
net_test.go:112: (dbg) Done: out/minikube-linux-arm64 start -p kindnet-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=kindnet --driver=docker  --container-runtime=containerd: (1m31.984782966s)
--- PASS: TestNetworkPlugins/group/kindnet/Start (91.98s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/Start (72.97s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-arm64 start -p calico-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=calico --driver=docker  --container-runtime=containerd
E0830 23:37:44.061697 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
net_test.go:112: (dbg) Done: out/minikube-linux-arm64 start -p calico-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=calico --driver=docker  --container-runtime=containerd: (1m12.970384909s)
--- PASS: TestNetworkPlugins/group/calico/Start (72.97s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/ControllerPod (5.04s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/ControllerPod
net_test.go:120: (dbg) TestNetworkPlugins/group/calico/ControllerPod: waiting 10m0s for pods matching "k8s-app=calico-node" in namespace "kube-system" ...
helpers_test.go:344: "calico-node-mgff2" [aa0f1e22-2231-445e-80c7-3e3ec7babdff] Running
net_test.go:120: (dbg) TestNetworkPlugins/group/calico/ControllerPod: k8s-app=calico-node healthy within 5.03053622s
--- PASS: TestNetworkPlugins/group/calico/ControllerPod (5.04s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/KubeletFlags (0.31s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-arm64 ssh -p calico-506633 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/calico/KubeletFlags (0.31s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/NetCatPod (9.43s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context calico-506633 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/calico/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-56589dfd74-w66rc" [0e048bbb-b256-489a-bde6-73dc58e91343] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
helpers_test.go:344: "netcat-56589dfd74-w66rc" [0e048bbb-b256-489a-bde6-73dc58e91343] Running
net_test.go:163: (dbg) TestNetworkPlugins/group/calico/NetCatPod: app=netcat healthy within 9.022383855s
--- PASS: TestNetworkPlugins/group/calico/NetCatPod (9.43s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/DNS (0.24s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/DNS
net_test.go:175: (dbg) Run:  kubectl --context calico-506633 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/calico/DNS (0.24s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/Localhost (0.21s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/Localhost
net_test.go:194: (dbg) Run:  kubectl --context calico-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/calico/Localhost (0.21s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/HairPin (0.21s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/HairPin
net_test.go:264: (dbg) Run:  kubectl --context calico-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/calico/HairPin (0.21s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/ControllerPod (5.04s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/ControllerPod
net_test.go:120: (dbg) TestNetworkPlugins/group/kindnet/ControllerPod: waiting 10m0s for pods matching "app=kindnet" in namespace "kube-system" ...
helpers_test.go:344: "kindnet-5td9n" [9d17cca5-2188-4fe4-a13c-168a15728505] Running
net_test.go:120: (dbg) TestNetworkPlugins/group/kindnet/ControllerPod: app=kindnet healthy within 5.038750301s
--- PASS: TestNetworkPlugins/group/kindnet/ControllerPod (5.04s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/KubeletFlags (0.47s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-arm64 ssh -p kindnet-506633 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/kindnet/KubeletFlags (0.47s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/NetCatPod (9.65s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context kindnet-506633 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/kindnet/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-56589dfd74-w4sml" [7f70dca5-2194-40b5-991b-025c1432b3f4] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
helpers_test.go:344: "netcat-56589dfd74-w4sml" [7f70dca5-2194-40b5-991b-025c1432b3f4] Running
net_test.go:163: (dbg) TestNetworkPlugins/group/kindnet/NetCatPod: app=netcat healthy within 9.025912923s
--- PASS: TestNetworkPlugins/group/kindnet/NetCatPod (9.65s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/DNS (0.28s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/DNS
net_test.go:175: (dbg) Run:  kubectl --context kindnet-506633 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/kindnet/DNS (0.28s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/Localhost (0.25s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/Localhost
net_test.go:194: (dbg) Run:  kubectl --context kindnet-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/kindnet/Localhost (0.25s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/HairPin (0.26s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/HairPin
net_test.go:264: (dbg) Run:  kubectl --context kindnet-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/kindnet/HairPin (0.26s)

                                                
                                    
x
+
TestNetworkPlugins/group/custom-flannel/Start (63.69s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/custom-flannel/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-arm64 start -p custom-flannel-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=testdata/kube-flannel.yaml --driver=docker  --container-runtime=containerd
net_test.go:112: (dbg) Done: out/minikube-linux-arm64 start -p custom-flannel-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=testdata/kube-flannel.yaml --driver=docker  --container-runtime=containerd: (1m3.693741521s)
--- PASS: TestNetworkPlugins/group/custom-flannel/Start (63.69s)

                                                
                                    
x
+
TestNetworkPlugins/group/enable-default-cni/Start (55.62s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/enable-default-cni/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-arm64 start -p enable-default-cni-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --enable-default-cni=true --driver=docker  --container-runtime=containerd
net_test.go:112: (dbg) Done: out/minikube-linux-arm64 start -p enable-default-cni-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --enable-default-cni=true --driver=docker  --container-runtime=containerd: (55.617613617s)
--- PASS: TestNetworkPlugins/group/enable-default-cni/Start (55.62s)

                                                
                                    
x
+
TestNetworkPlugins/group/custom-flannel/KubeletFlags (0.33s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/custom-flannel/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-arm64 ssh -p custom-flannel-506633 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/custom-flannel/KubeletFlags (0.33s)

                                                
                                    
x
+
TestNetworkPlugins/group/custom-flannel/NetCatPod (10.43s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/custom-flannel/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context custom-flannel-506633 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/custom-flannel/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-56589dfd74-mk2ht" [9929dfe0-b098-4e0d-9056-ec3cbf573cc7] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
E0830 23:40:24.311937 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
helpers_test.go:344: "netcat-56589dfd74-mk2ht" [9929dfe0-b098-4e0d-9056-ec3cbf573cc7] Running
net_test.go:163: (dbg) TestNetworkPlugins/group/custom-flannel/NetCatPod: app=netcat healthy within 10.027199437s
--- PASS: TestNetworkPlugins/group/custom-flannel/NetCatPod (10.43s)

                                                
                                    
x
+
TestNetworkPlugins/group/enable-default-cni/KubeletFlags (0.34s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/enable-default-cni/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-arm64 ssh -p enable-default-cni-506633 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/enable-default-cni/KubeletFlags (0.34s)

                                                
                                    
x
+
TestNetworkPlugins/group/custom-flannel/DNS (0.3s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/custom-flannel/DNS
net_test.go:175: (dbg) Run:  kubectl --context custom-flannel-506633 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/custom-flannel/DNS (0.30s)

                                                
                                    
x
+
TestNetworkPlugins/group/enable-default-cni/NetCatPod (8.54s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/enable-default-cni/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context enable-default-cni-506633 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/enable-default-cni/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-56589dfd74-btstm" [6498ec7f-95d1-4891-bb66-5eba4ba5fbb7] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
helpers_test.go:344: "netcat-56589dfd74-btstm" [6498ec7f-95d1-4891-bb66-5eba4ba5fbb7] Running
net_test.go:163: (dbg) TestNetworkPlugins/group/enable-default-cni/NetCatPod: app=netcat healthy within 8.020833146s
--- PASS: TestNetworkPlugins/group/enable-default-cni/NetCatPod (8.54s)

                                                
                                    
x
+
TestNetworkPlugins/group/custom-flannel/Localhost (0.27s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/custom-flannel/Localhost
net_test.go:194: (dbg) Run:  kubectl --context custom-flannel-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/custom-flannel/Localhost (0.27s)

                                                
                                    
x
+
TestNetworkPlugins/group/custom-flannel/HairPin (0.24s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/custom-flannel/HairPin
net_test.go:264: (dbg) Run:  kubectl --context custom-flannel-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/custom-flannel/HairPin (0.24s)

                                                
                                    
x
+
TestNetworkPlugins/group/enable-default-cni/DNS (0.29s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/enable-default-cni/DNS
net_test.go:175: (dbg) Run:  kubectl --context enable-default-cni-506633 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/enable-default-cni/DNS (0.29s)

                                                
                                    
x
+
TestNetworkPlugins/group/enable-default-cni/Localhost (0.27s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/enable-default-cni/Localhost
net_test.go:194: (dbg) Run:  kubectl --context enable-default-cni-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/enable-default-cni/Localhost (0.27s)

                                                
                                    
x
+
TestNetworkPlugins/group/enable-default-cni/HairPin (0.26s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/enable-default-cni/HairPin
net_test.go:264: (dbg) Run:  kubectl --context enable-default-cni-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/enable-default-cni/HairPin (0.26s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/Start (65.98s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-arm64 start -p flannel-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=flannel --driver=docker  --container-runtime=containerd
net_test.go:112: (dbg) Done: out/minikube-linux-arm64 start -p flannel-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=flannel --driver=docker  --container-runtime=containerd: (1m5.98247979s)
--- PASS: TestNetworkPlugins/group/flannel/Start (65.98s)

                                                
                                    
x
+
TestNetworkPlugins/group/bridge/Start (52.41s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/bridge/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-arm64 start -p bridge-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=bridge --driver=docker  --container-runtime=containerd
E0830 23:41:46.824642 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:41:46.830672 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:41:46.840881 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:41:46.861390 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:41:46.901679 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:41:46.981983 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:41:47.142983 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:41:47.463400 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:41:48.104567 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:41:49.385251 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:41:51.945734 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:41:57.065948 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
net_test.go:112: (dbg) Done: out/minikube-linux-arm64 start -p bridge-506633 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=bridge --driver=docker  --container-runtime=containerd: (52.409953409s)
--- PASS: TestNetworkPlugins/group/bridge/Start (52.41s)

                                                
                                    
x
+
TestNetworkPlugins/group/bridge/KubeletFlags (0.42s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/bridge/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-arm64 ssh -p bridge-506633 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/bridge/KubeletFlags (0.42s)

                                                
                                    
x
+
TestNetworkPlugins/group/bridge/NetCatPod (8.4s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/bridge/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context bridge-506633 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/bridge/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-56589dfd74-lg7bn" [9a6fd6f6-85fd-4667-a4f8-5c9fb9795cb7] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
E0830 23:42:06.148558 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
helpers_test.go:344: "netcat-56589dfd74-lg7bn" [9a6fd6f6-85fd-4667-a4f8-5c9fb9795cb7] Running
net_test.go:163: (dbg) TestNetworkPlugins/group/bridge/NetCatPod: app=netcat healthy within 8.011150566s
--- PASS: TestNetworkPlugins/group/bridge/NetCatPod (8.40s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/ControllerPod (5.03s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/ControllerPod
net_test.go:120: (dbg) TestNetworkPlugins/group/flannel/ControllerPod: waiting 10m0s for pods matching "app=flannel" in namespace "kube-flannel" ...
helpers_test.go:344: "kube-flannel-ds-mf5hr" [149ed2bf-fec5-4f1d-babc-1c64fc4b5e7d] Running
E0830 23:42:07.306795 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
net_test.go:120: (dbg) TestNetworkPlugins/group/flannel/ControllerPod: app=flannel healthy within 5.03059297s
--- PASS: TestNetworkPlugins/group/flannel/ControllerPod (5.03s)

                                                
                                    
x
+
TestNetworkPlugins/group/bridge/DNS (0.21s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/bridge/DNS
net_test.go:175: (dbg) Run:  kubectl --context bridge-506633 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/bridge/DNS (0.21s)

                                                
                                    
x
+
TestNetworkPlugins/group/bridge/Localhost (0.18s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/bridge/Localhost
net_test.go:194: (dbg) Run:  kubectl --context bridge-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/bridge/Localhost (0.18s)

                                                
                                    
x
+
TestNetworkPlugins/group/bridge/HairPin (0.18s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/bridge/HairPin
net_test.go:264: (dbg) Run:  kubectl --context bridge-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/bridge/HairPin (0.18s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/KubeletFlags (0.45s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-arm64 ssh -p flannel-506633 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/flannel/KubeletFlags (0.45s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/NetCatPod (9.35s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context flannel-506633 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/flannel/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-56589dfd74-q26pd" [f41b5e3e-2f61-4ac0-865f-5119e1d7d4a2] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
helpers_test.go:344: "netcat-56589dfd74-q26pd" [f41b5e3e-2f61-4ac0-865f-5119e1d7d4a2] Running
net_test.go:163: (dbg) TestNetworkPlugins/group/flannel/NetCatPod: app=netcat healthy within 9.012542039s
--- PASS: TestNetworkPlugins/group/flannel/NetCatPod (9.35s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/DNS (0.25s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/DNS
net_test.go:175: (dbg) Run:  kubectl --context flannel-506633 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/flannel/DNS (0.25s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/Localhost (0.27s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/Localhost
net_test.go:194: (dbg) Run:  kubectl --context flannel-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/flannel/Localhost (0.27s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/HairPin (0.27s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/HairPin
net_test.go:264: (dbg) Run:  kubectl --context flannel-506633 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/flannel/HairPin (0.27s)
E0831 00:00:23.425926 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0831 00:00:24.312488 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0831 00:00:31.318111 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0831 00:00:34.090586 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0831 00:01:12.278806 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0831 00:01:46.471274 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0831 00:01:46.824836 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0831 00:01:57.135252 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0831 00:02:03.267810 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0831 00:02:06.148552 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0831 00:02:06.971705 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0831 00:02:34.199531 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0831 00:02:44.061470 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/FirstStart (134.08s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/FirstStart
start_stop_delete_test.go:186: (dbg) Run:  out/minikube-linux-arm64 start -p old-k8s-version-537179 --memory=2200 --alsologtostderr --wait=true --kvm-network=default --kvm-qemu-uri=qemu:///system --disable-driver-mounts --keep-context=false --driver=docker  --container-runtime=containerd --kubernetes-version=v1.16.0
E0830 23:42:44.061542 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
start_stop_delete_test.go:186: (dbg) Done: out/minikube-linux-arm64 start -p old-k8s-version-537179 --memory=2200 --alsologtostderr --wait=true --kvm-network=default --kvm-qemu-uri=qemu:///system --disable-driver-mounts --keep-context=false --driver=docker  --container-runtime=containerd --kubernetes-version=v1.16.0: (2m14.082675621s)
--- PASS: TestStartStop/group/old-k8s-version/serial/FirstStart (134.08s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/FirstStart (90.41s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/FirstStart
start_stop_delete_test.go:186: (dbg) Run:  out/minikube-linux-arm64 start -p no-preload-386690 --memory=2200 --alsologtostderr --wait=true --preload=false --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1
E0830 23:43:08.747871 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:43:27.356140 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:43:38.132694 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:43:38.138008 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:43:38.148277 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:43:38.168514 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:43:38.208779 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:43:38.289017 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:43:38.449590 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:43:38.770371 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:43:39.410779 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:43:40.691605 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:43:43.252138 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:43:48.373280 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:43:54.103019 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:43:54.108329 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:43:54.118654 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:43:54.138905 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:43:54.179440 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:43:54.260490 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:43:54.421247 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:43:54.741406 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:43:55.382273 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:43:56.663009 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:43:58.614362 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:43:59.224159 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:44:04.344572 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:44:14.584788 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:44:19.094604 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
start_stop_delete_test.go:186: (dbg) Done: out/minikube-linux-arm64 start -p no-preload-386690 --memory=2200 --alsologtostderr --wait=true --preload=false --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1: (1m30.40751892s)
--- PASS: TestStartStop/group/no-preload/serial/FirstStart (90.41s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/DeployApp (8.47s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/DeployApp
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context no-preload-386690 create -f testdata/busybox.yaml
start_stop_delete_test.go:196: (dbg) TestStartStop/group/no-preload/serial/DeployApp: waiting 8m0s for pods matching "integration-test=busybox" in namespace "default" ...
helpers_test.go:344: "busybox" [e06a0aa9-763e-4b17-9fa0-6c276049ca27] Pending / Ready:ContainersNotReady (containers with unready status: [busybox]) / ContainersReady:ContainersNotReady (containers with unready status: [busybox])
helpers_test.go:344: "busybox" [e06a0aa9-763e-4b17-9fa0-6c276049ca27] Running
start_stop_delete_test.go:196: (dbg) TestStartStop/group/no-preload/serial/DeployApp: integration-test=busybox healthy within 8.029036782s
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context no-preload-386690 exec busybox -- /bin/sh -c "ulimit -n"
--- PASS: TestStartStop/group/no-preload/serial/DeployApp (8.47s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/EnableAddonWhileActive (1.24s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/EnableAddonWhileActive
start_stop_delete_test.go:205: (dbg) Run:  out/minikube-linux-arm64 addons enable metrics-server -p no-preload-386690 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain
E0830 23:44:30.668740 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
start_stop_delete_test.go:205: (dbg) Done: out/minikube-linux-arm64 addons enable metrics-server -p no-preload-386690 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain: (1.11800386s)
start_stop_delete_test.go:215: (dbg) Run:  kubectl --context no-preload-386690 describe deploy/metrics-server -n kube-system
--- PASS: TestStartStop/group/no-preload/serial/EnableAddonWhileActive (1.24s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/Stop (12.21s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/Stop
start_stop_delete_test.go:228: (dbg) Run:  out/minikube-linux-arm64 stop -p no-preload-386690 --alsologtostderr -v=3
E0830 23:44:35.065075 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
start_stop_delete_test.go:228: (dbg) Done: out/minikube-linux-arm64 stop -p no-preload-386690 --alsologtostderr -v=3: (12.214473718s)
--- PASS: TestStartStop/group/no-preload/serial/Stop (12.21s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/EnableAddonAfterStop (0.22s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/EnableAddonAfterStop
start_stop_delete_test.go:239: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p no-preload-386690 -n no-preload-386690
start_stop_delete_test.go:239: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.Host}} -p no-preload-386690 -n no-preload-386690: exit status 7 (82.078292ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:239: status error: exit status 7 (may be ok)
start_stop_delete_test.go:246: (dbg) Run:  out/minikube-linux-arm64 addons enable dashboard -p no-preload-386690 --images=MetricsScraper=registry.k8s.io/echoserver:1.4
--- PASS: TestStartStop/group/no-preload/serial/EnableAddonAfterStop (0.22s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/SecondStart (347.45s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/SecondStart
start_stop_delete_test.go:256: (dbg) Run:  out/minikube-linux-arm64 start -p no-preload-386690 --memory=2200 --alsologtostderr --wait=true --preload=false --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1
start_stop_delete_test.go:256: (dbg) Done: out/minikube-linux-arm64 start -p no-preload-386690 --memory=2200 --alsologtostderr --wait=true --preload=false --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1: (5m46.908420999s)
start_stop_delete_test.go:262: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p no-preload-386690 -n no-preload-386690
--- PASS: TestStartStop/group/no-preload/serial/SecondStart (347.45s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/DeployApp (9.53s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/DeployApp
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context old-k8s-version-537179 create -f testdata/busybox.yaml
start_stop_delete_test.go:196: (dbg) TestStartStop/group/old-k8s-version/serial/DeployApp: waiting 8m0s for pods matching "integration-test=busybox" in namespace "default" ...
helpers_test.go:344: "busybox" [defc704e-1e0c-42e9-ab00-e349fb93f781] Pending
helpers_test.go:344: "busybox" [defc704e-1e0c-42e9-ab00-e349fb93f781] Pending / Ready:ContainersNotReady (containers with unready status: [busybox]) / ContainersReady:ContainersNotReady (containers with unready status: [busybox])
helpers_test.go:344: "busybox" [defc704e-1e0c-42e9-ab00-e349fb93f781] Running
start_stop_delete_test.go:196: (dbg) TestStartStop/group/old-k8s-version/serial/DeployApp: integration-test=busybox healthy within 9.037270209s
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context old-k8s-version-537179 exec busybox -- /bin/sh -c "ulimit -n"
--- PASS: TestStartStop/group/old-k8s-version/serial/DeployApp (9.53s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/EnableAddonWhileActive (1.23s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/EnableAddonWhileActive
start_stop_delete_test.go:205: (dbg) Run:  out/minikube-linux-arm64 addons enable metrics-server -p old-k8s-version-537179 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain
E0830 23:45:00.057728 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
start_stop_delete_test.go:205: (dbg) Done: out/minikube-linux-arm64 addons enable metrics-server -p old-k8s-version-537179 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain: (1.091690193s)
start_stop_delete_test.go:215: (dbg) Run:  kubectl --context old-k8s-version-537179 describe deploy/metrics-server -n kube-system
--- PASS: TestStartStop/group/old-k8s-version/serial/EnableAddonWhileActive (1.23s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/Stop (12.47s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/Stop
start_stop_delete_test.go:228: (dbg) Run:  out/minikube-linux-arm64 stop -p old-k8s-version-537179 --alsologtostderr -v=3
start_stop_delete_test.go:228: (dbg) Done: out/minikube-linux-arm64 stop -p old-k8s-version-537179 --alsologtostderr -v=3: (12.474721126s)
--- PASS: TestStartStop/group/old-k8s-version/serial/Stop (12.47s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/EnableAddonAfterStop (0.29s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/EnableAddonAfterStop
start_stop_delete_test.go:239: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p old-k8s-version-537179 -n old-k8s-version-537179
start_stop_delete_test.go:239: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.Host}} -p old-k8s-version-537179 -n old-k8s-version-537179: exit status 7 (135.016808ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:239: status error: exit status 7 (may be ok)
start_stop_delete_test.go:246: (dbg) Run:  out/minikube-linux-arm64 addons enable dashboard -p old-k8s-version-537179 --images=MetricsScraper=registry.k8s.io/echoserver:1.4
--- PASS: TestStartStop/group/old-k8s-version/serial/EnableAddonAfterStop (0.29s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/SecondStart (617.31s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/SecondStart
start_stop_delete_test.go:256: (dbg) Run:  out/minikube-linux-arm64 start -p old-k8s-version-537179 --memory=2200 --alsologtostderr --wait=true --kvm-network=default --kvm-qemu-uri=qemu:///system --disable-driver-mounts --keep-context=false --driver=docker  --container-runtime=containerd --kubernetes-version=v1.16.0
E0830 23:45:16.025290 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:45:23.425452 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:45:23.430714 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:45:23.440952 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:45:23.461203 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:45:23.501566 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:45:23.581806 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:45:23.742723 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:45:24.063848 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:45:24.312072 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0830 23:45:24.704593 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:45:25.984978 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:45:28.545642 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:45:33.666853 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:45:34.090135 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:45:34.095448 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:45:34.105870 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:45:34.126152 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:45:34.167278 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:45:34.247756 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:45:34.408212 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:45:34.729346 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:45:35.370291 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:45:36.650424 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:45:39.210893 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:45:43.907940 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:45:44.331595 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:45:54.571929 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:46:04.389119 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:46:15.052749 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:46:21.978747 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:46:37.946342 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:46:45.349803 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:46:46.823911 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:46:56.013524 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:47:03.266933 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:03.272223 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:03.282520 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:03.302647 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:03.343123 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:03.423504 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:03.583771 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:03.904287 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:04.544534 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:05.824706 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:06.148386 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:47:06.971467 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:47:06.977240 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:47:06.987543 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:47:07.007783 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:47:07.048130 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:47:07.128501 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:47:07.288859 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:47:07.609763 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:47:08.250318 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:47:08.385814 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:09.531013 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:47:12.091544 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:47:13.506811 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:14.509839 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:47:17.211848 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:47:23.747577 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:27.452316 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:47:44.061315 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:47:44.228136 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:47:47.932539 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:48:07.270028 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:48:17.934544 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:48:25.188383 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:48:28.893612 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:48:38.133006 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:48:54.103313 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:49:05.819527 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:49:21.786823 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:49:47.108923 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:49:50.814325 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:50:23.425984 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:50:24.311930 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
start_stop_delete_test.go:256: (dbg) Done: out/minikube-linux-arm64 start -p old-k8s-version-537179 --memory=2200 --alsologtostderr --wait=true --kvm-network=default --kvm-qemu-uri=qemu:///system --disable-driver-mounts --keep-context=false --driver=docker  --container-runtime=containerd --kubernetes-version=v1.16.0: (10m16.904047103s)
start_stop_delete_test.go:262: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p old-k8s-version-537179 -n old-k8s-version-537179
--- PASS: TestStartStop/group/old-k8s-version/serial/SecondStart (617.31s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/UserAppExistsAfterStop (10.03s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/UserAppExistsAfterStop
start_stop_delete_test.go:274: (dbg) TestStartStop/group/no-preload/serial/UserAppExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-8694d4445c-fjslc" [2fc32e88-132c-470b-bd3c-2fec9a6e3cc0] Pending / Ready:ContainersNotReady (containers with unready status: [kubernetes-dashboard]) / ContainersReady:ContainersNotReady (containers with unready status: [kubernetes-dashboard])
E0830 23:50:34.090797 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
helpers_test.go:344: "kubernetes-dashboard-8694d4445c-fjslc" [2fc32e88-132c-470b-bd3c-2fec9a6e3cc0] Running
start_stop_delete_test.go:274: (dbg) TestStartStop/group/no-preload/serial/UserAppExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 10.029674515s
--- PASS: TestStartStop/group/no-preload/serial/UserAppExistsAfterStop (10.03s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/AddonExistsAfterStop (5.12s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/AddonExistsAfterStop
start_stop_delete_test.go:287: (dbg) TestStartStop/group/no-preload/serial/AddonExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-8694d4445c-fjslc" [2fc32e88-132c-470b-bd3c-2fec9a6e3cc0] Running
start_stop_delete_test.go:287: (dbg) TestStartStop/group/no-preload/serial/AddonExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 5.014195234s
start_stop_delete_test.go:291: (dbg) Run:  kubectl --context no-preload-386690 describe deploy/dashboard-metrics-scraper -n kubernetes-dashboard
--- PASS: TestStartStop/group/no-preload/serial/AddonExistsAfterStop (5.12s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/VerifyKubernetesImages (0.4s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/VerifyKubernetesImages
start_stop_delete_test.go:304: (dbg) Run:  out/minikube-linux-arm64 ssh -p no-preload-386690 "sudo crictl images -o json"
start_stop_delete_test.go:304: Found non-minikube image: kindest/kindnetd:v20230511-dc714da8
start_stop_delete_test.go:304: Found non-minikube image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
--- PASS: TestStartStop/group/no-preload/serial/VerifyKubernetesImages (0.40s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/Pause (4.85s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/Pause
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 pause -p no-preload-386690 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Done: out/minikube-linux-arm64 pause -p no-preload-386690 --alsologtostderr -v=1: (1.396715818s)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p no-preload-386690 -n no-preload-386690
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.APIServer}} -p no-preload-386690 -n no-preload-386690: exit status 2 (486.253941ms)

                                                
                                                
-- stdout --
	Paused

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Kubelet}} -p no-preload-386690 -n no-preload-386690
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.Kubelet}} -p no-preload-386690 -n no-preload-386690: exit status 2 (517.921604ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 unpause -p no-preload-386690 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Done: out/minikube-linux-arm64 unpause -p no-preload-386690 --alsologtostderr -v=1: (1.090940624s)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p no-preload-386690 -n no-preload-386690
E0830 23:50:51.111033 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Kubelet}} -p no-preload-386690 -n no-preload-386690
--- PASS: TestStartStop/group/no-preload/serial/Pause (4.85s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/FirstStart (85.68s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/FirstStart
start_stop_delete_test.go:186: (dbg) Run:  out/minikube-linux-arm64 start -p embed-certs-344283 --memory=2200 --alsologtostderr --wait=true --embed-certs --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1
E0830 23:51:01.774747 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
E0830 23:51:46.824700 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
E0830 23:51:49.192737 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:52:03.267233 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:52:06.148402 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:52:06.971516 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
start_stop_delete_test.go:186: (dbg) Done: out/minikube-linux-arm64 start -p embed-certs-344283 --memory=2200 --alsologtostderr --wait=true --embed-certs --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1: (1m25.678674118s)
--- PASS: TestStartStop/group/embed-certs/serial/FirstStart (85.68s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/DeployApp (9.5s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/DeployApp
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context embed-certs-344283 create -f testdata/busybox.yaml
start_stop_delete_test.go:196: (dbg) TestStartStop/group/embed-certs/serial/DeployApp: waiting 8m0s for pods matching "integration-test=busybox" in namespace "default" ...
helpers_test.go:344: "busybox" [c8aecb66-cf29-4e21-bcc9-524a1feeee01] Pending
helpers_test.go:344: "busybox" [c8aecb66-cf29-4e21-bcc9-524a1feeee01] Pending / Ready:ContainersNotReady (containers with unready status: [busybox]) / ContainersReady:ContainersNotReady (containers with unready status: [busybox])
helpers_test.go:344: "busybox" [c8aecb66-cf29-4e21-bcc9-524a1feeee01] Running
start_stop_delete_test.go:196: (dbg) TestStartStop/group/embed-certs/serial/DeployApp: integration-test=busybox healthy within 9.046535565s
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context embed-certs-344283 exec busybox -- /bin/sh -c "ulimit -n"
--- PASS: TestStartStop/group/embed-certs/serial/DeployApp (9.50s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/EnableAddonWhileActive (1.28s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/EnableAddonWhileActive
start_stop_delete_test.go:205: (dbg) Run:  out/minikube-linux-arm64 addons enable metrics-server -p embed-certs-344283 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain
E0830 23:52:30.950259 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
start_stop_delete_test.go:205: (dbg) Done: out/minikube-linux-arm64 addons enable metrics-server -p embed-certs-344283 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain: (1.148501373s)
start_stop_delete_test.go:215: (dbg) Run:  kubectl --context embed-certs-344283 describe deploy/metrics-server -n kube-system
--- PASS: TestStartStop/group/embed-certs/serial/EnableAddonWhileActive (1.28s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/Stop (12.19s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/Stop
start_stop_delete_test.go:228: (dbg) Run:  out/minikube-linux-arm64 stop -p embed-certs-344283 --alsologtostderr -v=3
E0830 23:52:34.654812 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
E0830 23:52:44.061559 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
start_stop_delete_test.go:228: (dbg) Done: out/minikube-linux-arm64 stop -p embed-certs-344283 --alsologtostderr -v=3: (12.18600747s)
--- PASS: TestStartStop/group/embed-certs/serial/Stop (12.19s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/EnableAddonAfterStop (0.21s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/EnableAddonAfterStop
start_stop_delete_test.go:239: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p embed-certs-344283 -n embed-certs-344283
start_stop_delete_test.go:239: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.Host}} -p embed-certs-344283 -n embed-certs-344283: exit status 7 (80.081626ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:239: status error: exit status 7 (may be ok)
start_stop_delete_test.go:246: (dbg) Run:  out/minikube-linux-arm64 addons enable dashboard -p embed-certs-344283 --images=MetricsScraper=registry.k8s.io/echoserver:1.4
--- PASS: TestStartStop/group/embed-certs/serial/EnableAddonAfterStop (0.21s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/SecondStart (342.19s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/SecondStart
start_stop_delete_test.go:256: (dbg) Run:  out/minikube-linux-arm64 start -p embed-certs-344283 --memory=2200 --alsologtostderr --wait=true --embed-certs --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1
E0830 23:53:38.132941 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0830 23:53:54.103149 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:54:22.358986 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:54:22.364306 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:54:22.374569 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:54:22.394802 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:54:22.435110 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:54:22.515749 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:54:22.675981 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:54:22.996537 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:54:23.637556 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:54:24.917806 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:54:27.478620 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:54:32.599420 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:54:42.839908 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:55:03.320986 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:55:23.425471 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/custom-flannel-506633/client.crt: no such file or directory
E0830 23:55:24.311531 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
start_stop_delete_test.go:256: (dbg) Done: out/minikube-linux-arm64 start -p embed-certs-344283 --memory=2200 --alsologtostderr --wait=true --embed-certs --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1: (5m41.670968234s)
start_stop_delete_test.go:262: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p embed-certs-344283 -n embed-certs-344283
--- PASS: TestStartStop/group/embed-certs/serial/SecondStart (342.19s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/UserAppExistsAfterStop (5.02s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/UserAppExistsAfterStop
start_stop_delete_test.go:274: (dbg) TestStartStop/group/old-k8s-version/serial/UserAppExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-84b68f675b-r6zl7" [01eb4f79-54a5-449b-b32a-94d76ba45893] Running
E0830 23:55:34.090493 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/enable-default-cni-506633/client.crt: no such file or directory
start_stop_delete_test.go:274: (dbg) TestStartStop/group/old-k8s-version/serial/UserAppExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 5.023192826s
--- PASS: TestStartStop/group/old-k8s-version/serial/UserAppExistsAfterStop (5.02s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/AddonExistsAfterStop (5.12s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/AddonExistsAfterStop
start_stop_delete_test.go:287: (dbg) TestStartStop/group/old-k8s-version/serial/AddonExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-84b68f675b-r6zl7" [01eb4f79-54a5-449b-b32a-94d76ba45893] Running
start_stop_delete_test.go:287: (dbg) TestStartStop/group/old-k8s-version/serial/AddonExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 5.008968968s
start_stop_delete_test.go:291: (dbg) Run:  kubectl --context old-k8s-version-537179 describe deploy/dashboard-metrics-scraper -n kubernetes-dashboard
--- PASS: TestStartStop/group/old-k8s-version/serial/AddonExistsAfterStop (5.12s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/VerifyKubernetesImages (0.37s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/VerifyKubernetesImages
start_stop_delete_test.go:304: (dbg) Run:  out/minikube-linux-arm64 ssh -p old-k8s-version-537179 "sudo crictl images -o json"
start_stop_delete_test.go:304: Found non-minikube image: kindest/kindnetd:v20210326-1e038dc5
start_stop_delete_test.go:304: Found non-minikube image: kindest/kindnetd:v20230511-dc714da8
start_stop_delete_test.go:304: Found non-minikube image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
--- PASS: TestStartStop/group/old-k8s-version/serial/VerifyKubernetesImages (0.37s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/Pause (3.66s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/Pause
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 pause -p old-k8s-version-537179 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p old-k8s-version-537179 -n old-k8s-version-537179
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.APIServer}} -p old-k8s-version-537179 -n old-k8s-version-537179: exit status 2 (381.140274ms)

                                                
                                                
-- stdout --
	Paused

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Kubelet}} -p old-k8s-version-537179 -n old-k8s-version-537179
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.Kubelet}} -p old-k8s-version-537179 -n old-k8s-version-537179: exit status 2 (371.202562ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 unpause -p old-k8s-version-537179 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p old-k8s-version-537179 -n old-k8s-version-537179
E0830 23:55:44.282143 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Kubelet}} -p old-k8s-version-537179 -n old-k8s-version-537179
--- PASS: TestStartStop/group/old-k8s-version/serial/Pause (3.66s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/FirstStart (60.29s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/FirstStart
start_stop_delete_test.go:186: (dbg) Run:  out/minikube-linux-arm64 start -p default-k8s-diff-port-310625 --memory=2200 --alsologtostderr --wait=true --apiserver-port=8444 --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1
E0830 23:56:46.824825 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
start_stop_delete_test.go:186: (dbg) Done: out/minikube-linux-arm64 start -p default-k8s-diff-port-310625 --memory=2200 --alsologtostderr --wait=true --apiserver-port=8444 --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1: (1m0.289913745s)
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/FirstStart (60.29s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/DeployApp (8.5s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/DeployApp
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context default-k8s-diff-port-310625 create -f testdata/busybox.yaml
start_stop_delete_test.go:196: (dbg) TestStartStop/group/default-k8s-diff-port/serial/DeployApp: waiting 8m0s for pods matching "integration-test=busybox" in namespace "default" ...
helpers_test.go:344: "busybox" [7167b3f0-31f5-49d2-bad4-45c530c9ee23] Pending
helpers_test.go:344: "busybox" [7167b3f0-31f5-49d2-bad4-45c530c9ee23] Pending / Ready:ContainersNotReady (containers with unready status: [busybox]) / ContainersReady:ContainersNotReady (containers with unready status: [busybox])
helpers_test.go:344: "busybox" [7167b3f0-31f5-49d2-bad4-45c530c9ee23] Running
start_stop_delete_test.go:196: (dbg) TestStartStop/group/default-k8s-diff-port/serial/DeployApp: integration-test=busybox healthy within 8.040818825s
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context default-k8s-diff-port-310625 exec busybox -- /bin/sh -c "ulimit -n"
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/DeployApp (8.50s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/EnableAddonWhileActive (1.26s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/EnableAddonWhileActive
start_stop_delete_test.go:205: (dbg) Run:  out/minikube-linux-arm64 addons enable metrics-server -p default-k8s-diff-port-310625 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain
start_stop_delete_test.go:205: (dbg) Done: out/minikube-linux-arm64 addons enable metrics-server -p default-k8s-diff-port-310625 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain: (1.136965572s)
start_stop_delete_test.go:215: (dbg) Run:  kubectl --context default-k8s-diff-port-310625 describe deploy/metrics-server -n kube-system
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/EnableAddonWhileActive (1.26s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/Stop (12.2s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/Stop
start_stop_delete_test.go:228: (dbg) Run:  out/minikube-linux-arm64 stop -p default-k8s-diff-port-310625 --alsologtostderr -v=3
E0830 23:57:03.267823 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/bridge-506633/client.crt: no such file or directory
E0830 23:57:06.148348 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/addons-015166/client.crt: no such file or directory
E0830 23:57:06.202535 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:57:06.971791 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/flannel-506633/client.crt: no such file or directory
start_stop_delete_test.go:228: (dbg) Done: out/minikube-linux-arm64 stop -p default-k8s-diff-port-310625 --alsologtostderr -v=3: (12.199499888s)
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/Stop (12.20s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/EnableAddonAfterStop (0.22s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/EnableAddonAfterStop
start_stop_delete_test.go:239: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p default-k8s-diff-port-310625 -n default-k8s-diff-port-310625
start_stop_delete_test.go:239: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.Host}} -p default-k8s-diff-port-310625 -n default-k8s-diff-port-310625: exit status 7 (85.145286ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:239: status error: exit status 7 (may be ok)
start_stop_delete_test.go:246: (dbg) Run:  out/minikube-linux-arm64 addons enable dashboard -p default-k8s-diff-port-310625 --images=MetricsScraper=registry.k8s.io/echoserver:1.4
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/EnableAddonAfterStop (0.22s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/SecondStart (342.48s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/SecondStart
start_stop_delete_test.go:256: (dbg) Run:  out/minikube-linux-arm64 start -p default-k8s-diff-port-310625 --memory=2200 --alsologtostderr --wait=true --apiserver-port=8444 --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1
E0830 23:57:27.105807 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:57:44.061876 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/functional-479614/client.crt: no such file or directory
E0830 23:58:09.870707 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/auto-506633/client.crt: no such file or directory
start_stop_delete_test.go:256: (dbg) Done: out/minikube-linux-arm64 start -p default-k8s-diff-port-310625 --memory=2200 --alsologtostderr --wait=true --apiserver-port=8444 --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1: (5m41.959160743s)
start_stop_delete_test.go:262: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p default-k8s-diff-port-310625 -n default-k8s-diff-port-310625
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/SecondStart (342.48s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/UserAppExistsAfterStop (14.03s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/UserAppExistsAfterStop
start_stop_delete_test.go:274: (dbg) TestStartStop/group/embed-certs/serial/UserAppExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-8694d4445c-clr2l" [dd8ce2ce-c609-46cc-8ef5-3495182a9ed4] Pending / Ready:ContainersNotReady (containers with unready status: [kubernetes-dashboard]) / ContainersReady:ContainersNotReady (containers with unready status: [kubernetes-dashboard])
helpers_test.go:344: "kubernetes-dashboard-8694d4445c-clr2l" [dd8ce2ce-c609-46cc-8ef5-3495182a9ed4] Running
E0830 23:58:38.132631 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
start_stop_delete_test.go:274: (dbg) TestStartStop/group/embed-certs/serial/UserAppExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 14.03087576s
--- PASS: TestStartStop/group/embed-certs/serial/UserAppExistsAfterStop (14.03s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/AddonExistsAfterStop (5.12s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/AddonExistsAfterStop
start_stop_delete_test.go:287: (dbg) TestStartStop/group/embed-certs/serial/AddonExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-8694d4445c-clr2l" [dd8ce2ce-c609-46cc-8ef5-3495182a9ed4] Running
start_stop_delete_test.go:287: (dbg) TestStartStop/group/embed-certs/serial/AddonExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 5.011190383s
start_stop_delete_test.go:291: (dbg) Run:  kubectl --context embed-certs-344283 describe deploy/dashboard-metrics-scraper -n kubernetes-dashboard
--- PASS: TestStartStop/group/embed-certs/serial/AddonExistsAfterStop (5.12s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/VerifyKubernetesImages (0.38s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/VerifyKubernetesImages
start_stop_delete_test.go:304: (dbg) Run:  out/minikube-linux-arm64 ssh -p embed-certs-344283 "sudo crictl images -o json"
start_stop_delete_test.go:304: Found non-minikube image: kindest/kindnetd:v20230511-dc714da8
start_stop_delete_test.go:304: Found non-minikube image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
--- PASS: TestStartStop/group/embed-certs/serial/VerifyKubernetesImages (0.38s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/Pause (3.52s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/Pause
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 pause -p embed-certs-344283 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p embed-certs-344283 -n embed-certs-344283
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.APIServer}} -p embed-certs-344283 -n embed-certs-344283: exit status 2 (371.763396ms)

                                                
                                                
-- stdout --
	Paused

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Kubelet}} -p embed-certs-344283 -n embed-certs-344283
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.Kubelet}} -p embed-certs-344283 -n embed-certs-344283: exit status 2 (375.536467ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 unpause -p embed-certs-344283 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p embed-certs-344283 -n embed-certs-344283
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Kubelet}} -p embed-certs-344283 -n embed-certs-344283
--- PASS: TestStartStop/group/embed-certs/serial/Pause (3.52s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/FirstStart (46.46s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/FirstStart
start_stop_delete_test.go:186: (dbg) Run:  out/minikube-linux-arm64 start -p newest-cni-540282 --memory=2200 --alsologtostderr --wait=apiserver,system_pods,default_sa --feature-gates ServerSideApply=true --network-plugin=cni --extra-config=kubeadm.pod-network-cidr=10.42.0.0/16 --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1
E0830 23:58:54.103213 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
E0830 23:59:22.359421 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
start_stop_delete_test.go:186: (dbg) Done: out/minikube-linux-arm64 start -p newest-cni-540282 --memory=2200 --alsologtostderr --wait=apiserver,system_pods,default_sa --feature-gates ServerSideApply=true --network-plugin=cni --extra-config=kubeadm.pod-network-cidr=10.42.0.0/16 --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1: (46.461300478s)
--- PASS: TestStartStop/group/newest-cni/serial/FirstStart (46.46s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/DeployApp (0s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/DeployApp
--- PASS: TestStartStop/group/newest-cni/serial/DeployApp (0.00s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/EnableAddonWhileActive (1.47s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/EnableAddonWhileActive
start_stop_delete_test.go:205: (dbg) Run:  out/minikube-linux-arm64 addons enable metrics-server -p newest-cni-540282 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain
start_stop_delete_test.go:205: (dbg) Done: out/minikube-linux-arm64 addons enable metrics-server -p newest-cni-540282 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain: (1.467665166s)
start_stop_delete_test.go:211: WARNING: cni mode requires additional setup before pods can schedule :(
--- PASS: TestStartStop/group/newest-cni/serial/EnableAddonWhileActive (1.47s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/Stop (1.32s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/Stop
start_stop_delete_test.go:228: (dbg) Run:  out/minikube-linux-arm64 stop -p newest-cni-540282 --alsologtostderr -v=3
start_stop_delete_test.go:228: (dbg) Done: out/minikube-linux-arm64 stop -p newest-cni-540282 --alsologtostderr -v=3: (1.321019853s)
--- PASS: TestStartStop/group/newest-cni/serial/Stop (1.32s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/EnableAddonAfterStop (0.23s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/EnableAddonAfterStop
start_stop_delete_test.go:239: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p newest-cni-540282 -n newest-cni-540282
start_stop_delete_test.go:239: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.Host}} -p newest-cni-540282 -n newest-cni-540282: exit status 7 (90.423443ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:239: status error: exit status 7 (may be ok)
start_stop_delete_test.go:246: (dbg) Run:  out/minikube-linux-arm64 addons enable dashboard -p newest-cni-540282 --images=MetricsScraper=registry.k8s.io/echoserver:1.4
--- PASS: TestStartStop/group/newest-cni/serial/EnableAddonAfterStop (0.23s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/SecondStart (31.95s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/SecondStart
start_stop_delete_test.go:256: (dbg) Run:  out/minikube-linux-arm64 start -p newest-cni-540282 --memory=2200 --alsologtostderr --wait=apiserver,system_pods,default_sa --feature-gates ServerSideApply=true --network-plugin=cni --extra-config=kubeadm.pod-network-cidr=10.42.0.0/16 --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1
E0830 23:59:50.043316 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/no-preload-386690/client.crt: no such file or directory
E0830 23:59:50.356402 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0830 23:59:50.361713 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0830 23:59:50.372048 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0830 23:59:50.392301 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0830 23:59:50.432831 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0830 23:59:50.513218 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0830 23:59:50.673653 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0830 23:59:50.994332 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0830 23:59:51.634827 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0830 23:59:52.915322 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0830 23:59:55.476068 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0831 00:00:00.596886 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
E0831 00:00:01.180398 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/calico-506633/client.crt: no such file or directory
E0831 00:00:07.356409 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
E0831 00:00:10.837923 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/old-k8s-version-537179/client.crt: no such file or directory
start_stop_delete_test.go:256: (dbg) Done: out/minikube-linux-arm64 start -p newest-cni-540282 --memory=2200 --alsologtostderr --wait=apiserver,system_pods,default_sa --feature-gates ServerSideApply=true --network-plugin=cni --extra-config=kubeadm.pod-network-cidr=10.42.0.0/16 --driver=docker  --container-runtime=containerd --kubernetes-version=v1.28.1: (31.521671206s)
start_stop_delete_test.go:262: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Host}} -p newest-cni-540282 -n newest-cni-540282
--- PASS: TestStartStop/group/newest-cni/serial/SecondStart (31.95s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/UserAppExistsAfterStop (0s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/UserAppExistsAfterStop
start_stop_delete_test.go:273: WARNING: cni mode requires additional setup before pods can schedule :(
--- PASS: TestStartStop/group/newest-cni/serial/UserAppExistsAfterStop (0.00s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/AddonExistsAfterStop (0s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/AddonExistsAfterStop
start_stop_delete_test.go:284: WARNING: cni mode requires additional setup before pods can schedule :(
--- PASS: TestStartStop/group/newest-cni/serial/AddonExistsAfterStop (0.00s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/VerifyKubernetesImages (0.38s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/VerifyKubernetesImages
start_stop_delete_test.go:304: (dbg) Run:  out/minikube-linux-arm64 ssh -p newest-cni-540282 "sudo crictl images -o json"
start_stop_delete_test.go:304: Found non-minikube image: kindest/kindnetd:v20230511-dc714da8
--- PASS: TestStartStop/group/newest-cni/serial/VerifyKubernetesImages (0.38s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/Pause (3.51s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/Pause
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 pause -p newest-cni-540282 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p newest-cni-540282 -n newest-cni-540282
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.APIServer}} -p newest-cni-540282 -n newest-cni-540282: exit status 2 (400.554822ms)

                                                
                                                
-- stdout --
	Paused

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Kubelet}} -p newest-cni-540282 -n newest-cni-540282
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.Kubelet}} -p newest-cni-540282 -n newest-cni-540282: exit status 2 (400.609214ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 unpause -p newest-cni-540282 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p newest-cni-540282 -n newest-cni-540282
E0831 00:00:17.147419 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/kindnet-506633/client.crt: no such file or directory
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Kubelet}} -p newest-cni-540282 -n newest-cni-540282
--- PASS: TestStartStop/group/newest-cni/serial/Pause (3.51s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/UserAppExistsAfterStop (9.03s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/UserAppExistsAfterStop
start_stop_delete_test.go:274: (dbg) TestStartStop/group/default-k8s-diff-port/serial/UserAppExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-8694d4445c-vmdd9" [2c3fcc29-b1f2-4b4e-a13a-f9adfe029893] Pending / Ready:ContainersNotReady (containers with unready status: [kubernetes-dashboard]) / ContainersReady:ContainersNotReady (containers with unready status: [kubernetes-dashboard])
helpers_test.go:344: "kubernetes-dashboard-8694d4445c-vmdd9" [2c3fcc29-b1f2-4b4e-a13a-f9adfe029893] Running
start_stop_delete_test.go:274: (dbg) TestStartStop/group/default-k8s-diff-port/serial/UserAppExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 9.028733563s
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/UserAppExistsAfterStop (9.03s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/AddonExistsAfterStop (5.14s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/AddonExistsAfterStop
start_stop_delete_test.go:287: (dbg) TestStartStop/group/default-k8s-diff-port/serial/AddonExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-8694d4445c-vmdd9" [2c3fcc29-b1f2-4b4e-a13a-f9adfe029893] Running
start_stop_delete_test.go:287: (dbg) TestStartStop/group/default-k8s-diff-port/serial/AddonExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 5.027228573s
start_stop_delete_test.go:291: (dbg) Run:  kubectl --context default-k8s-diff-port-310625 describe deploy/dashboard-metrics-scraper -n kubernetes-dashboard
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/AddonExistsAfterStop (5.14s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/VerifyKubernetesImages (0.37s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/VerifyKubernetesImages
start_stop_delete_test.go:304: (dbg) Run:  out/minikube-linux-arm64 ssh -p default-k8s-diff-port-310625 "sudo crictl images -o json"
start_stop_delete_test.go:304: Found non-minikube image: kindest/kindnetd:v20230511-dc714da8
start_stop_delete_test.go:304: Found non-minikube image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/VerifyKubernetesImages (0.37s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/Pause (3.36s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/Pause
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 pause -p default-k8s-diff-port-310625 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p default-k8s-diff-port-310625 -n default-k8s-diff-port-310625
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.APIServer}} -p default-k8s-diff-port-310625 -n default-k8s-diff-port-310625: exit status 2 (364.139976ms)

                                                
                                                
-- stdout --
	Paused

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Kubelet}} -p default-k8s-diff-port-310625 -n default-k8s-diff-port-310625
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-arm64 status --format={{.Kubelet}} -p default-k8s-diff-port-310625 -n default-k8s-diff-port-310625: exit status 2 (354.517814ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 unpause -p default-k8s-diff-port-310625 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.APIServer}} -p default-k8s-diff-port-310625 -n default-k8s-diff-port-310625
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-arm64 status --format={{.Kubelet}} -p default-k8s-diff-port-310625 -n default-k8s-diff-port-310625
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/Pause (3.36s)

                                                
                                    

Test skip (28/304)

x
+
TestDownloadOnly/v1.16.0/cached-images (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.16.0/cached-images
aaa_download_only_test.go:117: Preload exists, images won't be cached
--- SKIP: TestDownloadOnly/v1.16.0/cached-images (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.16.0/binaries (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.16.0/binaries
aaa_download_only_test.go:136: Preload exists, binaries are present within.
--- SKIP: TestDownloadOnly/v1.16.0/binaries (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.16.0/kubectl (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.16.0/kubectl
aaa_download_only_test.go:152: Test for darwin and windows
--- SKIP: TestDownloadOnly/v1.16.0/kubectl (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.28.1/cached-images (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.28.1/cached-images
aaa_download_only_test.go:117: Preload exists, images won't be cached
--- SKIP: TestDownloadOnly/v1.28.1/cached-images (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.28.1/binaries (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.28.1/binaries
aaa_download_only_test.go:136: Preload exists, binaries are present within.
--- SKIP: TestDownloadOnly/v1.28.1/binaries (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.28.1/kubectl (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.28.1/kubectl
aaa_download_only_test.go:152: Test for darwin and windows
--- SKIP: TestDownloadOnly/v1.28.1/kubectl (0.00s)

                                                
                                    
x
+
TestDownloadOnlyKic (0.6s)

                                                
                                                
=== RUN   TestDownloadOnlyKic
aaa_download_only_test.go:222: (dbg) Run:  out/minikube-linux-arm64 start --download-only -p download-docker-266424 --alsologtostderr --driver=docker  --container-runtime=containerd
aaa_download_only_test.go:234: Skip for arm64 platform. See https://github.com/kubernetes/minikube/issues/10144
helpers_test.go:175: Cleaning up "download-docker-266424" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p download-docker-266424
--- SKIP: TestDownloadOnlyKic (0.60s)

                                                
                                    
x
+
TestOffline (0s)

                                                
                                                
=== RUN   TestOffline
=== PAUSE TestOffline

                                                
                                                

                                                
                                                
=== CONT  TestOffline
aab_offline_test.go:35: skipping TestOffline - only docker runtime supported on arm64. See https://github.com/kubernetes/minikube/issues/10144
--- SKIP: TestOffline (0.00s)

                                                
                                    
x
+
TestAddons/parallel/HelmTiller (0s)

                                                
                                                
=== RUN   TestAddons/parallel/HelmTiller
=== PAUSE TestAddons/parallel/HelmTiller

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/HelmTiller
addons_test.go:420: skip Helm test on arm64
--- SKIP: TestAddons/parallel/HelmTiller (0.00s)

                                                
                                    
x
+
TestAddons/parallel/Olm (0s)

                                                
                                                
=== RUN   TestAddons/parallel/Olm
=== PAUSE TestAddons/parallel/Olm

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/Olm
addons_test.go:474: Skipping OLM addon test until https://github.com/operator-framework/operator-lifecycle-manager/issues/2534 is resolved
--- SKIP: TestAddons/parallel/Olm (0.00s)

                                                
                                    
x
+
TestDockerFlags (0s)

                                                
                                                
=== RUN   TestDockerFlags
docker_test.go:41: skipping: only runs with docker container runtime, currently testing containerd
--- SKIP: TestDockerFlags (0.00s)

                                                
                                    
x
+
TestKVMDriverInstallOrUpdate (0s)

                                                
                                                
=== RUN   TestKVMDriverInstallOrUpdate
driver_install_or_update_test.go:45: Skip if arm64. See https://github.com/kubernetes/minikube/issues/10144
--- SKIP: TestKVMDriverInstallOrUpdate (0.00s)

                                                
                                    
x
+
TestHyperKitDriverInstallOrUpdate (0s)

                                                
                                                
=== RUN   TestHyperKitDriverInstallOrUpdate
driver_install_or_update_test.go:105: Skip if not darwin.
--- SKIP: TestHyperKitDriverInstallOrUpdate (0.00s)

                                                
                                    
x
+
TestHyperkitDriverSkipUpgrade (0s)

                                                
                                                
=== RUN   TestHyperkitDriverSkipUpgrade
driver_install_or_update_test.go:169: Skip if not darwin.
--- SKIP: TestHyperkitDriverSkipUpgrade (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/MySQL (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/MySQL
=== PAUSE TestFunctional/parallel/MySQL

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/MySQL
functional_test.go:1783: arm64 is not supported by mysql. Skip the test. See https://github.com/kubernetes/minikube/issues/10144
--- SKIP: TestFunctional/parallel/MySQL (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/DockerEnv (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/DockerEnv
=== PAUSE TestFunctional/parallel/DockerEnv

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/DockerEnv
functional_test.go:459: only validate docker env with docker container runtime, currently testing containerd
--- SKIP: TestFunctional/parallel/DockerEnv (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/PodmanEnv (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/PodmanEnv
=== PAUSE TestFunctional/parallel/PodmanEnv

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/PodmanEnv
functional_test.go:546: only validate podman env with docker container runtime, currently testing containerd
--- SKIP: TestFunctional/parallel/PodmanEnv (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/DNSResolutionByDig (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/DNSResolutionByDig
functional_test_tunnel_test.go:99: DNS forwarding is only supported for Hyperkit on Darwin, skipping test DNS forwarding
--- SKIP: TestFunctional/parallel/TunnelCmd/serial/DNSResolutionByDig (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/DNSResolutionByDscacheutil (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/DNSResolutionByDscacheutil
functional_test_tunnel_test.go:99: DNS forwarding is only supported for Hyperkit on Darwin, skipping test DNS forwarding
--- SKIP: TestFunctional/parallel/TunnelCmd/serial/DNSResolutionByDscacheutil (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/AccessThroughDNS (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/AccessThroughDNS
functional_test_tunnel_test.go:99: DNS forwarding is only supported for Hyperkit on Darwin, skipping test DNS forwarding
--- SKIP: TestFunctional/parallel/TunnelCmd/serial/AccessThroughDNS (0.00s)

                                                
                                    
x
+
TestGvisorAddon (0s)

                                                
                                                
=== RUN   TestGvisorAddon
gvisor_addon_test.go:34: skipping test because --gvisor=false
--- SKIP: TestGvisorAddon (0.00s)

                                                
                                    
x
+
TestImageBuild (0s)

                                                
                                                
=== RUN   TestImageBuild
image_test.go:33: 
--- SKIP: TestImageBuild (0.00s)

                                                
                                    
x
+
TestChangeNoneUser (0s)

                                                
                                                
=== RUN   TestChangeNoneUser
none_test.go:38: Test requires none driver and SUDO_USER env to not be empty
--- SKIP: TestChangeNoneUser (0.00s)

                                                
                                    
x
+
TestScheduledStopWindows (0s)

                                                
                                                
=== RUN   TestScheduledStopWindows
scheduled_stop_test.go:42: test only runs on windows
--- SKIP: TestScheduledStopWindows (0.00s)

                                                
                                    
x
+
TestSkaffold (0s)

                                                
                                                
=== RUN   TestSkaffold
skaffold_test.go:45: skaffold requires docker-env, currently testing containerd container runtime
--- SKIP: TestSkaffold (0.00s)

                                                
                                    
x
+
TestNetworkPlugins/group/kubenet (4.04s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kubenet
net_test.go:93: Skipping the test as containerd container runtimes requires CNI
E0830 23:25:24.312359 1225375 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/ingress-addon-legacy-235029/client.crt: no such file or directory
panic.go:522: 
----------------------- debugLogs start: kubenet-506633 [pass: true] --------------------------------
>>> netcat: nslookup kubernetes.default:
Error in configuration: context was not found for specified context: kubenet-506633

                                                
                                                

                                                
                                                
>>> netcat: nslookup debug kubernetes.default a-records:
Error in configuration: context was not found for specified context: kubenet-506633

                                                
                                                

                                                
                                                
>>> netcat: dig search kubernetes.default:
Error in configuration: context was not found for specified context: kubenet-506633

                                                
                                                

                                                
                                                
>>> netcat: dig @10.96.0.10 kubernetes.default.svc.cluster.local udp/53:
Error in configuration: context was not found for specified context: kubenet-506633

                                                
                                                

                                                
                                                
>>> netcat: dig @10.96.0.10 kubernetes.default.svc.cluster.local tcp/53:
Error in configuration: context was not found for specified context: kubenet-506633

                                                
                                                

                                                
                                                
>>> netcat: nc 10.96.0.10 udp/53:
Error in configuration: context was not found for specified context: kubenet-506633

                                                
                                                

                                                
                                                
>>> netcat: nc 10.96.0.10 tcp/53:
Error in configuration: context was not found for specified context: kubenet-506633

                                                
                                                

                                                
                                                
>>> netcat: /etc/nsswitch.conf:
Error in configuration: context was not found for specified context: kubenet-506633

                                                
                                                

                                                
                                                
>>> netcat: /etc/hosts:
Error in configuration: context was not found for specified context: kubenet-506633

                                                
                                                

                                                
                                                
>>> netcat: /etc/resolv.conf:
Error in configuration: context was not found for specified context: kubenet-506633

                                                
                                                

                                                
                                                
>>> host: /etc/nsswitch.conf:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/hosts:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/resolv.conf:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> k8s: nodes, services, endpoints, daemon sets, deployments and pods, :
Error in configuration: context was not found for specified context: kubenet-506633

                                                
                                                

                                                
                                                
>>> host: crictl pods:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: crictl containers:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> k8s: describe netcat deployment:
error: context "kubenet-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe netcat pod(s):
error: context "kubenet-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: netcat logs:
error: context "kubenet-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe coredns deployment:
error: context "kubenet-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe coredns pods:
error: context "kubenet-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: coredns logs:
error: context "kubenet-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe api server pod(s):
error: context "kubenet-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: api server logs:
error: context "kubenet-506633" does not exist

                                                
                                                

                                                
                                                
>>> host: /etc/cni:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: ip a s:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: ip r s:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: iptables-save:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: iptables table nat:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> k8s: describe kube-proxy daemon set:
error: context "kubenet-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe kube-proxy pod(s):
error: context "kubenet-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: kube-proxy logs:
error: context "kubenet-506633" does not exist

                                                
                                                

                                                
                                                
>>> host: kubelet daemon status:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: kubelet daemon config:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> k8s: kubelet logs:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/kubernetes/kubelet.conf:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: /var/lib/kubelet/config.yaml:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> k8s: kubectl config:
apiVersion: v1
clusters:
- cluster:
certificate-authority: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt
extensions:
- extension:
last-update: Wed, 30 Aug 2023 23:24:58 UTC
provider: minikube.sigs.k8s.io
version: v1.31.2
name: cluster_info
server: https://192.168.67.2:8443
name: pause-498674
contexts:
- context:
cluster: pause-498674
extensions:
- extension:
last-update: Wed, 30 Aug 2023 23:24:58 UTC
provider: minikube.sigs.k8s.io
version: v1.31.2
name: context_info
namespace: default
user: pause-498674
name: pause-498674
current-context: pause-498674
kind: Config
preferences: {}
users:
- name: pause-498674
user:
client-certificate: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/pause-498674/client.crt
client-key: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/pause-498674/client.key

                                                
                                                

                                                
                                                
>>> k8s: cms:
Error in configuration: context was not found for specified context: kubenet-506633

                                                
                                                

                                                
                                                
>>> host: docker daemon status:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: docker daemon config:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/docker/daemon.json:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: docker system info:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: cri-docker daemon status:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: cri-docker daemon config:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/systemd/system/cri-docker.service.d/10-cni.conf:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: /usr/lib/systemd/system/cri-docker.service:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: cri-dockerd version:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: containerd daemon status:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: containerd daemon config:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: /lib/systemd/system/containerd.service:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/containerd/config.toml:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: containerd config dump:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: crio daemon status:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: crio daemon config:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/crio:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                

                                                
                                                
>>> host: crio config:
* Profile "kubenet-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p kubenet-506633"

                                                
                                                
----------------------- debugLogs end: kubenet-506633 [took: 3.847067011s] --------------------------------
helpers_test.go:175: Cleaning up "kubenet-506633" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p kubenet-506633
--- SKIP: TestNetworkPlugins/group/kubenet (4.04s)

                                                
                                    
x
+
TestNetworkPlugins/group/cilium (4.44s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/cilium
net_test.go:102: Skipping the test as it's interfering with other tests and is outdated
panic.go:522: 
----------------------- debugLogs start: cilium-506633 [pass: true] --------------------------------
>>> netcat: nslookup kubernetes.default:
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> netcat: nslookup debug kubernetes.default a-records:
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> netcat: dig search kubernetes.default:
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> netcat: dig @10.96.0.10 kubernetes.default.svc.cluster.local udp/53:
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> netcat: dig @10.96.0.10 kubernetes.default.svc.cluster.local tcp/53:
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> netcat: nc 10.96.0.10 udp/53:
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> netcat: nc 10.96.0.10 tcp/53:
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> netcat: /etc/nsswitch.conf:
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> netcat: /etc/hosts:
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> netcat: /etc/resolv.conf:
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> host: /etc/nsswitch.conf:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/hosts:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/resolv.conf:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> k8s: nodes, services, endpoints, daemon sets, deployments and pods, :
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> host: crictl pods:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: crictl containers:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> k8s: describe netcat deployment:
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe netcat pod(s):
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: netcat logs:
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe coredns deployment:
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe coredns pods:
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: coredns logs:
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe api server pod(s):
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: api server logs:
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> host: /etc/cni:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: ip a s:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: ip r s:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: iptables-save:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: iptables table nat:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> k8s: describe cilium daemon set:
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> k8s: describe cilium daemon set pod(s):
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> k8s: cilium daemon set container(s) logs (current):
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: cilium daemon set container(s) logs (previous):
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe cilium deployment:
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> k8s: describe cilium deployment pod(s):
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> k8s: cilium deployment container(s) logs (current):
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: cilium deployment container(s) logs (previous):
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe kube-proxy daemon set:
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe kube-proxy pod(s):
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> k8s: kube-proxy logs:
error: context "cilium-506633" does not exist

                                                
                                                

                                                
                                                
>>> host: kubelet daemon status:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: kubelet daemon config:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> k8s: kubelet logs:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/kubernetes/kubelet.conf:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: /var/lib/kubelet/config.yaml:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> k8s: kubectl config:
apiVersion: v1
clusters:
- cluster:
certificate-authority: /home/jenkins/minikube-integration/17114-1219981/.minikube/ca.crt
extensions:
- extension:
last-update: Wed, 30 Aug 2023 23:24:58 UTC
provider: minikube.sigs.k8s.io
version: v1.31.2
name: cluster_info
server: https://192.168.67.2:8443
name: pause-498674
contexts:
- context:
cluster: pause-498674
extensions:
- extension:
last-update: Wed, 30 Aug 2023 23:24:58 UTC
provider: minikube.sigs.k8s.io
version: v1.31.2
name: context_info
namespace: default
user: pause-498674
name: pause-498674
current-context: pause-498674
kind: Config
preferences: {}
users:
- name: pause-498674
user:
client-certificate: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/pause-498674/client.crt
client-key: /home/jenkins/minikube-integration/17114-1219981/.minikube/profiles/pause-498674/client.key

                                                
                                                

                                                
                                                
>>> k8s: cms:
Error in configuration: context was not found for specified context: cilium-506633

                                                
                                                

                                                
                                                
>>> host: docker daemon status:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: docker daemon config:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/docker/daemon.json:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: docker system info:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: cri-docker daemon status:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: cri-docker daemon config:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/systemd/system/cri-docker.service.d/10-cni.conf:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: /usr/lib/systemd/system/cri-docker.service:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: cri-dockerd version:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: containerd daemon status:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: containerd daemon config:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: /lib/systemd/system/containerd.service:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/containerd/config.toml:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: containerd config dump:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: crio daemon status:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: crio daemon config:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: /etc/crio:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                

                                                
                                                
>>> host: crio config:
* Profile "cilium-506633" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-506633"

                                                
                                                
----------------------- debugLogs end: cilium-506633 [took: 4.25394624s] --------------------------------
helpers_test.go:175: Cleaning up "cilium-506633" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p cilium-506633
--- SKIP: TestNetworkPlugins/group/cilium (4.44s)

                                                
                                    
x
+
TestStartStop/group/disable-driver-mounts (0.19s)

                                                
                                                
=== RUN   TestStartStop/group/disable-driver-mounts
=== PAUSE TestStartStop/group/disable-driver-mounts

                                                
                                                

                                                
                                                
=== CONT  TestStartStop/group/disable-driver-mounts
start_stop_delete_test.go:103: skipping TestStartStop/group/disable-driver-mounts - only runs on virtualbox
helpers_test.go:175: Cleaning up "disable-driver-mounts-811488" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-arm64 delete -p disable-driver-mounts-811488
--- SKIP: TestStartStop/group/disable-driver-mounts (0.19s)

                                                
                                    
Copied to clipboard