Test Report: Docker_Linux 19636

                    
                      a6feba20ebb4dc887776b248ea5c810d31cc7846:2024-09-13:36198
                    
                

Test fail (1/342)

Order failed test Duration
33 TestAddons/parallel/Registry 72.37
x
+
TestAddons/parallel/Registry (72.37s)

                                                
                                                
=== RUN   TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/Registry
addons_test.go:328: registry stabilized in 1.950457ms
addons_test.go:330: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-66c9cd494c-cvtgf" [58047cca-a75a-41df-bc9f-a91dc5a547ca] Running
addons_test.go:330: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 5.002440512s
addons_test.go:333: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-f7ltj" [7dfd72e6-65ac-495b-9832-0e900d22d7e6] Running
addons_test.go:333: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.002630036s
addons_test.go:338: (dbg) Run:  kubectl --context addons-315216 delete po -l run=registry-test --now
addons_test.go:343: (dbg) Run:  kubectl --context addons-315216 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:343: (dbg) Non-zero exit: kubectl --context addons-315216 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.068810795s)

                                                
                                                
-- stdout --
	pod "registry-test" deleted

                                                
                                                
-- /stdout --
** stderr ** 
	error: timed out waiting for the condition

                                                
                                                
** /stderr **
addons_test.go:345: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-315216 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:349: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:357: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 ip
2024/09/13 18:33:55 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:386: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======>  post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run:  docker inspect addons-315216
helpers_test.go:235: (dbg) docker inspect addons-315216:

                                                
                                                
-- stdout --
	[
	    {
	        "Id": "597eb6e49b799c2c80ebeda6fe4177eac5914f7d15d7d4f6d592a9307b9dd73a",
	        "Created": "2024-09-13T18:21:00.949028173Z",
	        "Path": "/usr/local/bin/entrypoint",
	        "Args": [
	            "/sbin/init"
	        ],
	        "State": {
	            "Status": "running",
	            "Running": true,
	            "Paused": false,
	            "Restarting": false,
	            "OOMKilled": false,
	            "Dead": false,
	            "Pid": 12513,
	            "ExitCode": 0,
	            "Error": "",
	            "StartedAt": "2024-09-13T18:21:01.074591628Z",
	            "FinishedAt": "0001-01-01T00:00:00Z"
	        },
	        "Image": "sha256:0dbfa4e10de72ad5b19052941be6a0a31f45a6011543305e0465758d912df3d4",
	        "ResolvConfPath": "/var/lib/docker/containers/597eb6e49b799c2c80ebeda6fe4177eac5914f7d15d7d4f6d592a9307b9dd73a/resolv.conf",
	        "HostnamePath": "/var/lib/docker/containers/597eb6e49b799c2c80ebeda6fe4177eac5914f7d15d7d4f6d592a9307b9dd73a/hostname",
	        "HostsPath": "/var/lib/docker/containers/597eb6e49b799c2c80ebeda6fe4177eac5914f7d15d7d4f6d592a9307b9dd73a/hosts",
	        "LogPath": "/var/lib/docker/containers/597eb6e49b799c2c80ebeda6fe4177eac5914f7d15d7d4f6d592a9307b9dd73a/597eb6e49b799c2c80ebeda6fe4177eac5914f7d15d7d4f6d592a9307b9dd73a-json.log",
	        "Name": "/addons-315216",
	        "RestartCount": 0,
	        "Driver": "overlay2",
	        "Platform": "linux",
	        "MountLabel": "",
	        "ProcessLabel": "",
	        "AppArmorProfile": "unconfined",
	        "ExecIDs": null,
	        "HostConfig": {
	            "Binds": [
	                "addons-315216:/var",
	                "/lib/modules:/lib/modules:ro"
	            ],
	            "ContainerIDFile": "",
	            "LogConfig": {
	                "Type": "json-file",
	                "Config": {
	                    "max-size": "100m"
	                }
	            },
	            "NetworkMode": "addons-315216",
	            "PortBindings": {
	                "22/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "2376/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "32443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "5000/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ],
	                "8443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": ""
	                    }
	                ]
	            },
	            "RestartPolicy": {
	                "Name": "no",
	                "MaximumRetryCount": 0
	            },
	            "AutoRemove": false,
	            "VolumeDriver": "",
	            "VolumesFrom": null,
	            "ConsoleSize": [
	                0,
	                0
	            ],
	            "CapAdd": null,
	            "CapDrop": null,
	            "CgroupnsMode": "host",
	            "Dns": [],
	            "DnsOptions": [],
	            "DnsSearch": [],
	            "ExtraHosts": null,
	            "GroupAdd": null,
	            "IpcMode": "private",
	            "Cgroup": "",
	            "Links": null,
	            "OomScoreAdj": 0,
	            "PidMode": "",
	            "Privileged": true,
	            "PublishAllPorts": false,
	            "ReadonlyRootfs": false,
	            "SecurityOpt": [
	                "seccomp=unconfined",
	                "apparmor=unconfined",
	                "label=disable"
	            ],
	            "Tmpfs": {
	                "/run": "",
	                "/tmp": ""
	            },
	            "UTSMode": "",
	            "UsernsMode": "",
	            "ShmSize": 67108864,
	            "Runtime": "runc",
	            "Isolation": "",
	            "CpuShares": 0,
	            "Memory": 4194304000,
	            "NanoCpus": 2000000000,
	            "CgroupParent": "",
	            "BlkioWeight": 0,
	            "BlkioWeightDevice": [],
	            "BlkioDeviceReadBps": [],
	            "BlkioDeviceWriteBps": [],
	            "BlkioDeviceReadIOps": [],
	            "BlkioDeviceWriteIOps": [],
	            "CpuPeriod": 0,
	            "CpuQuota": 0,
	            "CpuRealtimePeriod": 0,
	            "CpuRealtimeRuntime": 0,
	            "CpusetCpus": "",
	            "CpusetMems": "",
	            "Devices": [],
	            "DeviceCgroupRules": null,
	            "DeviceRequests": null,
	            "MemoryReservation": 0,
	            "MemorySwap": 8388608000,
	            "MemorySwappiness": null,
	            "OomKillDisable": false,
	            "PidsLimit": null,
	            "Ulimits": [],
	            "CpuCount": 0,
	            "CpuPercent": 0,
	            "IOMaximumIOps": 0,
	            "IOMaximumBandwidth": 0,
	            "MaskedPaths": null,
	            "ReadonlyPaths": null
	        },
	        "GraphDriver": {
	            "Data": {
	                "LowerDir": "/var/lib/docker/overlay2/cd747b1eb8e45dbe960648592581b099af7779cbf13094bf4b15e3727b2c3b1b-init/diff:/var/lib/docker/overlay2/e031d5ccc36ea399e2cfba0c5681544df856a4f055e288cbaea3e8f185f63675/diff",
	                "MergedDir": "/var/lib/docker/overlay2/cd747b1eb8e45dbe960648592581b099af7779cbf13094bf4b15e3727b2c3b1b/merged",
	                "UpperDir": "/var/lib/docker/overlay2/cd747b1eb8e45dbe960648592581b099af7779cbf13094bf4b15e3727b2c3b1b/diff",
	                "WorkDir": "/var/lib/docker/overlay2/cd747b1eb8e45dbe960648592581b099af7779cbf13094bf4b15e3727b2c3b1b/work"
	            },
	            "Name": "overlay2"
	        },
	        "Mounts": [
	            {
	                "Type": "volume",
	                "Name": "addons-315216",
	                "Source": "/var/lib/docker/volumes/addons-315216/_data",
	                "Destination": "/var",
	                "Driver": "local",
	                "Mode": "z",
	                "RW": true,
	                "Propagation": ""
	            },
	            {
	                "Type": "bind",
	                "Source": "/lib/modules",
	                "Destination": "/lib/modules",
	                "Mode": "ro",
	                "RW": false,
	                "Propagation": "rprivate"
	            }
	        ],
	        "Config": {
	            "Hostname": "addons-315216",
	            "Domainname": "",
	            "User": "",
	            "AttachStdin": false,
	            "AttachStdout": false,
	            "AttachStderr": false,
	            "ExposedPorts": {
	                "22/tcp": {},
	                "2376/tcp": {},
	                "32443/tcp": {},
	                "5000/tcp": {},
	                "8443/tcp": {}
	            },
	            "Tty": true,
	            "OpenStdin": false,
	            "StdinOnce": false,
	            "Env": [
	                "container=docker",
	                "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
	            ],
	            "Cmd": null,
	            "Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e",
	            "Volumes": null,
	            "WorkingDir": "/",
	            "Entrypoint": [
	                "/usr/local/bin/entrypoint",
	                "/sbin/init"
	            ],
	            "OnBuild": null,
	            "Labels": {
	                "created_by.minikube.sigs.k8s.io": "true",
	                "mode.minikube.sigs.k8s.io": "addons-315216",
	                "name.minikube.sigs.k8s.io": "addons-315216",
	                "role.minikube.sigs.k8s.io": ""
	            },
	            "StopSignal": "SIGRTMIN+3"
	        },
	        "NetworkSettings": {
	            "Bridge": "",
	            "SandboxID": "9b76faff0cac9600db0f94d9c74357424a1defc891e3e0cf0df8516e8a4d997c",
	            "SandboxKey": "/var/run/docker/netns/9b76faff0cac",
	            "Ports": {
	                "22/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "32768"
	                    }
	                ],
	                "2376/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "32769"
	                    }
	                ],
	                "32443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "32772"
	                    }
	                ],
	                "5000/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "32770"
	                    }
	                ],
	                "8443/tcp": [
	                    {
	                        "HostIp": "127.0.0.1",
	                        "HostPort": "32771"
	                    }
	                ]
	            },
	            "HairpinMode": false,
	            "LinkLocalIPv6Address": "",
	            "LinkLocalIPv6PrefixLen": 0,
	            "SecondaryIPAddresses": null,
	            "SecondaryIPv6Addresses": null,
	            "EndpointID": "",
	            "Gateway": "",
	            "GlobalIPv6Address": "",
	            "GlobalIPv6PrefixLen": 0,
	            "IPAddress": "",
	            "IPPrefixLen": 0,
	            "IPv6Gateway": "",
	            "MacAddress": "",
	            "Networks": {
	                "addons-315216": {
	                    "IPAMConfig": {
	                        "IPv4Address": "192.168.49.2"
	                    },
	                    "Links": null,
	                    "Aliases": null,
	                    "MacAddress": "02:42:c0:a8:31:02",
	                    "DriverOpts": null,
	                    "NetworkID": "7ae42d435adaf1ee1f67d007ea424a7745532b78bcb0da4cb69d5ac692087e5b",
	                    "EndpointID": "c3b6f36f89bbdedea52a50d42fc7edd52ba979da79a15f30e0c6026678d21542",
	                    "Gateway": "192.168.49.1",
	                    "IPAddress": "192.168.49.2",
	                    "IPPrefixLen": 24,
	                    "IPv6Gateway": "",
	                    "GlobalIPv6Address": "",
	                    "GlobalIPv6PrefixLen": 0,
	                    "DNSNames": [
	                        "addons-315216",
	                        "597eb6e49b79"
	                    ]
	                }
	            }
	        }
	    }
	]

                                                
                                                
-- /stdout --
helpers_test.go:239: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Host}} -p addons-315216 -n addons-315216
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======>  post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 logs -n 25
helpers_test.go:252: TestAddons/parallel/Registry logs: 
-- stdout --
	
	==> Audit <==
	|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
	| Command |                                            Args                                             |        Profile         |  User   | Version |     Start Time      |      End Time       |
	|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
	| start   | --download-only -p                                                                          | download-docker-782708 | jenkins | v1.34.0 | 13 Sep 24 18:20 UTC |                     |
	|         | download-docker-782708                                                                      |                        |         |         |                     |                     |
	|         | --alsologtostderr                                                                           |                        |         |         |                     |                     |
	|         | --driver=docker                                                                             |                        |         |         |                     |                     |
	|         | --container-runtime=docker                                                                  |                        |         |         |                     |                     |
	| delete  | -p download-docker-782708                                                                   | download-docker-782708 | jenkins | v1.34.0 | 13 Sep 24 18:20 UTC | 13 Sep 24 18:20 UTC |
	| start   | --download-only -p                                                                          | binary-mirror-486689   | jenkins | v1.34.0 | 13 Sep 24 18:20 UTC |                     |
	|         | binary-mirror-486689                                                                        |                        |         |         |                     |                     |
	|         | --alsologtostderr                                                                           |                        |         |         |                     |                     |
	|         | --binary-mirror                                                                             |                        |         |         |                     |                     |
	|         | http://127.0.0.1:41715                                                                      |                        |         |         |                     |                     |
	|         | --driver=docker                                                                             |                        |         |         |                     |                     |
	|         | --container-runtime=docker                                                                  |                        |         |         |                     |                     |
	| delete  | -p binary-mirror-486689                                                                     | binary-mirror-486689   | jenkins | v1.34.0 | 13 Sep 24 18:20 UTC | 13 Sep 24 18:20 UTC |
	| addons  | enable dashboard -p                                                                         | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:20 UTC |                     |
	|         | addons-315216                                                                               |                        |         |         |                     |                     |
	| addons  | disable dashboard -p                                                                        | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:20 UTC |                     |
	|         | addons-315216                                                                               |                        |         |         |                     |                     |
	| start   | -p addons-315216 --wait=true                                                                | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:20 UTC | 13 Sep 24 18:24 UTC |
	|         | --memory=4000 --alsologtostderr                                                             |                        |         |         |                     |                     |
	|         | --addons=registry                                                                           |                        |         |         |                     |                     |
	|         | --addons=metrics-server                                                                     |                        |         |         |                     |                     |
	|         | --addons=volumesnapshots                                                                    |                        |         |         |                     |                     |
	|         | --addons=csi-hostpath-driver                                                                |                        |         |         |                     |                     |
	|         | --addons=gcp-auth                                                                           |                        |         |         |                     |                     |
	|         | --addons=cloud-spanner                                                                      |                        |         |         |                     |                     |
	|         | --addons=inspektor-gadget                                                                   |                        |         |         |                     |                     |
	|         | --addons=storage-provisioner-rancher                                                        |                        |         |         |                     |                     |
	|         | --addons=nvidia-device-plugin                                                               |                        |         |         |                     |                     |
	|         | --addons=yakd --addons=volcano                                                              |                        |         |         |                     |                     |
	|         | --driver=docker                                                                             |                        |         |         |                     |                     |
	|         | --container-runtime=docker                                                                  |                        |         |         |                     |                     |
	|         | --addons=ingress                                                                            |                        |         |         |                     |                     |
	|         | --addons=ingress-dns                                                                        |                        |         |         |                     |                     |
	| addons  | addons-315216 addons disable                                                                | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:24 UTC | 13 Sep 24 18:24 UTC |
	|         | volcano --alsologtostderr -v=1                                                              |                        |         |         |                     |                     |
	| addons  | enable headlamp                                                                             | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:32 UTC | 13 Sep 24 18:32 UTC |
	|         | -p addons-315216                                                                            |                        |         |         |                     |                     |
	|         | --alsologtostderr -v=1                                                                      |                        |         |         |                     |                     |
	| addons  | disable inspektor-gadget -p                                                                 | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:32 UTC | 13 Sep 24 18:32 UTC |
	|         | addons-315216                                                                               |                        |         |         |                     |                     |
	| addons  | addons-315216 addons                                                                        | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:32 UTC | 13 Sep 24 18:32 UTC |
	|         | disable metrics-server                                                                      |                        |         |         |                     |                     |
	|         | --alsologtostderr -v=1                                                                      |                        |         |         |                     |                     |
	| addons  | addons-315216 addons disable                                                                | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:32 UTC | 13 Sep 24 18:33 UTC |
	|         | headlamp --alsologtostderr                                                                  |                        |         |         |                     |                     |
	|         | -v=1                                                                                        |                        |         |         |                     |                     |
	| addons  | disable nvidia-device-plugin                                                                | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:33 UTC | 13 Sep 24 18:33 UTC |
	|         | -p addons-315216                                                                            |                        |         |         |                     |                     |
	| addons  | addons-315216 addons disable                                                                | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:33 UTC | 13 Sep 24 18:33 UTC |
	|         | yakd --alsologtostderr -v=1                                                                 |                        |         |         |                     |                     |
	| ssh     | addons-315216 ssh curl -s                                                                   | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:33 UTC | 13 Sep 24 18:33 UTC |
	|         | http://127.0.0.1/ -H 'Host:                                                                 |                        |         |         |                     |                     |
	|         | nginx.example.com'                                                                          |                        |         |         |                     |                     |
	| ip      | addons-315216 ip                                                                            | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:33 UTC | 13 Sep 24 18:33 UTC |
	| addons  | addons-315216 addons disable                                                                | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:33 UTC | 13 Sep 24 18:33 UTC |
	|         | ingress-dns --alsologtostderr                                                               |                        |         |         |                     |                     |
	|         | -v=1                                                                                        |                        |         |         |                     |                     |
	| addons  | addons-315216 addons disable                                                                | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:33 UTC | 13 Sep 24 18:33 UTC |
	|         | ingress --alsologtostderr -v=1                                                              |                        |         |         |                     |                     |
	| ssh     | addons-315216 ssh cat                                                                       | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:33 UTC | 13 Sep 24 18:33 UTC |
	|         | /opt/local-path-provisioner/pvc-2743739b-b850-4f4a-8324-d1e883dc17e5_default_test-pvc/file1 |                        |         |         |                     |                     |
	| addons  | addons-315216 addons disable                                                                | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:33 UTC |                     |
	|         | storage-provisioner-rancher                                                                 |                        |         |         |                     |                     |
	|         | --alsologtostderr -v=1                                                                      |                        |         |         |                     |                     |
	| addons  | disable cloud-spanner -p                                                                    | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:33 UTC | 13 Sep 24 18:33 UTC |
	|         | addons-315216                                                                               |                        |         |         |                     |                     |
	| addons  | addons-315216 addons                                                                        | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:33 UTC | 13 Sep 24 18:33 UTC |
	|         | disable csi-hostpath-driver                                                                 |                        |         |         |                     |                     |
	|         | --alsologtostderr -v=1                                                                      |                        |         |         |                     |                     |
	| addons  | addons-315216 addons                                                                        | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:33 UTC | 13 Sep 24 18:33 UTC |
	|         | disable volumesnapshots                                                                     |                        |         |         |                     |                     |
	|         | --alsologtostderr -v=1                                                                      |                        |         |         |                     |                     |
	| ip      | addons-315216 ip                                                                            | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:33 UTC | 13 Sep 24 18:33 UTC |
	| addons  | addons-315216 addons disable                                                                | addons-315216          | jenkins | v1.34.0 | 13 Sep 24 18:33 UTC | 13 Sep 24 18:33 UTC |
	|         | registry --alsologtostderr                                                                  |                        |         |         |                     |                     |
	|         | -v=1                                                                                        |                        |         |         |                     |                     |
	|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
	
	
	==> Last Start <==
	Log file created at: 2024/09/13 18:20:37
	Running on machine: ubuntu-20-agent-14
	Binary: Built with gc go1.23.0 for linux/amd64
	Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
	I0913 18:20:37.622079   11774 out.go:345] Setting OutFile to fd 1 ...
	I0913 18:20:37.622164   11774 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:20:37.622171   11774 out.go:358] Setting ErrFile to fd 2...
	I0913 18:20:37.622175   11774 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:20:37.622329   11774 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
	I0913 18:20:37.622838   11774 out.go:352] Setting JSON to false
	I0913 18:20:37.623599   11774 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-14","uptime":189,"bootTime":1726251449,"procs":176,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1068-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
	I0913 18:20:37.623685   11774 start.go:139] virtualization: kvm guest
	I0913 18:20:37.625620   11774 out.go:177] * [addons-315216] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
	I0913 18:20:37.626906   11774 notify.go:220] Checking for updates...
	I0913 18:20:37.626912   11774 out.go:177]   - MINIKUBE_LOCATION=19636
	I0913 18:20:37.628079   11774 out.go:177]   - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0913 18:20:37.629591   11774 out.go:177]   - KUBECONFIG=/home/jenkins/minikube-integration/19636-3724/kubeconfig
	I0913 18:20:37.630753   11774 out.go:177]   - MINIKUBE_HOME=/home/jenkins/minikube-integration/19636-3724/.minikube
	I0913 18:20:37.631775   11774 out.go:177]   - MINIKUBE_BIN=out/minikube-linux-amd64
	I0913 18:20:37.632836   11774 out.go:177]   - MINIKUBE_FORCE_SYSTEMD=
	I0913 18:20:37.633886   11774 driver.go:394] Setting default libvirt URI to qemu:///system
	I0913 18:20:37.653342   11774 docker.go:123] docker version: linux-27.2.1:Docker Engine - Community
	I0913 18:20:37.653406   11774 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0913 18:20:37.699455   11774 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:30 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-13 18:20:37.689365898 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
	I0913 18:20:37.699553   11774 docker.go:318] overlay module found
	I0913 18:20:37.701342   11774 out.go:177] * Using the docker driver based on user configuration
	I0913 18:20:37.702551   11774 start.go:297] selected driver: docker
	I0913 18:20:37.702562   11774 start.go:901] validating driver "docker" against <nil>
	I0913 18:20:37.702571   11774 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
	I0913 18:20:37.703389   11774 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0913 18:20:37.746659   11774 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-13 18:20:37.738713127 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
	I0913 18:20:37.746804   11774 start_flags.go:310] no existing cluster config was found, will generate one from the flags 
	I0913 18:20:37.747024   11774 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
	I0913 18:20:37.748612   11774 out.go:177] * Using Docker driver with root privileges
	I0913 18:20:37.749809   11774 cni.go:84] Creating CNI manager for ""
	I0913 18:20:37.749874   11774 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
	I0913 18:20:37.749885   11774 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
	I0913 18:20:37.749950   11774 start.go:340] cluster config:
	{Name:addons-315216 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-315216 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
	I0913 18:20:37.751216   11774 out.go:177] * Starting "addons-315216" primary control-plane node in "addons-315216" cluster
	I0913 18:20:37.752214   11774 cache.go:121] Beginning downloading kic base image for docker with docker
	I0913 18:20:37.753232   11774 out.go:177] * Pulling base image v0.0.45-1726193793-19634 ...
	I0913 18:20:37.754137   11774 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
	I0913 18:20:37.754169   11774 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19636-3724/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4
	I0913 18:20:37.754181   11774 cache.go:56] Caching tarball of preloaded images
	I0913 18:20:37.754219   11774 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e in local docker daemon
	I0913 18:20:37.754271   11774 preload.go:172] Found /home/jenkins/minikube-integration/19636-3724/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4 in cache, skipping download
	I0913 18:20:37.754284   11774 cache.go:59] Finished verifying existence of preloaded tar for v1.31.1 on docker
	I0913 18:20:37.754634   11774 profile.go:143] Saving config to /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/config.json ...
	I0913 18:20:37.754654   11774 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/config.json: {Name:mkac6589fc12791c34c9067d069a69a8559b9e44 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:20:37.768267   11774 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e to local cache
	I0913 18:20:37.768342   11774 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e in local cache directory
	I0913 18:20:37.768355   11774 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e in local cache directory, skipping pull
	I0913 18:20:37.768359   11774 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e exists in cache, skipping pull
	I0913 18:20:37.768368   11774 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e as a tarball
	I0913 18:20:37.768375   11774 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e from local cache
	I0913 18:20:49.395078   11774 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e from cached tarball
	I0913 18:20:49.395109   11774 cache.go:194] Successfully downloaded all kic artifacts
	I0913 18:20:49.395147   11774 start.go:360] acquireMachinesLock for addons-315216: {Name:mkd469c6369f105e9f5fe2a932062bc589aa74d2 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
	I0913 18:20:49.395237   11774 start.go:364] duration metric: took 68.605µs to acquireMachinesLock for "addons-315216"
	I0913 18:20:49.395263   11774 start.go:93] Provisioning new machine with config: &{Name:addons-315216 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-315216 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
	I0913 18:20:49.395350   11774 start.go:125] createHost starting for "" (driver="docker")
	I0913 18:20:49.397777   11774 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
	I0913 18:20:49.398017   11774 start.go:159] libmachine.API.Create for "addons-315216" (driver="docker")
	I0913 18:20:49.398050   11774 client.go:168] LocalClient.Create starting
	I0913 18:20:49.398152   11774 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19636-3724/.minikube/certs/ca.pem
	I0913 18:20:49.631678   11774 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19636-3724/.minikube/certs/cert.pem
	I0913 18:20:49.831979   11774 cli_runner.go:164] Run: docker network inspect addons-315216 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	W0913 18:20:49.846976   11774 cli_runner.go:211] docker network inspect addons-315216 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
	I0913 18:20:49.847056   11774 network_create.go:284] running [docker network inspect addons-315216] to gather additional debugging logs...
	I0913 18:20:49.847073   11774 cli_runner.go:164] Run: docker network inspect addons-315216
	W0913 18:20:49.862020   11774 cli_runner.go:211] docker network inspect addons-315216 returned with exit code 1
	I0913 18:20:49.862050   11774 network_create.go:287] error running [docker network inspect addons-315216]: docker network inspect addons-315216: exit status 1
	stdout:
	[]
	
	stderr:
	Error response from daemon: network addons-315216 not found
	I0913 18:20:49.862062   11774 network_create.go:289] output of [docker network inspect addons-315216]: -- stdout --
	[]
	
	-- /stdout --
	** stderr ** 
	Error response from daemon: network addons-315216 not found
	
	** /stderr **
	I0913 18:20:49.862133   11774 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	I0913 18:20:49.876956   11774 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc00191a910}
	I0913 18:20:49.876999   11774 network_create.go:124] attempt to create docker network addons-315216 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
	I0913 18:20:49.877037   11774 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-315216 addons-315216
	I0913 18:20:49.931960   11774 network_create.go:108] docker network addons-315216 192.168.49.0/24 created
	I0913 18:20:49.931990   11774 kic.go:121] calculated static IP "192.168.49.2" for the "addons-315216" container
	I0913 18:20:49.932042   11774 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
	I0913 18:20:49.946536   11774 cli_runner.go:164] Run: docker volume create addons-315216 --label name.minikube.sigs.k8s.io=addons-315216 --label created_by.minikube.sigs.k8s.io=true
	I0913 18:20:49.962350   11774 oci.go:103] Successfully created a docker volume addons-315216
	I0913 18:20:49.962437   11774 cli_runner.go:164] Run: docker run --rm --name addons-315216-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-315216 --entrypoint /usr/bin/test -v addons-315216:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e -d /var/lib
	I0913 18:20:57.101515   11774 cli_runner.go:217] Completed: docker run --rm --name addons-315216-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-315216 --entrypoint /usr/bin/test -v addons-315216:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e -d /var/lib: (7.139033373s)
	I0913 18:20:57.101549   11774 oci.go:107] Successfully prepared a docker volume addons-315216
	I0913 18:20:57.101565   11774 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
	I0913 18:20:57.101584   11774 kic.go:194] Starting extracting preloaded images to volume ...
	I0913 18:20:57.101636   11774 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19636-3724/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-315216:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e -I lz4 -xf /preloaded.tar -C /extractDir
	I0913 18:21:00.887434   11774 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19636-3724/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-315216:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e -I lz4 -xf /preloaded.tar -C /extractDir: (3.785767467s)
	I0913 18:21:00.887459   11774 kic.go:203] duration metric: took 3.78587336s to extract preloaded images to volume ...
	W0913 18:21:00.887610   11774 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
	I0913 18:21:00.887733   11774 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
	I0913 18:21:00.934295   11774 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-315216 --name addons-315216 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-315216 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-315216 --network addons-315216 --ip 192.168.49.2 --volume addons-315216:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e
	I0913 18:21:01.251847   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Running}}
	I0913 18:21:01.269216   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:01.287109   11774 cli_runner.go:164] Run: docker exec addons-315216 stat /var/lib/dpkg/alternatives/iptables
	I0913 18:21:01.328940   11774 oci.go:144] the created container "addons-315216" has a running status.
	I0913 18:21:01.328975   11774 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa...
	I0913 18:21:01.503353   11774 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
	I0913 18:21:01.526219   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:01.543637   11774 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
	I0913 18:21:01.543662   11774 kic_runner.go:114] Args: [docker exec --privileged addons-315216 chown docker:docker /home/docker/.ssh/authorized_keys]
	I0913 18:21:01.608418   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:01.628565   11774 machine.go:93] provisionDockerMachine start ...
	I0913 18:21:01.628640   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:01.647681   11774 main.go:141] libmachine: Using SSH client type: native
	I0913 18:21:01.647862   11774 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil>  [] 0s} 127.0.0.1 32768 <nil> <nil>}
	I0913 18:21:01.647875   11774 main.go:141] libmachine: About to run SSH command:
	hostname
	I0913 18:21:01.840545   11774 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-315216
	
	I0913 18:21:01.840583   11774 ubuntu.go:169] provisioning hostname "addons-315216"
	I0913 18:21:01.840645   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:01.857070   11774 main.go:141] libmachine: Using SSH client type: native
	I0913 18:21:01.857244   11774 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil>  [] 0s} 127.0.0.1 32768 <nil> <nil>}
	I0913 18:21:01.857269   11774 main.go:141] libmachine: About to run SSH command:
	sudo hostname addons-315216 && echo "addons-315216" | sudo tee /etc/hostname
	I0913 18:21:01.990975   11774 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-315216
	
	I0913 18:21:01.991061   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:02.006511   11774 main.go:141] libmachine: Using SSH client type: native
	I0913 18:21:02.006691   11774 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil>  [] 0s} 127.0.0.1 32768 <nil> <nil>}
	I0913 18:21:02.006707   11774 main.go:141] libmachine: About to run SSH command:
	
			if ! grep -xq '.*\saddons-315216' /etc/hosts; then
				if grep -xq '127.0.1.1\s.*' /etc/hosts; then
					sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-315216/g' /etc/hosts;
				else 
					echo '127.0.1.1 addons-315216' | sudo tee -a /etc/hosts; 
				fi
			fi
	I0913 18:21:02.125286   11774 main.go:141] libmachine: SSH cmd err, output: <nil>: 
	I0913 18:21:02.125312   11774 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19636-3724/.minikube CaCertPath:/home/jenkins/minikube-integration/19636-3724/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19636-3724/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19636-3724/.minikube}
	I0913 18:21:02.125353   11774 ubuntu.go:177] setting up certificates
	I0913 18:21:02.125365   11774 provision.go:84] configureAuth start
	I0913 18:21:02.125416   11774 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-315216
	I0913 18:21:02.140622   11774 provision.go:143] copyHostCerts
	I0913 18:21:02.140687   11774 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19636-3724/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19636-3724/.minikube/ca.pem (1078 bytes)
	I0913 18:21:02.140811   11774 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19636-3724/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19636-3724/.minikube/cert.pem (1123 bytes)
	I0913 18:21:02.140888   11774 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19636-3724/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19636-3724/.minikube/key.pem (1679 bytes)
	I0913 18:21:02.140948   11774 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19636-3724/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19636-3724/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19636-3724/.minikube/certs/ca-key.pem org=jenkins.addons-315216 san=[127.0.0.1 192.168.49.2 addons-315216 localhost minikube]
	I0913 18:21:02.426237   11774 provision.go:177] copyRemoteCerts
	I0913 18:21:02.426293   11774 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
	I0913 18:21:02.426326   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:02.441726   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:02.529150   11774 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19636-3724/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
	I0913 18:21:02.549166   11774 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19636-3724/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
	I0913 18:21:02.568923   11774 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19636-3724/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
	I0913 18:21:02.588387   11774 provision.go:87] duration metric: took 463.009888ms to configureAuth
	I0913 18:21:02.588412   11774 ubuntu.go:193] setting minikube options for container-runtime
	I0913 18:21:02.588567   11774 config.go:182] Loaded profile config "addons-315216": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
	I0913 18:21:02.588619   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:02.603648   11774 main.go:141] libmachine: Using SSH client type: native
	I0913 18:21:02.603815   11774 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil>  [] 0s} 127.0.0.1 32768 <nil> <nil>}
	I0913 18:21:02.603827   11774 main.go:141] libmachine: About to run SSH command:
	df --output=fstype / | tail -n 1
	I0913 18:21:02.721394   11774 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
	
	I0913 18:21:02.721420   11774 ubuntu.go:71] root file system type: overlay
	I0913 18:21:02.721534   11774 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
	I0913 18:21:02.721596   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:02.736743   11774 main.go:141] libmachine: Using SSH client type: native
	I0913 18:21:02.736905   11774 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil>  [] 0s} 127.0.0.1 32768 <nil> <nil>}
	I0913 18:21:02.736961   11774 main.go:141] libmachine: About to run SSH command:
	sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
	Description=Docker Application Container Engine
	Documentation=https://docs.docker.com
	BindsTo=containerd.service
	After=network-online.target firewalld.service containerd.service
	Wants=network-online.target
	Requires=docker.socket
	StartLimitBurst=3
	StartLimitIntervalSec=60
	
	[Service]
	Type=notify
	Restart=on-failure
	
	
	
	# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
	# The base configuration already specifies an 'ExecStart=...' command. The first directive
	# here is to clear out that command inherited from the base configuration. Without this,
	# the command from the base configuration and the command specified here are treated as
	# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
	# will catch this invalid input and refuse to start the service with an error like:
	#  Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
	
	# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
	# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
	ExecStart=
	ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12 
	ExecReload=/bin/kill -s HUP \$MAINPID
	
	# Having non-zero Limit*s causes performance problems due to accounting overhead
	# in the kernel. We recommend using cgroups to do container-local accounting.
	LimitNOFILE=infinity
	LimitNPROC=infinity
	LimitCORE=infinity
	
	# Uncomment TasksMax if your systemd version supports it.
	# Only systemd 226 and above support this version.
	TasksMax=infinity
	TimeoutStartSec=0
	
	# set delegate yes so that systemd does not reset the cgroups of docker containers
	Delegate=yes
	
	# kill only the docker process, not all processes in the cgroup
	KillMode=process
	
	[Install]
	WantedBy=multi-user.target
	" | sudo tee /lib/systemd/system/docker.service.new
	I0913 18:21:02.866701   11774 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
	Description=Docker Application Container Engine
	Documentation=https://docs.docker.com
	BindsTo=containerd.service
	After=network-online.target firewalld.service containerd.service
	Wants=network-online.target
	Requires=docker.socket
	StartLimitBurst=3
	StartLimitIntervalSec=60
	
	[Service]
	Type=notify
	Restart=on-failure
	
	
	
	# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
	# The base configuration already specifies an 'ExecStart=...' command. The first directive
	# here is to clear out that command inherited from the base configuration. Without this,
	# the command from the base configuration and the command specified here are treated as
	# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
	# will catch this invalid input and refuse to start the service with an error like:
	#  Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
	
	# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
	# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
	ExecStart=
	ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12 
	ExecReload=/bin/kill -s HUP $MAINPID
	
	# Having non-zero Limit*s causes performance problems due to accounting overhead
	# in the kernel. We recommend using cgroups to do container-local accounting.
	LimitNOFILE=infinity
	LimitNPROC=infinity
	LimitCORE=infinity
	
	# Uncomment TasksMax if your systemd version supports it.
	# Only systemd 226 and above support this version.
	TasksMax=infinity
	TimeoutStartSec=0
	
	# set delegate yes so that systemd does not reset the cgroups of docker containers
	Delegate=yes
	
	# kill only the docker process, not all processes in the cgroup
	KillMode=process
	
	[Install]
	WantedBy=multi-user.target
	
	I0913 18:21:02.866764   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:02.882223   11774 main.go:141] libmachine: Using SSH client type: native
	I0913 18:21:02.882432   11774 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil>  [] 0s} 127.0.0.1 32768 <nil> <nil>}
	I0913 18:21:02.882458   11774 main.go:141] libmachine: About to run SSH command:
	sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
	I0913 18:21:03.551281   11774 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service	2024-09-06 12:06:41.000000000 +0000
	+++ /lib/systemd/system/docker.service.new	2024-09-13 18:21:02.859175682 +0000
	@@ -1,46 +1,49 @@
	 [Unit]
	 Description=Docker Application Container Engine
	 Documentation=https://docs.docker.com
	-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
	-Wants=network-online.target containerd.service
	+BindsTo=containerd.service
	+After=network-online.target firewalld.service containerd.service
	+Wants=network-online.target
	 Requires=docker.socket
	+StartLimitBurst=3
	+StartLimitIntervalSec=60
	 
	 [Service]
	 Type=notify
	-# the default is not to use systemd for cgroups because the delegate issues still
	-# exists and systemd currently does not support the cgroup feature set required
	-# for containers run by docker
	-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
	-ExecReload=/bin/kill -s HUP $MAINPID
	-TimeoutStartSec=0
	-RestartSec=2
	-Restart=always
	+Restart=on-failure
	 
	-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
	-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
	-# to make them work for either version of systemd.
	-StartLimitBurst=3
	 
	-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
	-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
	-# this option work for either version of systemd.
	-StartLimitInterval=60s
	+
	+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
	+# The base configuration already specifies an 'ExecStart=...' command. The first directive
	+# here is to clear out that command inherited from the base configuration. Without this,
	+# the command from the base configuration and the command specified here are treated as
	+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
	+# will catch this invalid input and refuse to start the service with an error like:
	+#  Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
	+
	+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
	+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
	+ExecStart=
	+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12 
	+ExecReload=/bin/kill -s HUP $MAINPID
	 
	 # Having non-zero Limit*s causes performance problems due to accounting overhead
	 # in the kernel. We recommend using cgroups to do container-local accounting.
	+LimitNOFILE=infinity
	 LimitNPROC=infinity
	 LimitCORE=infinity
	 
	-# Comment TasksMax if your systemd version does not support it.
	-# Only systemd 226 and above support this option.
	+# Uncomment TasksMax if your systemd version supports it.
	+# Only systemd 226 and above support this version.
	 TasksMax=infinity
	+TimeoutStartSec=0
	 
	 # set delegate yes so that systemd does not reset the cgroups of docker containers
	 Delegate=yes
	 
	 # kill only the docker process, not all processes in the cgroup
	 KillMode=process
	-OOMScoreAdjust=-500
	 
	 [Install]
	 WantedBy=multi-user.target
	Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
	Executing: /lib/systemd/systemd-sysv-install enable docker
	
	I0913 18:21:03.551322   11774 machine.go:96] duration metric: took 1.922738555s to provisionDockerMachine
	I0913 18:21:03.551336   11774 client.go:171] duration metric: took 14.153274624s to LocalClient.Create
	I0913 18:21:03.551354   11774 start.go:167] duration metric: took 14.153337196s to libmachine.API.Create "addons-315216"
	I0913 18:21:03.551365   11774 start.go:293] postStartSetup for "addons-315216" (driver="docker")
	I0913 18:21:03.551375   11774 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
	I0913 18:21:03.551429   11774 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
	I0913 18:21:03.551470   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:03.567684   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:03.653633   11774 ssh_runner.go:195] Run: cat /etc/os-release
	I0913 18:21:03.656359   11774 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
	I0913 18:21:03.656388   11774 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
	I0913 18:21:03.656395   11774 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
	I0913 18:21:03.656401   11774 info.go:137] Remote host: Ubuntu 22.04.4 LTS
	I0913 18:21:03.656409   11774 filesync.go:126] Scanning /home/jenkins/minikube-integration/19636-3724/.minikube/addons for local assets ...
	I0913 18:21:03.656465   11774 filesync.go:126] Scanning /home/jenkins/minikube-integration/19636-3724/.minikube/files for local assets ...
	I0913 18:21:03.656496   11774 start.go:296] duration metric: took 105.124671ms for postStartSetup
	I0913 18:21:03.656803   11774 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-315216
	I0913 18:21:03.672056   11774 profile.go:143] Saving config to /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/config.json ...
	I0913 18:21:03.672302   11774 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0913 18:21:03.672349   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:03.688467   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:03.773937   11774 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
	I0913 18:21:03.777821   11774 start.go:128] duration metric: took 14.38245778s to createHost
	I0913 18:21:03.777850   11774 start.go:83] releasing machines lock for "addons-315216", held for 14.382600069s
	I0913 18:21:03.777902   11774 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-315216
	I0913 18:21:03.793501   11774 ssh_runner.go:195] Run: cat /version.json
	I0913 18:21:03.793550   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:03.793558   11774 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
	I0913 18:21:03.793605   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:03.811796   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:03.812062   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:03.962985   11774 ssh_runner.go:195] Run: systemctl --version
	I0913 18:21:03.966714   11774 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
	I0913 18:21:03.970270   11774 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
	I0913 18:21:03.990577   11774 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
	I0913 18:21:03.990634   11774 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
	I0913 18:21:04.013991   11774 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
	I0913 18:21:04.014015   11774 start.go:495] detecting cgroup driver to use...
	I0913 18:21:04.014042   11774 detect.go:187] detected "cgroupfs" cgroup driver on host os
	I0913 18:21:04.014129   11774 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
	" | sudo tee /etc/crictl.yaml"
	I0913 18:21:04.027539   11774 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
	I0913 18:21:04.035545   11774 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
	I0913 18:21:04.043618   11774 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
	I0913 18:21:04.043672   11774 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
	I0913 18:21:04.051741   11774 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
	I0913 18:21:04.059445   11774 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
	I0913 18:21:04.067342   11774 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
	I0913 18:21:04.075352   11774 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
	I0913 18:21:04.082813   11774 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
	I0913 18:21:04.090742   11774 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
	I0913 18:21:04.098665   11774 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1  enable_unprivileged_ports = true|' /etc/containerd/config.toml"
	I0913 18:21:04.106587   11774 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
	I0913 18:21:04.113303   11774 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
	I0913 18:21:04.120092   11774 ssh_runner.go:195] Run: sudo systemctl daemon-reload
	I0913 18:21:04.196082   11774 ssh_runner.go:195] Run: sudo systemctl restart containerd
	I0913 18:21:04.280699   11774 start.go:495] detecting cgroup driver to use...
	I0913 18:21:04.280747   11774 detect.go:187] detected "cgroupfs" cgroup driver on host os
	I0913 18:21:04.280786   11774 ssh_runner.go:195] Run: sudo systemctl cat docker.service
	I0913 18:21:04.290789   11774 cruntime.go:279] skipping containerd shutdown because we are bound to it
	I0913 18:21:04.290836   11774 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
	I0913 18:21:04.301202   11774 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
	" | sudo tee /etc/crictl.yaml"
	I0913 18:21:04.314644   11774 ssh_runner.go:195] Run: which cri-dockerd
	I0913 18:21:04.317659   11774 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
	I0913 18:21:04.326208   11774 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
	I0913 18:21:04.341135   11774 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
	I0913 18:21:04.432349   11774 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
	I0913 18:21:04.526370   11774 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
	I0913 18:21:04.526544   11774 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
	I0913 18:21:04.541742   11774 ssh_runner.go:195] Run: sudo systemctl daemon-reload
	I0913 18:21:04.613180   11774 ssh_runner.go:195] Run: sudo systemctl restart docker
	I0913 18:21:04.848601   11774 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
	I0913 18:21:04.858900   11774 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
	I0913 18:21:04.868919   11774 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
	I0913 18:21:04.949098   11774 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
	I0913 18:21:05.020308   11774 ssh_runner.go:195] Run: sudo systemctl daemon-reload
	I0913 18:21:05.097993   11774 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
	I0913 18:21:05.109101   11774 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
	I0913 18:21:05.118172   11774 ssh_runner.go:195] Run: sudo systemctl daemon-reload
	I0913 18:21:05.189195   11774 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
	I0913 18:21:05.242648   11774 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
	I0913 18:21:05.242734   11774 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
	I0913 18:21:05.246025   11774 start.go:563] Will wait 60s for crictl version
	I0913 18:21:05.246072   11774 ssh_runner.go:195] Run: which crictl
	I0913 18:21:05.248863   11774 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
	I0913 18:21:05.277695   11774 start.go:579] Version:  0.1.0
	RuntimeName:  docker
	RuntimeVersion:  27.2.1
	RuntimeApiVersion:  v1
	I0913 18:21:05.277757   11774 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
	I0913 18:21:05.298177   11774 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
	I0913 18:21:05.320914   11774 out.go:235] * Preparing Kubernetes v1.31.1 on Docker 27.2.1 ...
	I0913 18:21:05.320991   11774 cli_runner.go:164] Run: docker network inspect addons-315216 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
	I0913 18:21:05.335847   11774 ssh_runner.go:195] Run: grep 192.168.49.1	host.minikube.internal$ /etc/hosts
	I0913 18:21:05.338863   11774 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1	host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
	I0913 18:21:05.348005   11774 kubeadm.go:883] updating cluster {Name:addons-315216 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-315216 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
	I0913 18:21:05.348133   11774 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
	I0913 18:21:05.348190   11774 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
	I0913 18:21:05.364878   11774 docker.go:685] Got preloaded images: -- stdout --
	registry.k8s.io/kube-apiserver:v1.31.1
	registry.k8s.io/kube-controller-manager:v1.31.1
	registry.k8s.io/kube-scheduler:v1.31.1
	registry.k8s.io/kube-proxy:v1.31.1
	registry.k8s.io/coredns/coredns:v1.11.3
	registry.k8s.io/etcd:3.5.15-0
	registry.k8s.io/pause:3.10
	gcr.io/k8s-minikube/storage-provisioner:v5
	
	-- /stdout --
	I0913 18:21:05.364895   11774 docker.go:615] Images already preloaded, skipping extraction
	I0913 18:21:05.364945   11774 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
	I0913 18:21:05.380835   11774 docker.go:685] Got preloaded images: -- stdout --
	registry.k8s.io/kube-apiserver:v1.31.1
	registry.k8s.io/kube-scheduler:v1.31.1
	registry.k8s.io/kube-controller-manager:v1.31.1
	registry.k8s.io/kube-proxy:v1.31.1
	registry.k8s.io/coredns/coredns:v1.11.3
	registry.k8s.io/etcd:3.5.15-0
	registry.k8s.io/pause:3.10
	gcr.io/k8s-minikube/storage-provisioner:v5
	
	-- /stdout --
	I0913 18:21:05.380860   11774 cache_images.go:84] Images are preloaded, skipping loading
	I0913 18:21:05.380871   11774 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.1 docker true true} ...
	I0913 18:21:05.380961   11774 kubeadm.go:946] kubelet [Unit]
	Wants=docker.socket
	
	[Service]
	ExecStart=
	ExecStart=/var/lib/minikube/binaries/v1.31.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-315216 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
	
	[Install]
	 config:
	{KubernetesVersion:v1.31.1 ClusterName:addons-315216 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
	I0913 18:21:05.381004   11774 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
	I0913 18:21:05.421844   11774 cni.go:84] Creating CNI manager for ""
	I0913 18:21:05.421870   11774 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
	I0913 18:21:05.421879   11774 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
	I0913 18:21:05.421896   11774 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-315216 NodeName:addons-315216 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
	I0913 18:21:05.422026   11774 kubeadm.go:187] kubeadm config:
	apiVersion: kubeadm.k8s.io/v1beta3
	kind: InitConfiguration
	localAPIEndpoint:
	  advertiseAddress: 192.168.49.2
	  bindPort: 8443
	bootstrapTokens:
	  - groups:
	      - system:bootstrappers:kubeadm:default-node-token
	    ttl: 24h0m0s
	    usages:
	      - signing
	      - authentication
	nodeRegistration:
	  criSocket: unix:///var/run/cri-dockerd.sock
	  name: "addons-315216"
	  kubeletExtraArgs:
	    node-ip: 192.168.49.2
	  taints: []
	---
	apiVersion: kubeadm.k8s.io/v1beta3
	kind: ClusterConfiguration
	apiServer:
	  certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
	  extraArgs:
	    enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
	controllerManager:
	  extraArgs:
	    allocate-node-cidrs: "true"
	    leader-elect: "false"
	scheduler:
	  extraArgs:
	    leader-elect: "false"
	certificatesDir: /var/lib/minikube/certs
	clusterName: mk
	controlPlaneEndpoint: control-plane.minikube.internal:8443
	etcd:
	  local:
	    dataDir: /var/lib/minikube/etcd
	    extraArgs:
	      proxy-refresh-interval: "70000"
	kubernetesVersion: v1.31.1
	networking:
	  dnsDomain: cluster.local
	  podSubnet: "10.244.0.0/16"
	  serviceSubnet: 10.96.0.0/12
	---
	apiVersion: kubelet.config.k8s.io/v1beta1
	kind: KubeletConfiguration
	authentication:
	  x509:
	    clientCAFile: /var/lib/minikube/certs/ca.crt
	cgroupDriver: cgroupfs
	containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
	hairpinMode: hairpin-veth
	runtimeRequestTimeout: 15m
	clusterDomain: "cluster.local"
	# disable disk resource management by default
	imageGCHighThresholdPercent: 100
	evictionHard:
	  nodefs.available: "0%"
	  nodefs.inodesFree: "0%"
	  imagefs.available: "0%"
	failSwapOn: false
	staticPodPath: /etc/kubernetes/manifests
	---
	apiVersion: kubeproxy.config.k8s.io/v1alpha1
	kind: KubeProxyConfiguration
	clusterCIDR: "10.244.0.0/16"
	metricsBindAddress: 0.0.0.0:10249
	conntrack:
	  maxPerCore: 0
	# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
	  tcpEstablishedTimeout: 0s
	# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
	  tcpCloseWaitTimeout: 0s
	
	I0913 18:21:05.422075   11774 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.1
	I0913 18:21:05.430258   11774 binaries.go:44] Found k8s binaries, skipping transfer
	I0913 18:21:05.430332   11774 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
	I0913 18:21:05.437511   11774 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
	I0913 18:21:05.452089   11774 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
	I0913 18:21:05.466469   11774 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
	I0913 18:21:05.480733   11774 ssh_runner.go:195] Run: grep 192.168.49.2	control-plane.minikube.internal$ /etc/hosts
	I0913 18:21:05.483446   11774 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2	control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
	I0913 18:21:05.492057   11774 ssh_runner.go:195] Run: sudo systemctl daemon-reload
	I0913 18:21:05.562559   11774 ssh_runner.go:195] Run: sudo systemctl start kubelet
	I0913 18:21:05.574309   11774 certs.go:68] Setting up /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216 for IP: 192.168.49.2
	I0913 18:21:05.574329   11774 certs.go:194] generating shared ca certs ...
	I0913 18:21:05.574355   11774 certs.go:226] acquiring lock for ca certs: {Name:mk30906b62223b99152c4d67bec8517fa308d1fb Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:21:05.574458   11774 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19636-3724/.minikube/ca.key
	I0913 18:21:05.794726   11774 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19636-3724/.minikube/ca.crt ...
	I0913 18:21:05.794750   11774 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19636-3724/.minikube/ca.crt: {Name:mkc80074970e1242bfacc3c9f057072b02a9533e Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:21:05.794899   11774 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19636-3724/.minikube/ca.key ...
	I0913 18:21:05.794908   11774 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19636-3724/.minikube/ca.key: {Name:mk66ef0bf208c9f3e45eefb23304d19ebf6e1743 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:21:05.794975   11774 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19636-3724/.minikube/proxy-client-ca.key
	I0913 18:21:05.899577   11774 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19636-3724/.minikube/proxy-client-ca.crt ...
	I0913 18:21:05.899601   11774 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19636-3724/.minikube/proxy-client-ca.crt: {Name:mk3de351ff27390f75702fb762ff7275b36b8083 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:21:05.899732   11774 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19636-3724/.minikube/proxy-client-ca.key ...
	I0913 18:21:05.899741   11774 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19636-3724/.minikube/proxy-client-ca.key: {Name:mk6287a32ce037416b89e1fb9c3406f9f9919e9d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:21:05.899801   11774 certs.go:256] generating profile certs ...
	I0913 18:21:05.899852   11774 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.key
	I0913 18:21:05.899871   11774 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt with IP's: []
	I0913 18:21:06.054981   11774 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt ...
	I0913 18:21:06.055009   11774 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: {Name:mk8ffe00c34c62d38ca12589d6401134b22742c9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:21:06.055150   11774 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.key ...
	I0913 18:21:06.055156   11774 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.key: {Name:mk3c6cc57dd1367570cd38344b98be400024fc30 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:21:06.055221   11774 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/apiserver.key.1603746e
	I0913 18:21:06.055239   11774 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/apiserver.crt.1603746e with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
	I0913 18:21:06.335406   11774 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/apiserver.crt.1603746e ...
	I0913 18:21:06.335429   11774 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/apiserver.crt.1603746e: {Name:mke8879123003ad5180e44d22dc9118aef6500a3 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:21:06.335561   11774 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/apiserver.key.1603746e ...
	I0913 18:21:06.335572   11774 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/apiserver.key.1603746e: {Name:mk95c7d82a90ec8eda1ddb0bb2d7ad29e87a894c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:21:06.335635   11774 certs.go:381] copying /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/apiserver.crt.1603746e -> /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/apiserver.crt
	I0913 18:21:06.335701   11774 certs.go:385] copying /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/apiserver.key.1603746e -> /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/apiserver.key
	I0913 18:21:06.335746   11774 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/proxy-client.key
	I0913 18:21:06.335762   11774 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/proxy-client.crt with IP's: []
	I0913 18:21:06.487619   11774 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/proxy-client.crt ...
	I0913 18:21:06.487643   11774 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/proxy-client.crt: {Name:mkb891460f76649e120fcc26379551a0a1e2bb9f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:21:06.487770   11774 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/proxy-client.key ...
	I0913 18:21:06.487780   11774 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/proxy-client.key: {Name:mked965dcfee3f6c105d8ad01fca0550fd7bc13f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:21:06.487925   11774 certs.go:484] found cert: /home/jenkins/minikube-integration/19636-3724/.minikube/certs/ca-key.pem (1675 bytes)
	I0913 18:21:06.487955   11774 certs.go:484] found cert: /home/jenkins/minikube-integration/19636-3724/.minikube/certs/ca.pem (1078 bytes)
	I0913 18:21:06.487974   11774 certs.go:484] found cert: /home/jenkins/minikube-integration/19636-3724/.minikube/certs/cert.pem (1123 bytes)
	I0913 18:21:06.487998   11774 certs.go:484] found cert: /home/jenkins/minikube-integration/19636-3724/.minikube/certs/key.pem (1679 bytes)
	I0913 18:21:06.488507   11774 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19636-3724/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
	I0913 18:21:06.509188   11774 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19636-3724/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
	I0913 18:21:06.528472   11774 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19636-3724/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
	I0913 18:21:06.548151   11774 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19636-3724/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
	I0913 18:21:06.568082   11774 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
	I0913 18:21:06.587847   11774 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
	I0913 18:21:06.607148   11774 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
	I0913 18:21:06.626503   11774 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
	I0913 18:21:06.646216   11774 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19636-3724/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
	I0913 18:21:06.665743   11774 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
	I0913 18:21:06.680252   11774 ssh_runner.go:195] Run: openssl version
	I0913 18:21:06.684764   11774 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
	I0913 18:21:06.692329   11774 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
	I0913 18:21:06.695101   11774 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 13 18:21 /usr/share/ca-certificates/minikubeCA.pem
	I0913 18:21:06.695136   11774 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
	I0913 18:21:06.700774   11774 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
	I0913 18:21:06.708737   11774 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
	I0913 18:21:06.711449   11774 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
	stdout:
	
	stderr:
	stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
	I0913 18:21:06.711490   11774 kubeadm.go:392] StartCluster: {Name:addons-315216 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-315216 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
	I0913 18:21:06.711611   11774 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
	I0913 18:21:06.727237   11774 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
	I0913 18:21:06.734660   11774 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
	I0913 18:21:06.742046   11774 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
	I0913 18:21:06.742083   11774 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
	I0913 18:21:06.749198   11774 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
	stdout:
	
	stderr:
	ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
	ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
	ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
	ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
	I0913 18:21:06.749211   11774 kubeadm.go:157] found existing configuration files:
	
	I0913 18:21:06.749242   11774 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
	I0913 18:21:06.756257   11774 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
	stdout:
	
	stderr:
	grep: /etc/kubernetes/admin.conf: No such file or directory
	I0913 18:21:06.756304   11774 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
	I0913 18:21:06.762924   11774 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
	I0913 18:21:06.769838   11774 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
	stdout:
	
	stderr:
	grep: /etc/kubernetes/kubelet.conf: No such file or directory
	I0913 18:21:06.769887   11774 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
	I0913 18:21:06.776376   11774 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
	I0913 18:21:06.782968   11774 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
	stdout:
	
	stderr:
	grep: /etc/kubernetes/controller-manager.conf: No such file or directory
	I0913 18:21:06.782998   11774 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
	I0913 18:21:06.789373   11774 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
	I0913 18:21:06.795894   11774 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
	stdout:
	
	stderr:
	grep: /etc/kubernetes/scheduler.conf: No such file or directory
	I0913 18:21:06.795927   11774 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
	I0913 18:21:06.802409   11774 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml  --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
	I0913 18:21:06.833231   11774 kubeadm.go:310] [init] Using Kubernetes version: v1.31.1
	I0913 18:21:06.833296   11774 kubeadm.go:310] [preflight] Running pre-flight checks
	I0913 18:21:06.849891   11774 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
	I0913 18:21:06.849971   11774 kubeadm.go:310] KERNEL_VERSION: 5.15.0-1068-gcp
	I0913 18:21:06.850018   11774 kubeadm.go:310] OS: Linux
	I0913 18:21:06.850081   11774 kubeadm.go:310] CGROUPS_CPU: enabled
	I0913 18:21:06.850127   11774 kubeadm.go:310] CGROUPS_CPUACCT: enabled
	I0913 18:21:06.850165   11774 kubeadm.go:310] CGROUPS_CPUSET: enabled
	I0913 18:21:06.850209   11774 kubeadm.go:310] CGROUPS_DEVICES: enabled
	I0913 18:21:06.850300   11774 kubeadm.go:310] CGROUPS_FREEZER: enabled
	I0913 18:21:06.850387   11774 kubeadm.go:310] CGROUPS_MEMORY: enabled
	I0913 18:21:06.850427   11774 kubeadm.go:310] CGROUPS_PIDS: enabled
	I0913 18:21:06.850469   11774 kubeadm.go:310] CGROUPS_HUGETLB: enabled
	I0913 18:21:06.850512   11774 kubeadm.go:310] CGROUPS_BLKIO: enabled
	I0913 18:21:06.895712   11774 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
	I0913 18:21:06.895829   11774 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
	I0913 18:21:06.895941   11774 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
	I0913 18:21:06.905948   11774 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
	I0913 18:21:06.909010   11774 out.go:235]   - Generating certificates and keys ...
	I0913 18:21:06.909094   11774 kubeadm.go:310] [certs] Using existing ca certificate authority
	I0913 18:21:06.909172   11774 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
	I0913 18:21:07.019266   11774 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
	I0913 18:21:07.174731   11774 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
	I0913 18:21:07.378516   11774 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
	I0913 18:21:07.506797   11774 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
	I0913 18:21:07.692384   11774 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
	I0913 18:21:07.692561   11774 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-315216 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
	I0913 18:21:07.777337   11774 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
	I0913 18:21:07.777498   11774 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-315216 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
	I0913 18:21:07.876125   11774 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
	I0913 18:21:08.190849   11774 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
	I0913 18:21:08.544927   11774 kubeadm.go:310] [certs] Generating "sa" key and public key
	I0913 18:21:08.545003   11774 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
	I0913 18:21:08.846251   11774 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
	I0913 18:21:09.124086   11774 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
	I0913 18:21:09.210796   11774 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
	I0913 18:21:09.280743   11774 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
	I0913 18:21:09.411368   11774 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
	I0913 18:21:09.411856   11774 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
	I0913 18:21:09.414006   11774 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
	I0913 18:21:09.416285   11774 out.go:235]   - Booting up control plane ...
	I0913 18:21:09.416375   11774 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
	I0913 18:21:09.416444   11774 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
	I0913 18:21:09.416971   11774 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
	I0913 18:21:09.425414   11774 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
	I0913 18:21:09.430415   11774 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
	I0913 18:21:09.430486   11774 kubeadm.go:310] [kubelet-start] Starting the kubelet
	I0913 18:21:09.505859   11774 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
	I0913 18:21:09.505998   11774 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
	I0913 18:21:10.007228   11774 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 501.46246ms
	I0913 18:21:10.007355   11774 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
	I0913 18:21:14.508844   11774 kubeadm.go:310] [api-check] The API server is healthy after 4.501572296s
	I0913 18:21:14.519024   11774 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
	I0913 18:21:14.527163   11774 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
	I0913 18:21:14.542540   11774 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
	I0913 18:21:14.542785   11774 kubeadm.go:310] [mark-control-plane] Marking the node addons-315216 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
	I0913 18:21:14.548747   11774 kubeadm.go:310] [bootstrap-token] Using token: tuhy79.cpnl0w4vi8mb0vd6
	I0913 18:21:14.549913   11774 out.go:235]   - Configuring RBAC rules ...
	I0913 18:21:14.550071   11774 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
	I0913 18:21:14.552758   11774 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
	I0913 18:21:14.557871   11774 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
	I0913 18:21:14.559932   11774 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
	I0913 18:21:14.562063   11774 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
	I0913 18:21:14.564104   11774 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
	I0913 18:21:14.914781   11774 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
	I0913 18:21:15.332104   11774 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
	I0913 18:21:15.915151   11774 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
	I0913 18:21:15.915898   11774 kubeadm.go:310] 
	I0913 18:21:15.916003   11774 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
	I0913 18:21:15.916016   11774 kubeadm.go:310] 
	I0913 18:21:15.916149   11774 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
	I0913 18:21:15.916161   11774 kubeadm.go:310] 
	I0913 18:21:15.916196   11774 kubeadm.go:310]   mkdir -p $HOME/.kube
	I0913 18:21:15.916315   11774 kubeadm.go:310]   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
	I0913 18:21:15.916405   11774 kubeadm.go:310]   sudo chown $(id -u):$(id -g) $HOME/.kube/config
	I0913 18:21:15.916423   11774 kubeadm.go:310] 
	I0913 18:21:15.916508   11774 kubeadm.go:310] Alternatively, if you are the root user, you can run:
	I0913 18:21:15.916517   11774 kubeadm.go:310] 
	I0913 18:21:15.916587   11774 kubeadm.go:310]   export KUBECONFIG=/etc/kubernetes/admin.conf
	I0913 18:21:15.916595   11774 kubeadm.go:310] 
	I0913 18:21:15.916668   11774 kubeadm.go:310] You should now deploy a pod network to the cluster.
	I0913 18:21:15.916775   11774 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
	I0913 18:21:15.916870   11774 kubeadm.go:310]   https://kubernetes.io/docs/concepts/cluster-administration/addons/
	I0913 18:21:15.916893   11774 kubeadm.go:310] 
	I0913 18:21:15.916995   11774 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
	I0913 18:21:15.917082   11774 kubeadm.go:310] and service account keys on each node and then running the following as root:
	I0913 18:21:15.917093   11774 kubeadm.go:310] 
	I0913 18:21:15.917200   11774 kubeadm.go:310]   kubeadm join control-plane.minikube.internal:8443 --token tuhy79.cpnl0w4vi8mb0vd6 \
	I0913 18:21:15.917333   11774 kubeadm.go:310] 	--discovery-token-ca-cert-hash sha256:ff52e83c10dc1cb593d25925eeb92aea4da9258f323e73d474dae3b251f98a2a \
	I0913 18:21:15.917365   11774 kubeadm.go:310] 	--control-plane 
	I0913 18:21:15.917376   11774 kubeadm.go:310] 
	I0913 18:21:15.917508   11774 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
	I0913 18:21:15.917523   11774 kubeadm.go:310] 
	I0913 18:21:15.917648   11774 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token tuhy79.cpnl0w4vi8mb0vd6 \
	I0913 18:21:15.917809   11774 kubeadm.go:310] 	--discovery-token-ca-cert-hash sha256:ff52e83c10dc1cb593d25925eeb92aea4da9258f323e73d474dae3b251f98a2a 
	I0913 18:21:15.919538   11774 kubeadm.go:310] W0913 18:21:06.830855    1922 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
	I0913 18:21:15.919824   11774 kubeadm.go:310] W0913 18:21:06.831452    1922 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
	I0913 18:21:15.920012   11774 kubeadm.go:310] 	[WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1068-gcp\n", err: exit status 1
	I0913 18:21:15.920180   11774 kubeadm.go:310] 	[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
	I0913 18:21:15.920206   11774 cni.go:84] Creating CNI manager for ""
	I0913 18:21:15.920219   11774 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
	I0913 18:21:15.921669   11774 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
	I0913 18:21:15.922983   11774 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
	I0913 18:21:15.930625   11774 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
	I0913 18:21:15.945389   11774 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
	I0913 18:21:15.945433   11774 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
	I0913 18:21:15.945467   11774 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-315216 minikube.k8s.io/updated_at=2024_09_13T18_21_15_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=fdd33bebc6743cfd1c61ec7fe066add478610a92 minikube.k8s.io/name=addons-315216 minikube.k8s.io/primary=true
	I0913 18:21:16.006268   11774 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0913 18:21:16.006305   11774 ops.go:34] apiserver oom_adj: -16
	I0913 18:21:16.506462   11774 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0913 18:21:17.006456   11774 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0913 18:21:17.506390   11774 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0913 18:21:18.006856   11774 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0913 18:21:18.506859   11774 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0913 18:21:19.006412   11774 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0913 18:21:19.506846   11774 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0913 18:21:20.006430   11774 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0913 18:21:20.506320   11774 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
	I0913 18:21:20.569353   11774 kubeadm.go:1113] duration metric: took 4.623961565s to wait for elevateKubeSystemPrivileges
	I0913 18:21:20.569380   11774 kubeadm.go:394] duration metric: took 13.857893144s to StartCluster
	I0913 18:21:20.569400   11774 settings.go:142] acquiring lock: {Name:mkdfe968303b166acd0b853425fddda0e662ae88 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:21:20.569517   11774 settings.go:150] Updating kubeconfig:  /home/jenkins/minikube-integration/19636-3724/kubeconfig
	I0913 18:21:20.569914   11774 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19636-3724/kubeconfig: {Name:mkb3de2d521cdfaf51b480e5ea10ef30c8160b52 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
	I0913 18:21:20.570138   11774 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
	I0913 18:21:20.570262   11774 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
	I0913 18:21:20.570276   11774 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
	I0913 18:21:20.570408   11774 addons.go:69] Setting yakd=true in profile "addons-315216"
	I0913 18:21:20.570415   11774 addons.go:69] Setting inspektor-gadget=true in profile "addons-315216"
	I0913 18:21:20.570431   11774 addons.go:234] Setting addon yakd=true in "addons-315216"
	I0913 18:21:20.570443   11774 addons.go:234] Setting addon inspektor-gadget=true in "addons-315216"
	I0913 18:21:20.570433   11774 addons.go:69] Setting storage-provisioner=true in profile "addons-315216"
	I0913 18:21:20.570460   11774 addons.go:234] Setting addon storage-provisioner=true in "addons-315216"
	I0913 18:21:20.570461   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.570457   11774 addons.go:69] Setting volcano=true in profile "addons-315216"
	I0913 18:21:20.570471   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.570477   11774 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-315216"
	I0913 18:21:20.570489   11774 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-315216"
	I0913 18:21:20.570495   11774 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-315216"
	I0913 18:21:20.570499   11774 addons.go:69] Setting registry=true in profile "addons-315216"
	I0913 18:21:20.570503   11774 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-315216"
	I0913 18:21:20.570510   11774 addons.go:234] Setting addon registry=true in "addons-315216"
	I0913 18:21:20.570514   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.570516   11774 addons.go:69] Setting metrics-server=true in profile "addons-315216"
	I0913 18:21:20.570528   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.570537   11774 addons.go:234] Setting addon metrics-server=true in "addons-315216"
	I0913 18:21:20.570552   11774 addons.go:69] Setting volumesnapshots=true in profile "addons-315216"
	I0913 18:21:20.570564   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.570572   11774 addons.go:234] Setting addon volumesnapshots=true in "addons-315216"
	I0913 18:21:20.570591   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.570479   11774 addons.go:234] Setting addon volcano=true in "addons-315216"
	I0913 18:21:20.570640   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.570820   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.570971   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.571011   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.571026   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.571050   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.571055   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.571189   11774 addons.go:69] Setting default-storageclass=true in profile "addons-315216"
	I0913 18:21:20.571212   11774 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-315216"
	I0913 18:21:20.571328   11774 addons.go:69] Setting cloud-spanner=true in profile "addons-315216"
	I0913 18:21:20.571348   11774 addons.go:234] Setting addon cloud-spanner=true in "addons-315216"
	I0913 18:21:20.571385   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.571485   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.571897   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.570490   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.572741   11774 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-315216"
	I0913 18:21:20.572849   11774 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-315216"
	I0913 18:21:20.572876   11774 addons.go:69] Setting ingress=true in profile "addons-315216"
	I0913 18:21:20.572891   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.572898   11774 addons.go:234] Setting addon ingress=true in "addons-315216"
	I0913 18:21:20.572932   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.573983   11774 out.go:177] * Verifying Kubernetes components...
	I0913 18:21:20.570465   11774 config.go:182] Loaded profile config "addons-315216": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
	I0913 18:21:20.575003   11774 addons.go:69] Setting ingress-dns=true in profile "addons-315216"
	I0913 18:21:20.575020   11774 addons.go:234] Setting addon ingress-dns=true in "addons-315216"
	I0913 18:21:20.575052   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.575356   11774 ssh_runner.go:195] Run: sudo systemctl daemon-reload
	I0913 18:21:20.575372   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.572862   11774 addons.go:69] Setting gcp-auth=true in profile "addons-315216"
	I0913 18:21:20.575425   11774 mustload.go:65] Loading cluster: addons-315216
	I0913 18:21:20.601545   11774 addons.go:234] Setting addon default-storageclass=true in "addons-315216"
	I0913 18:21:20.601582   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.601973   11774 config.go:182] Loaded profile config "addons-315216": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
	I0913 18:21:20.602041   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.602115   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.602234   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.602271   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.602934   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.603262   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.604249   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.607578   11774 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-315216"
	I0913 18:21:20.607622   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.608048   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:20.610729   11774 out.go:177]   - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0
	I0913 18:21:20.612581   11774 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
	I0913 18:21:20.612603   11774 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
	I0913 18:21:20.612662   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.614579   11774 out.go:177]   - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
	I0913 18:21:20.614579   11774 out.go:177]   - Using image docker.io/registry:2.8.3
	I0913 18:21:20.616642   11774 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
	I0913 18:21:20.616664   11774 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
	I0913 18:21:20.616727   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.620949   11774 out.go:177]   - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
	I0913 18:21:20.622591   11774 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
	I0913 18:21:20.622614   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
	I0913 18:21:20.622638   11774 out.go:177]   - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
	I0913 18:21:20.622660   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.623704   11774 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
	I0913 18:21:20.623725   11774 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
	I0913 18:21:20.623774   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.628386   11774 out.go:177]   - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
	I0913 18:21:20.631097   11774 out.go:177]   - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
	I0913 18:21:20.632280   11774 out.go:177]   - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
	I0913 18:21:20.634785   11774 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
	I0913 18:21:20.634807   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
	I0913 18:21:20.634857   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.653297   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:20.660387   11774 out.go:177]   - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
	I0913 18:21:20.661518   11774 out.go:177]   - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
	I0913 18:21:20.665344   11774 out.go:177]   - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
	I0913 18:21:20.665580   11774 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
	I0913 18:21:20.665599   11774 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
	I0913 18:21:20.665694   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.665842   11774 out.go:177]   - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
	I0913 18:21:20.666768   11774 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
	I0913 18:21:20.666792   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
	I0913 18:21:20.666874   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.668703   11774 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
	I0913 18:21:20.668723   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
	I0913 18:21:20.668870   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.677520   11774 out.go:177]   - Using image docker.io/rancher/local-path-provisioner:v0.0.22
	I0913 18:21:20.677717   11774 out.go:177]   - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
	I0913 18:21:20.679180   11774 out.go:177]   - Using image docker.io/busybox:stable
	I0913 18:21:20.679250   11774 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
	I0913 18:21:20.679262   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
	I0913 18:21:20.679303   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.680243   11774 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
	I0913 18:21:20.680263   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
	I0913 18:21:20.680315   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.680603   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:20.685494   11774 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
	I0913 18:21:20.686690   11774 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
	I0913 18:21:20.687650   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:20.689807   11774 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
	I0913 18:21:20.690391   11774 out.go:177]   - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.23
	I0913 18:21:20.691485   11774 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
	I0913 18:21:20.692490   11774 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
	I0913 18:21:20.693103   11774 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
	I0913 18:21:20.693119   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
	I0913 18:21:20.693169   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.697714   11774 out.go:177]   - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
	I0913 18:21:20.699795   11774 out.go:177]   - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
	I0913 18:21:20.700847   11774 out.go:177]   - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
	I0913 18:21:20.701758   11774 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
	I0913 18:21:20.701781   11774 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
	I0913 18:21:20.701839   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.701883   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:20.702410   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:20.709500   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:20.717198   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:20.719913   11774 out.go:177]   - Using image gcr.io/k8s-minikube/storage-provisioner:v5
	I0913 18:21:20.721391   11774 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
	I0913 18:21:20.721410   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
	I0913 18:21:20.721460   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.730894   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:20.732089   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:20.732102   11774 out.go:177]   - Using image docker.io/marcnuri/yakd:0.0.5
	I0913 18:21:20.733210   11774 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
	I0913 18:21:20.733233   11774 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
	I0913 18:21:20.733288   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:20.735602   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:20.737183   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:20.746730   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:20.752534   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:20.755015   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:20.759601   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:21.000184   11774 ssh_runner.go:195] Run: sudo systemctl start kubelet
	I0913 18:21:21.000234   11774 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^        forward . \/etc\/resolv.conf.*/i \        hosts {\n           192.168.49.1 host.minikube.internal\n           fallthrough\n        }' -e '/^        errors *$/i \        log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
	I0913 18:21:21.097656   11774 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
	I0913 18:21:21.097696   11774 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
	I0913 18:21:21.099511   11774 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
	I0913 18:21:21.099530   11774 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
	I0913 18:21:21.113541   11774 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
	I0913 18:21:21.113568   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
	I0913 18:21:21.117247   11774 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
	I0913 18:21:21.117270   11774 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
	I0913 18:21:21.206385   11774 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
	I0913 18:21:21.206413   11774 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
	I0913 18:21:21.208540   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
	I0913 18:21:21.209125   11774 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
	I0913 18:21:21.209147   11774 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
	I0913 18:21:21.295028   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
	I0913 18:21:21.295317   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
	I0913 18:21:21.298881   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
	I0913 18:21:21.306064   11774 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
	I0913 18:21:21.306085   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
	I0913 18:21:21.394313   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
	I0913 18:21:21.395057   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
	I0913 18:21:21.395523   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
	I0913 18:21:21.396534   11774 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
	I0913 18:21:21.396553   11774 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
	I0913 18:21:21.400048   11774 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
	I0913 18:21:21.400068   11774 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
	I0913 18:21:21.405246   11774 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
	I0913 18:21:21.405264   11774 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
	I0913 18:21:21.417477   11774 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
	I0913 18:21:21.417567   11774 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
	I0913 18:21:21.496972   11774 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
	I0913 18:21:21.497000   11774 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
	I0913 18:21:21.596153   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
	I0913 18:21:21.715723   11774 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
	I0913 18:21:21.715804   11774 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
	I0913 18:21:21.716694   11774 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
	I0913 18:21:21.716746   11774 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
	I0913 18:21:21.803596   11774 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
	I0913 18:21:21.803670   11774 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
	I0913 18:21:21.806508   11774 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
	I0913 18:21:21.806525   11774 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
	I0913 18:21:21.813445   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
	I0913 18:21:22.011664   11774 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
	I0913 18:21:22.011695   11774 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
	I0913 18:21:22.095601   11774 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
	I0913 18:21:22.095631   11774 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
	I0913 18:21:22.497385   11774 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
	I0913 18:21:22.497415   11774 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
	I0913 18:21:22.516142   11774 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
	I0913 18:21:22.516224   11774 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
	I0913 18:21:22.594265   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
	I0913 18:21:22.606708   11774 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
	I0913 18:21:22.606739   11774 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
	I0913 18:21:22.699398   11774 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
	I0913 18:21:22.699482   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
	I0913 18:21:22.806777   11774 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^        forward . \/etc\/resolv.conf.*/i \        hosts {\n           192.168.49.1 host.minikube.internal\n           fallthrough\n        }' -e '/^        errors *$/i \        log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.806505092s)
	I0913 18:21:22.806871   11774 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
	I0913 18:21:22.808152   11774 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.8079396s)
	I0913 18:21:22.809102   11774 node_ready.go:35] waiting up to 6m0s for node "addons-315216" to be "Ready" ...
	I0913 18:21:22.817613   11774 node_ready.go:49] node "addons-315216" has status "Ready":"True"
	I0913 18:21:22.817634   11774 node_ready.go:38] duration metric: took 8.466753ms for node "addons-315216" to be "Ready" ...
	I0913 18:21:22.817646   11774 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
	I0913 18:21:22.909529   11774 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-6hp86" in "kube-system" namespace to be "Ready" ...
	I0913 18:21:22.913215   11774 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
	I0913 18:21:22.913281   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
	I0913 18:21:22.997782   11774 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
	I0913 18:21:22.997866   11774 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
	I0913 18:21:23.012764   11774 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
	I0913 18:21:23.012846   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
	I0913 18:21:23.114746   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
	I0913 18:21:23.311433   11774 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-315216" context rescaled to 1 replicas
	I0913 18:21:23.396145   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
	I0913 18:21:23.415943   11774 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
	I0913 18:21:23.415968   11774 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
	I0913 18:21:23.511834   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (2.303259754s)
	I0913 18:21:23.794546   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
	I0913 18:21:24.009567   11774 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
	I0913 18:21:24.009654   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
	I0913 18:21:24.999241   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-6hp86" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:25.400146   11774 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
	I0913 18:21:25.400229   11774 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
	I0913 18:21:25.695583   11774 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
	I0913 18:21:25.695681   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
	I0913 18:21:26.005534   11774 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
	I0913 18:21:26.005559   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
	I0913 18:21:26.603064   11774 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
	I0913 18:21:26.603140   11774 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
	I0913 18:21:26.907208   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
	I0913 18:21:27.417155   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-6hp86" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:27.796753   11774 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
	I0913 18:21:27.796883   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:27.823889   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:28.594301   11774 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
	I0913 18:21:28.714379   11774 addons.go:234] Setting addon gcp-auth=true in "addons-315216"
	I0913 18:21:28.714482   11774 host.go:66] Checking if "addons-315216" exists ...
	I0913 18:21:28.715031   11774 cli_runner.go:164] Run: docker container inspect addons-315216 --format={{.State.Status}}
	I0913 18:21:28.734258   11774 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
	I0913 18:21:28.734299   11774 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-315216
	I0913 18:21:28.748760   11774 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/addons-315216/id_rsa Username:docker}
	I0913 18:21:29.414567   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (8.119500486s)
	I0913 18:21:29.414709   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (8.119362945s)
	I0913 18:21:29.414764   11774 addons.go:475] Verifying addon ingress=true in "addons-315216"
	I0913 18:21:29.417004   11774 out.go:177] * Verifying ingress addon...
	I0913 18:21:29.419653   11774 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
	I0913 18:21:29.498488   11774 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
	I0913 18:21:29.498567   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:29.919662   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-6hp86" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:29.923441   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:30.500533   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:31.000341   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:31.598913   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:31.924367   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:32.302789   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (11.003859833s)
	I0913 18:21:32.302954   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (10.908597259s)
	I0913 18:21:32.303011   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (10.907936142s)
	I0913 18:21:32.303198   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (10.907644072s)
	I0913 18:21:32.303284   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (10.707048658s)
	I0913 18:21:32.303341   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (10.489875349s)
	I0913 18:21:32.303360   11774 addons.go:475] Verifying addon registry=true in "addons-315216"
	I0913 18:21:32.303601   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (9.70923476s)
	I0913 18:21:32.303721   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (9.188888773s)
	I0913 18:21:32.303795   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (8.907546037s)
	I0913 18:21:32.303833   11774 addons.go:475] Verifying addon metrics-server=true in "addons-315216"
	I0913 18:21:32.303903   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (8.509321749s)
	W0913 18:21:32.303935   11774 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
	stdout:
	customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
	customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
	customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
	serviceaccount/snapshot-controller created
	clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
	clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
	role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
	rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
	deployment.apps/snapshot-controller created
	
	stderr:
	error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
	ensure CRDs are installed first
	I0913 18:21:32.303953   11774 retry.go:31] will retry after 224.652163ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
	stdout:
	customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
	customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
	customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
	serviceaccount/snapshot-controller created
	clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
	clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
	role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
	rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
	deployment.apps/snapshot-controller created
	
	stderr:
	error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
	ensure CRDs are installed first
	I0913 18:21:32.305090   11774 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
	
		minikube -p addons-315216 service yakd-dashboard -n yakd-dashboard
	
	I0913 18:21:32.305198   11774 out.go:177] * Verifying registry addon...
	I0913 18:21:32.308136   11774 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
	W0913 18:21:32.311988   11774 out.go:270] ! Enabling 'default-storageclass' returned an error: running callbacks: [Error making standard the default storage class: Error while marking storage class local-path as non-default: Operation cannot be fulfilled on storageclasses.storage.k8s.io "local-path": the object has been modified; please apply your changes to the latest version and try again]
	I0913 18:21:32.313154   11774 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
	I0913 18:21:32.313209   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:32.419695   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-6hp86" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:32.497263   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:32.529635   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
	I0913 18:21:32.811790   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:33.006171   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:33.311535   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:33.426789   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:33.811916   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:33.819832   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (6.912494384s)
	I0913 18:21:33.819888   11774 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-315216"
	I0913 18:21:33.820106   11774 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (5.085810144s)
	I0913 18:21:33.821541   11774 out.go:177] * Verifying csi-hostpath-driver addon...
	I0913 18:21:33.821600   11774 out.go:177]   - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
	I0913 18:21:33.822720   11774 out.go:177]   - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
	I0913 18:21:33.823625   11774 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
	I0913 18:21:33.823832   11774 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
	I0913 18:21:33.823850   11774 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
	I0913 18:21:33.904039   11774 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
	I0913 18:21:33.904064   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:33.917802   11774 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
	I0913 18:21:33.917826   11774 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
	I0913 18:21:33.923230   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:34.005112   11774 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
	I0913 18:21:34.005138   11774 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
	I0913 18:21:34.100384   11774 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
	I0913 18:21:34.312061   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:34.398003   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:34.423260   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:34.798826   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.269143586s)
	I0913 18:21:34.812936   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:34.897457   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:34.917582   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-6hp86" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:34.922870   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:35.314915   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:35.413771   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:35.419146   11774 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.318696529s)
	I0913 18:21:35.421424   11774 addons.go:475] Verifying addon gcp-auth=true in "addons-315216"
	I0913 18:21:35.423187   11774 out.go:177] * Verifying gcp-auth addon...
	I0913 18:21:35.425314   11774 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
	I0913 18:21:35.513108   11774 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
	I0913 18:21:35.514446   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:35.811833   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:35.827965   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:35.915764   11774 pod_ready.go:98] pod "coredns-7c65d6cfc9-6hp86" in "kube-system" namespace has status phase "Succeeded" (skipping!): {Phase:Succeeded Conditions:[{Type:PodReadyToStartContainers Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-13 18:21:35 +0000 UTC Reason: Message:} {Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-13 18:21:21 +0000 UTC Reason:PodCompleted Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-13 18:21:21 +0000 UTC Reason:PodCompleted Message:} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-13 18:21:21 +0000 UTC Reason:PodCompleted Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-13 18:21:21 +0000 UTC Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.49.2 HostIPs:[{IP:192.168.49.2
}] PodIP:10.244.0.3 PodIPs:[{IP:10.244.0.3}] StartTime:2024-09-13 18:21:21 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:coredns State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2024-09-13 18:21:23 +0000 UTC,FinishedAt:2024-09-13 18:21:34 +0000 UTC,ContainerID:docker://5ca960f2c2b21944355698ab2396f91ab5e7561023a354e921920997231a4b0f,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:registry.k8s.io/coredns/coredns:v1.11.3 ImageID:docker-pullable://registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e ContainerID:docker://5ca960f2c2b21944355698ab2396f91ab5e7561023a354e921920997231a4b0f Started:0xc00098c6d0 AllocatedResources:map[] Resources:nil VolumeMounts:[{Name:config-volume MountPath:/etc/coredns ReadOnly:true RecursiveReadOnly:0xc001df07f0} {Name:kube-api-access-6cpc6 MountPath:/var/run/secrets/kubernetes.io/serviceaccount
ReadOnly:true RecursiveReadOnly:0xc001df0800}] User:nil AllocatedResourcesStatus:[]}] QOSClass:Burstable EphemeralContainerStatuses:[] Resize: ResourceClaimStatuses:[]}
	I0913 18:21:35.915798   11774 pod_ready.go:82] duration metric: took 13.006229311s for pod "coredns-7c65d6cfc9-6hp86" in "kube-system" namespace to be "Ready" ...
	E0913 18:21:35.915812   11774 pod_ready.go:67] WaitExtra: waitPodCondition: pod "coredns-7c65d6cfc9-6hp86" in "kube-system" namespace has status phase "Succeeded" (skipping!): {Phase:Succeeded Conditions:[{Type:PodReadyToStartContainers Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-13 18:21:35 +0000 UTC Reason: Message:} {Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-13 18:21:21 +0000 UTC Reason:PodCompleted Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-13 18:21:21 +0000 UTC Reason:PodCompleted Message:} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-13 18:21:21 +0000 UTC Reason:PodCompleted Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-13 18:21:21 +0000 UTC Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.4
9.2 HostIPs:[{IP:192.168.49.2}] PodIP:10.244.0.3 PodIPs:[{IP:10.244.0.3}] StartTime:2024-09-13 18:21:21 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:coredns State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2024-09-13 18:21:23 +0000 UTC,FinishedAt:2024-09-13 18:21:34 +0000 UTC,ContainerID:docker://5ca960f2c2b21944355698ab2396f91ab5e7561023a354e921920997231a4b0f,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:registry.k8s.io/coredns/coredns:v1.11.3 ImageID:docker-pullable://registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e ContainerID:docker://5ca960f2c2b21944355698ab2396f91ab5e7561023a354e921920997231a4b0f Started:0xc00098c6d0 AllocatedResources:map[] Resources:nil VolumeMounts:[{Name:config-volume MountPath:/etc/coredns ReadOnly:true RecursiveReadOnly:0xc001df07f0} {Name:kube-api-access-6cpc6 MountPath:/var/run/secrets
/kubernetes.io/serviceaccount ReadOnly:true RecursiveReadOnly:0xc001df0800}] User:nil AllocatedResourcesStatus:[]}] QOSClass:Burstable EphemeralContainerStatuses:[] Resize: ResourceClaimStatuses:[]}
	I0913 18:21:35.915823   11774 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace to be "Ready" ...
	I0913 18:21:35.922638   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:36.312014   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:36.328305   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:36.423380   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:36.812286   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:36.827185   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:36.923030   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:37.312582   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:37.328938   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:37.423195   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:37.811764   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:37.827831   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:37.921294   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:37.923517   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:38.311884   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:38.328299   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:38.423274   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:38.811514   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:38.827260   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:38.922892   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:39.312529   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:39.327893   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:39.422899   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:39.811394   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:39.827763   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:39.923191   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:39.923659   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:40.311866   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:40.327862   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:40.423465   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:40.812317   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:40.827790   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:40.923115   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:41.311748   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:41.332333   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:41.422521   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:41.811711   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:41.827492   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:41.922754   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:42.311635   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:42.327257   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:42.420938   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:42.423259   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:42.812083   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:42.828807   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:42.923150   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:43.312092   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:43.328201   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:43.422787   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:43.812316   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:43.827446   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:43.923091   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:44.311183   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:44.327610   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:44.421849   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:44.422408   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:44.812385   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:44.827739   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:44.923038   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:45.316448   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:45.327583   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:45.423085   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:45.812210   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:45.828217   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:45.923346   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:46.312170   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:46.328139   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:46.423010   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:46.812384   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:46.827920   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:46.920626   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:46.923234   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:47.311375   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:47.327411   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:47.422896   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:47.812370   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:47.827866   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:47.922908   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:48.314929   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:48.419098   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:48.423541   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:48.811636   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:48.827303   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:48.922759   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:48.923463   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:49.312024   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:49.331408   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:49.423186   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:49.812036   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:49.828393   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:49.923124   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:50.314726   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:50.327943   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:50.423668   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:50.812013   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:50.827620   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:50.922564   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:51.312787   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:51.327792   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:51.421593   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:51.423714   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:51.811725   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:51.828099   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:51.923461   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:52.312108   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:52.328603   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:52.422942   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:52.812256   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:52.826518   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:52.923001   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:53.312640   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
	I0913 18:21:53.328216   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:53.423353   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:53.811520   11774 kapi.go:107] duration metric: took 21.503379767s to wait for kubernetes.io/minikube-addons=registry ...
	I0913 18:21:53.827552   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:54.000552   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:54.001550   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:54.328967   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:54.422891   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:54.828157   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:54.922795   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:55.327713   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:55.423148   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:55.828607   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:55.922867   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:56.328241   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:56.421773   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:56.423089   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:56.827926   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:56.927813   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:57.328200   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:57.423043   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:57.828804   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:57.924354   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:58.328609   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:58.422187   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace has status "Ready":"False"
	I0913 18:21:58.422904   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:58.828101   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:58.927866   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:59.408827   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:59.478301   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:21:59.828083   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:21:59.923103   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:00.328029   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:00.423426   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:00.828694   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:00.923292   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:00.926988   11774 pod_ready.go:103] pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace has status "Ready":"False"
	I0913 18:22:01.329441   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:01.429479   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:01.827812   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:01.923509   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:02.328129   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:02.423900   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:02.423983   11774 pod_ready.go:93] pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace has status "Ready":"True"
	I0913 18:22:02.424001   11774 pod_ready.go:82] duration metric: took 26.508168678s for pod "coredns-7c65d6cfc9-nj2wd" in "kube-system" namespace to be "Ready" ...
	I0913 18:22:02.424012   11774 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-315216" in "kube-system" namespace to be "Ready" ...
	I0913 18:22:02.428411   11774 pod_ready.go:93] pod "etcd-addons-315216" in "kube-system" namespace has status "Ready":"True"
	I0913 18:22:02.428485   11774 pod_ready.go:82] duration metric: took 4.463179ms for pod "etcd-addons-315216" in "kube-system" namespace to be "Ready" ...
	I0913 18:22:02.428507   11774 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-315216" in "kube-system" namespace to be "Ready" ...
	I0913 18:22:02.432686   11774 pod_ready.go:93] pod "kube-apiserver-addons-315216" in "kube-system" namespace has status "Ready":"True"
	I0913 18:22:02.432705   11774 pod_ready.go:82] duration metric: took 4.183014ms for pod "kube-apiserver-addons-315216" in "kube-system" namespace to be "Ready" ...
	I0913 18:22:02.432739   11774 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-315216" in "kube-system" namespace to be "Ready" ...
	I0913 18:22:02.436825   11774 pod_ready.go:93] pod "kube-controller-manager-addons-315216" in "kube-system" namespace has status "Ready":"True"
	I0913 18:22:02.436844   11774 pod_ready.go:82] duration metric: took 4.096685ms for pod "kube-controller-manager-addons-315216" in "kube-system" namespace to be "Ready" ...
	I0913 18:22:02.436855   11774 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-tfw7l" in "kube-system" namespace to be "Ready" ...
	I0913 18:22:02.440526   11774 pod_ready.go:93] pod "kube-proxy-tfw7l" in "kube-system" namespace has status "Ready":"True"
	I0913 18:22:02.440544   11774 pod_ready.go:82] duration metric: took 3.683ms for pod "kube-proxy-tfw7l" in "kube-system" namespace to be "Ready" ...
	I0913 18:22:02.440553   11774 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-315216" in "kube-system" namespace to be "Ready" ...
	I0913 18:22:02.820128   11774 pod_ready.go:93] pod "kube-scheduler-addons-315216" in "kube-system" namespace has status "Ready":"True"
	I0913 18:22:02.820150   11774 pod_ready.go:82] duration metric: took 379.591304ms for pod "kube-scheduler-addons-315216" in "kube-system" namespace to be "Ready" ...
	I0913 18:22:02.820157   11774 pod_ready.go:39] duration metric: took 40.002499793s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
	I0913 18:22:02.820175   11774 api_server.go:52] waiting for apiserver process to appear ...
	I0913 18:22:02.820229   11774 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0913 18:22:02.830758   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:02.903719   11774 api_server.go:72] duration metric: took 42.333539309s to wait for apiserver process to appear ...
	I0913 18:22:02.903749   11774 api_server.go:88] waiting for apiserver healthz status ...
	I0913 18:22:02.903776   11774 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
	I0913 18:22:02.909144   11774 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
	ok
	I0913 18:22:02.910049   11774 api_server.go:141] control plane version: v1.31.1
	I0913 18:22:02.910071   11774 api_server.go:131] duration metric: took 6.315438ms to wait for apiserver health ...
	I0913 18:22:02.910078   11774 system_pods.go:43] waiting for kube-system pods to appear ...
	I0913 18:22:02.923903   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:03.025508   11774 system_pods.go:59] 17 kube-system pods found
	I0913 18:22:03.025550   11774 system_pods.go:61] "coredns-7c65d6cfc9-nj2wd" [fd21cc8a-bae9-439d-bd4b-88a00fc27467] Running
	I0913 18:22:03.025563   11774 system_pods.go:61] "csi-hostpath-attacher-0" [be57972c-58d1-4808-8ff4-ca67190811c9] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
	I0913 18:22:03.025571   11774 system_pods.go:61] "csi-hostpath-resizer-0" [58cef56b-414e-4c64-add2-32d236f09463] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
	I0913 18:22:03.025583   11774 system_pods.go:61] "csi-hostpathplugin-6xs52" [5f3cdfcc-6d9c-4003-b6ad-4fa772f91e6b] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
	I0913 18:22:03.025590   11774 system_pods.go:61] "etcd-addons-315216" [c7b8c5da-ed34-4550-88cb-b57283301f35] Running
	I0913 18:22:03.025597   11774 system_pods.go:61] "kube-apiserver-addons-315216" [65bc5844-551a-424d-b9c4-f84ae4e5e50b] Running
	I0913 18:22:03.025603   11774 system_pods.go:61] "kube-controller-manager-addons-315216" [1eaef35a-ef2c-45ba-964a-d4c711bf0a58] Running
	I0913 18:22:03.025609   11774 system_pods.go:61] "kube-ingress-dns-minikube" [3c02732d-d3f3-4f79-890e-2843e593ca75] Running
	I0913 18:22:03.025615   11774 system_pods.go:61] "kube-proxy-tfw7l" [7125370e-1597-42d4-b4a4-fa01e11f5ceb] Running
	I0913 18:22:03.025624   11774 system_pods.go:61] "kube-scheduler-addons-315216" [7b24d0a6-8ae7-44b1-931b-542172178637] Running
	I0913 18:22:03.025629   11774 system_pods.go:61] "metrics-server-84c5f94fbc-5r2xh" [9fd33765-d89a-4aca-9d83-b2ef709128c4] Running
	I0913 18:22:03.025638   11774 system_pods.go:61] "nvidia-device-plugin-daemonset-tvp28" [845bb4f2-4781-45e0-a654-5b57584b3a70] Running
	I0913 18:22:03.025643   11774 system_pods.go:61] "registry-66c9cd494c-cvtgf" [58047cca-a75a-41df-bc9f-a91dc5a547ca] Running
	I0913 18:22:03.025650   11774 system_pods.go:61] "registry-proxy-f7ltj" [7dfd72e6-65ac-495b-9832-0e900d22d7e6] Running
	I0913 18:22:03.025654   11774 system_pods.go:61] "snapshot-controller-56fcc65765-ntlvm" [41f0b45e-62eb-4d4b-9bee-ebfbe102b16e] Running
	I0913 18:22:03.025660   11774 system_pods.go:61] "snapshot-controller-56fcc65765-vsk8h" [af53978d-78a7-4c1a-960b-e203fade05e4] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
	I0913 18:22:03.025669   11774 system_pods.go:61] "storage-provisioner" [74ceb6a0-d02e-4906-9c7b-1fff3590e1d4] Running
	I0913 18:22:03.026393   11774 system_pods.go:74] duration metric: took 116.302133ms to wait for pod list to return data ...
	I0913 18:22:03.026413   11774 default_sa.go:34] waiting for default service account to be created ...
	I0913 18:22:03.219552   11774 default_sa.go:45] found service account: "default"
	I0913 18:22:03.219576   11774 default_sa.go:55] duration metric: took 193.156753ms for default service account to be created ...
	I0913 18:22:03.219585   11774 system_pods.go:116] waiting for k8s-apps to be running ...
	I0913 18:22:03.328839   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:03.423119   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:03.425924   11774 system_pods.go:86] 17 kube-system pods found
	I0913 18:22:03.425952   11774 system_pods.go:89] "coredns-7c65d6cfc9-nj2wd" [fd21cc8a-bae9-439d-bd4b-88a00fc27467] Running
	I0913 18:22:03.425966   11774 system_pods.go:89] "csi-hostpath-attacher-0" [be57972c-58d1-4808-8ff4-ca67190811c9] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
	I0913 18:22:03.425974   11774 system_pods.go:89] "csi-hostpath-resizer-0" [58cef56b-414e-4c64-add2-32d236f09463] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
	I0913 18:22:03.425984   11774 system_pods.go:89] "csi-hostpathplugin-6xs52" [5f3cdfcc-6d9c-4003-b6ad-4fa772f91e6b] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
	I0913 18:22:03.425993   11774 system_pods.go:89] "etcd-addons-315216" [c7b8c5da-ed34-4550-88cb-b57283301f35] Running
	I0913 18:22:03.426002   11774 system_pods.go:89] "kube-apiserver-addons-315216" [65bc5844-551a-424d-b9c4-f84ae4e5e50b] Running
	I0913 18:22:03.426011   11774 system_pods.go:89] "kube-controller-manager-addons-315216" [1eaef35a-ef2c-45ba-964a-d4c711bf0a58] Running
	I0913 18:22:03.426020   11774 system_pods.go:89] "kube-ingress-dns-minikube" [3c02732d-d3f3-4f79-890e-2843e593ca75] Running
	I0913 18:22:03.426028   11774 system_pods.go:89] "kube-proxy-tfw7l" [7125370e-1597-42d4-b4a4-fa01e11f5ceb] Running
	I0913 18:22:03.426037   11774 system_pods.go:89] "kube-scheduler-addons-315216" [7b24d0a6-8ae7-44b1-931b-542172178637] Running
	I0913 18:22:03.426042   11774 system_pods.go:89] "metrics-server-84c5f94fbc-5r2xh" [9fd33765-d89a-4aca-9d83-b2ef709128c4] Running
	I0913 18:22:03.426051   11774 system_pods.go:89] "nvidia-device-plugin-daemonset-tvp28" [845bb4f2-4781-45e0-a654-5b57584b3a70] Running
	I0913 18:22:03.426057   11774 system_pods.go:89] "registry-66c9cd494c-cvtgf" [58047cca-a75a-41df-bc9f-a91dc5a547ca] Running
	I0913 18:22:03.426062   11774 system_pods.go:89] "registry-proxy-f7ltj" [7dfd72e6-65ac-495b-9832-0e900d22d7e6] Running
	I0913 18:22:03.426067   11774 system_pods.go:89] "snapshot-controller-56fcc65765-ntlvm" [41f0b45e-62eb-4d4b-9bee-ebfbe102b16e] Running
	I0913 18:22:03.426081   11774 system_pods.go:89] "snapshot-controller-56fcc65765-vsk8h" [af53978d-78a7-4c1a-960b-e203fade05e4] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
	I0913 18:22:03.426091   11774 system_pods.go:89] "storage-provisioner" [74ceb6a0-d02e-4906-9c7b-1fff3590e1d4] Running
	I0913 18:22:03.426102   11774 system_pods.go:126] duration metric: took 206.509869ms to wait for k8s-apps to be running ...
	I0913 18:22:03.426114   11774 system_svc.go:44] waiting for kubelet service to be running ....
	I0913 18:22:03.426166   11774 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0913 18:22:03.440087   11774 system_svc.go:56] duration metric: took 13.963656ms WaitForService to wait for kubelet
	I0913 18:22:03.440115   11774 kubeadm.go:582] duration metric: took 42.869940343s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
	I0913 18:22:03.440138   11774 node_conditions.go:102] verifying NodePressure condition ...
	I0913 18:22:03.619308   11774 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
	I0913 18:22:03.619339   11774 node_conditions.go:123] node cpu capacity is 8
	I0913 18:22:03.619353   11774 node_conditions.go:105] duration metric: took 179.209734ms to run NodePressure ...
	I0913 18:22:03.619365   11774 start.go:241] waiting for startup goroutines ...
	I0913 18:22:03.828211   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:03.923858   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:04.328752   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:04.428154   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:04.828836   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:04.923056   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:05.328604   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:05.428337   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:05.828362   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:05.924131   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:06.327943   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:06.428768   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:06.828271   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:06.928410   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:07.327650   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:07.423524   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:07.827283   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:07.928592   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:08.328624   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:08.423998   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:08.827766   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:08.923148   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:09.328852   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:09.423828   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:09.827795   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:09.923403   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:10.327890   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:10.423587   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:10.828165   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:10.925947   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:11.328365   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:11.423861   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:11.827639   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:11.923489   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:12.329584   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:12.423540   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:12.827696   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:12.922663   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:13.328168   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:13.423570   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:13.828623   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:13.924069   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:14.327833   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:14.428634   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:14.828009   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:14.923921   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:15.328565   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:15.428830   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:15.827459   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:15.924286   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:16.328226   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
	I0913 18:22:16.423636   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:16.827512   11774 kapi.go:107] duration metric: took 43.003884236s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
	I0913 18:22:16.923643   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:17.423081   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:17.923836   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:18.423478   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:18.923988   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:19.424151   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:19.924069   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:20.423554   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:20.923373   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:21.423731   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:21.923566   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:22.423431   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:22.923111   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:23.423436   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:23.923085   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:24.424168   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:24.923607   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:25.422931   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:25.923591   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:26.423759   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:26.923250   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:27.423759   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:27.924068   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:28.423908   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:28.923416   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:29.422743   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:29.922647   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:30.423280   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:30.924261   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:31.423189   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:31.923932   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:32.422814   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:32.923845   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:33.423434   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:33.924398   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:34.424657   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:34.924489   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:35.423970   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:35.924145   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:36.424452   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:36.923966   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:37.436679   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:37.930282   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:38.423777   11774 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
	I0913 18:22:38.923494   11774 kapi.go:107] duration metric: took 1m9.503838891s to wait for app.kubernetes.io/name=ingress-nginx ...
	I0913 18:22:57.928840   11774 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
	I0913 18:22:57.928861   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:22:58.428475   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:22:58.928031   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:22:59.429012   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:22:59.928819   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:00.428737   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:00.928541   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:01.428440   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:01.928469   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:02.428264   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:02.928355   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:03.427759   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:03.928695   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:04.428282   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:04.928335   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:05.429025   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:05.928632   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:06.428227   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:06.927992   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:07.428790   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:07.929348   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:08.428028   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:08.928472   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:09.428110   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:09.928761   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:10.428651   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:10.928577   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:11.428307   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:11.928227   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:12.428978   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:12.928837   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:13.428742   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:13.928783   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:14.428536   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:14.928413   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:15.428720   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:15.928412   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:16.427896   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:16.928521   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:17.428557   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:17.928338   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:18.427873   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:18.928684   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:19.428257   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:19.928044   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:20.428665   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:20.928561   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:21.428624   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:21.928384   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:22.428129   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:22.927855   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:23.428433   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:23.928343   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:24.428018   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:24.928947   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:25.428623   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:25.928609   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:26.428525   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:26.928478   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:27.428662   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:27.928889   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:28.428907   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:28.927816   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:29.428629   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:29.928280   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:30.427776   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:30.928748   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:31.428023   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:31.929145   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:32.429134   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:32.928563   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:33.428047   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:33.928900   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:34.428343   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:34.928140   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:35.429006   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:35.929194   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:36.429265   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:36.927940   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:37.428796   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:37.928973   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:38.428449   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:38.928252   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:39.428197   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:39.929175   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:40.428781   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:40.928698   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:41.428090   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:41.929092   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:42.428788   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:42.928666   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:43.428325   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:43.928335   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:44.428677   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:44.928449   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:45.428157   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:45.928993   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:46.429559   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:46.928449   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:47.428421   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:47.928736   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:48.428168   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:48.927912   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:49.496278   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:49.928861   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:50.428353   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:50.928045   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:51.428831   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:51.928673   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:52.428324   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:52.928192   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:53.427895   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:53.928982   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:54.428786   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:54.928836   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:55.428524   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:55.928247   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:56.427809   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:56.928832   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:57.428639   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:57.928623   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:58.428680   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:58.928297   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:59.427978   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:23:59.929111   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:24:00.428966   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:24:00.928817   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:24:01.428920   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:24:01.928713   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:24:02.429379   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:24:02.928953   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:24:03.428772   11774 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
	I0913 18:24:03.932958   11774 kapi.go:107] duration metric: took 2m28.507642144s to wait for kubernetes.io/minikube-addons=gcp-auth ...
	I0913 18:24:03.934510   11774 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-315216 cluster.
	I0913 18:24:03.935941   11774 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
	I0913 18:24:03.937252   11774 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
	I0913 18:24:03.938751   11774 out.go:177] * Enabled addons: cloud-spanner, storage-provisioner, volcano, nvidia-device-plugin, ingress-dns, inspektor-gadget, metrics-server, yakd, storage-provisioner-rancher, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
	I0913 18:24:03.940012   11774 addons.go:510] duration metric: took 2m43.369745411s for enable addons: enabled=[cloud-spanner storage-provisioner volcano nvidia-device-plugin ingress-dns inspektor-gadget metrics-server yakd storage-provisioner-rancher volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
	I0913 18:24:03.940054   11774 start.go:246] waiting for cluster config update ...
	I0913 18:24:03.940080   11774 start.go:255] writing updated cluster config ...
	I0913 18:24:03.940339   11774 ssh_runner.go:195] Run: rm -f paused
	I0913 18:24:03.987557   11774 start.go:600] kubectl: 1.31.0, cluster: 1.31.1 (minor skew: 0)
	I0913 18:24:03.989066   11774 out.go:177] * Done! kubectl is now configured to use "addons-315216" cluster and "default" namespace by default
	
	
	==> Docker <==
	Sep 13 18:33:41 addons-315216 dockerd[1338]: time="2024-09-13T18:33:41.114665735Z" level=info msg="ignoring event" container=e730bad4eddfd7fb61660e09ab8137d98aa942330f61f66d46c582668394c761 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:41 addons-315216 dockerd[1338]: time="2024-09-13T18:33:41.195305005Z" level=info msg="ignoring event" container=7c5d10e708e012586e28c5d7200eb2152c2e0a2fd57814fcecabddd014b32a48 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:41 addons-315216 dockerd[1338]: time="2024-09-13T18:33:41.196097947Z" level=info msg="ignoring event" container=cf61891d73b2104be45270a6ba0b802f886583a164958e949b07341a92218472 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:41 addons-315216 dockerd[1338]: time="2024-09-13T18:33:41.198427008Z" level=info msg="ignoring event" container=dc0098d012d2d3e6ca87eff20568cfb85c46a2523a9584158a87cdbfca02a427 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:41 addons-315216 dockerd[1338]: time="2024-09-13T18:33:41.212268222Z" level=info msg="ignoring event" container=f851c23625d4d68da50faabfd3a76baa1ae03df2e65cf5d6884150c569390476 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:41 addons-315216 cri-dockerd[1603]: time="2024-09-13T18:33:41Z" level=info msg="Failed to read pod IP from plugin/docker: networkPlugin cni failed on the status hook for pod \"csi-hostpath-attacher-0_kube-system\": unexpected command output Device \"eth0\" does not exist.\n with error: exit status 1"
	Sep 13 18:33:41 addons-315216 dockerd[1338]: time="2024-09-13T18:33:41.444842210Z" level=info msg="ignoring event" container=2d973abe7158676f97e3fb2d3334d4e481fd46dabbc7bcee62762437547831ee module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:41 addons-315216 dockerd[1338]: time="2024-09-13T18:33:41.444889078Z" level=info msg="ignoring event" container=6f5c559beae39d16452a4da942e086b4a679558a678dc38e54196eefc5fd99ac module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:41 addons-315216 dockerd[1338]: time="2024-09-13T18:33:41.494139502Z" level=info msg="ignoring event" container=78419849c7facdbe3ce7e508513b1590bc833d2df12b70e84b09470401005ed2 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:47 addons-315216 dockerd[1338]: time="2024-09-13T18:33:47.411872476Z" level=info msg="ignoring event" container=6848075bd0d508d26582ee4fb1b1c4ada191ba9ff66dd6575d248a64aac3097f module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:47 addons-315216 dockerd[1338]: time="2024-09-13T18:33:47.412646493Z" level=info msg="ignoring event" container=e2fdf049e04be2fd30424ae2c5bd79d81da97f7d3dcdfbde3fb6f66af1a327fc module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:47 addons-315216 dockerd[1338]: time="2024-09-13T18:33:47.529143660Z" level=info msg="ignoring event" container=d51d26561449424d32a70227cfd0dcb41ddcec5a2e7dd4921d3ce6c0538e6569 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:47 addons-315216 cri-dockerd[1603]: time="2024-09-13T18:33:47Z" level=info msg="Failed to read pod IP from plugin/docker: networkPlugin cni failed on the status hook for pod \"snapshot-controller-56fcc65765-vsk8h_kube-system\": unexpected command output nsenter: cannot open /proc/4895/ns/net: No such file or directory\n with error: exit status 1"
	Sep 13 18:33:47 addons-315216 dockerd[1338]: time="2024-09-13T18:33:47.560498656Z" level=info msg="ignoring event" container=7f013354760312ca2ffcd868d797a8cec827b4839470ec0992f6b03b19567672 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:50 addons-315216 dockerd[1338]: time="2024-09-13T18:33:50.646989223Z" level=info msg="Container failed to exit within 30s of signal 15 - using the force" container=3005df670852e05dfec0a608f9fb5e529fe3817074331f2268b2b0b1020d4dab
	Sep 13 18:33:50 addons-315216 dockerd[1338]: time="2024-09-13T18:33:50.666846765Z" level=info msg="ignoring event" container=3005df670852e05dfec0a608f9fb5e529fe3817074331f2268b2b0b1020d4dab module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:50 addons-315216 dockerd[1338]: time="2024-09-13T18:33:50.787351635Z" level=info msg="ignoring event" container=a00ad65edb70a6c7fc9cca7a5aeabd324dc28adfc15eb3c204908118a6438321 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:55 addons-315216 dockerd[1338]: time="2024-09-13T18:33:55.124834694Z" level=info msg="ignoring event" container=f75e3d31f40ea6f09a03d467bb467771387d6ae116de584d6d8de2aaf952e55f module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:55 addons-315216 cri-dockerd[1603]: time="2024-09-13T18:33:55Z" level=error msg="error getting RW layer size for container ID '3005df670852e05dfec0a608f9fb5e529fe3817074331f2268b2b0b1020d4dab': Error response from daemon: No such container: 3005df670852e05dfec0a608f9fb5e529fe3817074331f2268b2b0b1020d4dab"
	Sep 13 18:33:55 addons-315216 cri-dockerd[1603]: time="2024-09-13T18:33:55Z" level=error msg="Set backoffDuration to : 1m0s for container ID '3005df670852e05dfec0a608f9fb5e529fe3817074331f2268b2b0b1020d4dab'"
	Sep 13 18:33:55 addons-315216 dockerd[1338]: time="2024-09-13T18:33:55.606474918Z" level=info msg="ignoring event" container=09c283ca052c8d8f8ac7439d6bcbbef27e851b72f8c2f976ccb6214a8335f2b8 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:55 addons-315216 dockerd[1338]: time="2024-09-13T18:33:55.669783905Z" level=info msg="ignoring event" container=a9c1d39517cc60fb6d08fe5da04a0d2d8c507457c8e761d1465b185c411788ff module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:55 addons-315216 cri-dockerd[1603]: time="2024-09-13T18:33:55Z" level=info msg="Failed to read pod IP from plugin/docker: networkPlugin cni failed on the status hook for pod \"registry-66c9cd494c-cvtgf_kube-system\": unexpected command output nsenter: cannot open /proc/3808/ns/net: No such file or directory\n with error: exit status 1"
	Sep 13 18:33:55 addons-315216 dockerd[1338]: time="2024-09-13T18:33:55.751124777Z" level=info msg="ignoring event" container=ed8ef66e14f9f2445129d2a9231ef5ad8ce6310b62bc37b87b17e442c698c5bb module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	Sep 13 18:33:55 addons-315216 dockerd[1338]: time="2024-09-13T18:33:55.816006917Z" level=info msg="ignoring event" container=4228b78f9cbe6d607372551f7d4bacb06ffe8a6c97b96940a48405acdad25c6c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
	
	
	==> container status <==
	CONTAINER           IMAGE                                                                                                             CREATED             STATE               NAME                      ATTEMPT             POD ID              POD
	341cbdd33017a       a416a98b71e22                                                                                                     36 seconds ago      Exited              helper-pod                0                   e88dbbbc27eb8       helper-pod-delete-pvc-2743739b-b850-4f4a-8324-d1e883dc17e5
	b755e69b74345       busybox@sha256:c230832bd3b0be59a6c47ed64294f9ce71e91b327957920b6929a0caa8353140                                   39 seconds ago      Exited              busybox                   0                   cd6890e3c93d8       test-local-path
	6eeed28aac9c1       busybox@sha256:3fbc632167424a6d997e74f52b878d7cc478225cffac6bc977eedfe51c7f4e79                                   43 seconds ago      Exited              helper-pod                0                   c1608097129fa       helper-pod-create-pvc-2743739b-b850-4f4a-8324-d1e883dc17e5
	8f79e6d5fe2e1       kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6                       45 seconds ago      Running             hello-world-app           0                   39794026d32fc       hello-world-app-55bf9c44b4-s2vv8
	469485a0cc670       nginx@sha256:a5127daff3d6f4606be3100a252419bfa84fd6ee5cd74d0feaca1a5068f97dcf                                     53 seconds ago      Running             nginx                     0                   0da52d43730a0       nginx
	8a279ca7ee383       gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb      9 minutes ago       Running             gcp-auth                  0                   2ebf1e9d2262f       gcp-auth-89d5ffd79-2nz72
	a9c1d39517cc6       gcr.io/k8s-minikube/kube-registry-proxy@sha256:b3fa0b2df8737fdb85ad5918a7e2652527463e357afff83a5e5bb966bcedc367   12 minutes ago      Exited              registry-proxy            0                   4228b78f9cbe6       registry-proxy-f7ltj
	09c283ca052c8       registry@sha256:ac0192b549007e22998eb74e8d8488dcfe70f1489520c3b144a6047ac5efbe90                                  12 minutes ago      Exited              registry                  0                   ed8ef66e14f9f       registry-66c9cd494c-cvtgf
	a9ce8589ddeaa       6e38f40d628db                                                                                                     12 minutes ago      Running             storage-provisioner       0                   09486a9e994bb       storage-provisioner
	544a45c4cd5b9       c69fa2e9cbf5f                                                                                                     12 minutes ago      Running             coredns                   0                   1f5ef07f6fd07       coredns-7c65d6cfc9-nj2wd
	0308e52ea005b       60c005f310ff3                                                                                                     12 minutes ago      Running             kube-proxy                0                   346d1f687f17f       kube-proxy-tfw7l
	67ee7bebe548c       2e96e5913fc06                                                                                                     12 minutes ago      Running             etcd                      0                   0cc635df3d092       etcd-addons-315216
	40d4f39434b28       175ffd71cce3d                                                                                                     12 minutes ago      Running             kube-controller-manager   0                   87f22ce877b01       kube-controller-manager-addons-315216
	8a77b7004be60       9aa1fad941575                                                                                                     12 minutes ago      Running             kube-scheduler            0                   042932805c58e       kube-scheduler-addons-315216
	1513e7b25742a       6bab7719df100                                                                                                     12 minutes ago      Running             kube-apiserver            0                   7b03ca6f9b68b       kube-apiserver-addons-315216
	
	
	==> coredns [544a45c4cd5b] <==
	[INFO] 10.244.0.21:50623 - 54815 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005236642s
	[INFO] 10.244.0.21:33573 - 52588 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.005465878s
	[INFO] 10.244.0.21:59053 - 26920 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.005074627s
	[INFO] 10.244.0.21:56063 - 29935 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.001307241s
	[INFO] 10.244.0.21:55934 - 45252 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.004258563s
	[INFO] 10.244.0.21:50623 - 30256 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005750034s
	[INFO] 10.244.0.21:50623 - 39039 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000081157s
	[INFO] 10.244.0.21:33573 - 62657 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004918161s
	[INFO] 10.244.0.21:59053 - 51155 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003282374s
	[INFO] 10.244.0.21:55934 - 42925 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004312214s
	[INFO] 10.244.0.21:56063 - 61286 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005186376s
	[INFO] 10.244.0.21:56063 - 12933 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003175245s
	[INFO] 10.244.0.21:33573 - 48080 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003357526s
	[INFO] 10.244.0.21:55934 - 21435 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003594358s
	[INFO] 10.244.0.21:59053 - 22280 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.001570503s
	[INFO] 10.244.0.21:59053 - 58034 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.00006771s
	[INFO] 10.244.0.21:56063 - 595 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000158883s
	[INFO] 10.244.0.21:55934 - 49891 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000131037s
	[INFO] 10.244.0.21:33573 - 63494 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000106943s
	[INFO] 10.244.0.21:58133 - 50298 "AAAA IN hello-world-app.default.svc.cluster.local.us-east4-a.c.k8s-minikube.internal. udp 94 false 512" NXDOMAIN qr,rd,ra 94 0.004874632s
	[INFO] 10.244.0.21:58133 - 15964 "A IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.004980408s
	[INFO] 10.244.0.21:58133 - 8754 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.079609181s
	[INFO] 10.244.0.21:58133 - 23877 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004613704s
	[INFO] 10.244.0.21:58133 - 44057 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004799811s
	[INFO] 10.244.0.21:58133 - 6483 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000133819s
	
	
	==> describe nodes <==
	Name:               addons-315216
	Roles:              control-plane
	Labels:             beta.kubernetes.io/arch=amd64
	                    beta.kubernetes.io/os=linux
	                    kubernetes.io/arch=amd64
	                    kubernetes.io/hostname=addons-315216
	                    kubernetes.io/os=linux
	                    minikube.k8s.io/commit=fdd33bebc6743cfd1c61ec7fe066add478610a92
	                    minikube.k8s.io/name=addons-315216
	                    minikube.k8s.io/primary=true
	                    minikube.k8s.io/updated_at=2024_09_13T18_21_15_0700
	                    minikube.k8s.io/version=v1.34.0
	                    node-role.kubernetes.io/control-plane=
	                    node.kubernetes.io/exclude-from-external-load-balancers=
	                    topology.hostpath.csi/node=addons-315216
	Annotations:        kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
	                    node.alpha.kubernetes.io/ttl: 0
	                    volumes.kubernetes.io/controller-managed-attach-detach: true
	CreationTimestamp:  Fri, 13 Sep 2024 18:21:13 +0000
	Taints:             <none>
	Unschedulable:      false
	Lease:
	  HolderIdentity:  addons-315216
	  AcquireTime:     <unset>
	  RenewTime:       Fri, 13 Sep 2024 18:33:50 +0000
	Conditions:
	  Type             Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message
	  ----             ------  -----------------                 ------------------                ------                       -------
	  MemoryPressure   False   Fri, 13 Sep 2024 18:33:49 +0000   Fri, 13 Sep 2024 18:21:11 +0000   KubeletHasSufficientMemory   kubelet has sufficient memory available
	  DiskPressure     False   Fri, 13 Sep 2024 18:33:49 +0000   Fri, 13 Sep 2024 18:21:11 +0000   KubeletHasNoDiskPressure     kubelet has no disk pressure
	  PIDPressure      False   Fri, 13 Sep 2024 18:33:49 +0000   Fri, 13 Sep 2024 18:21:11 +0000   KubeletHasSufficientPID      kubelet has sufficient PID available
	  Ready            True    Fri, 13 Sep 2024 18:33:49 +0000   Fri, 13 Sep 2024 18:21:13 +0000   KubeletReady                 kubelet is posting ready status
	Addresses:
	  InternalIP:  192.168.49.2
	  Hostname:    addons-315216
	Capacity:
	  cpu:                8
	  ephemeral-storage:  304681132Ki
	  hugepages-1Gi:      0
	  hugepages-2Mi:      0
	  memory:             32859320Ki
	  pods:               110
	Allocatable:
	  cpu:                8
	  ephemeral-storage:  304681132Ki
	  hugepages-1Gi:      0
	  hugepages-2Mi:      0
	  memory:             32859320Ki
	  pods:               110
	System Info:
	  Machine ID:                 20bec4f7953f4f749c6214db2c7a3ff3
	  System UUID:                36576c8a-8490-4e74-9790-97effe6fd933
	  Boot ID:                    0dbba4ce-a2df-496d-8af2-30d1b719613b
	  Kernel Version:             5.15.0-1068-gcp
	  OS Image:                   Ubuntu 22.04.4 LTS
	  Operating System:           linux
	  Architecture:               amd64
	  Container Runtime Version:  docker://27.2.1
	  Kubelet Version:            v1.31.1
	  Kube-Proxy Version:         v1.31.1
	PodCIDR:                      10.244.0.0/24
	PodCIDRs:                     10.244.0.0/24
	Non-terminated Pods:          (11 in total)
	  Namespace                   Name                                     CPU Requests  CPU Limits  Memory Requests  Memory Limits  Age
	  ---------                   ----                                     ------------  ----------  ---------------  -------------  ---
	  default                     busybox                                  0 (0%)        0 (0%)      0 (0%)           0 (0%)         9m14s
	  default                     hello-world-app-55bf9c44b4-s2vv8         0 (0%)        0 (0%)      0 (0%)           0 (0%)         46s
	  default                     nginx                                    0 (0%)        0 (0%)      0 (0%)           0 (0%)         55s
	  gcp-auth                    gcp-auth-89d5ffd79-2nz72                 0 (0%)        0 (0%)      0 (0%)           0 (0%)         10m
	  kube-system                 coredns-7c65d6cfc9-nj2wd                 100m (1%)     0 (0%)      70Mi (0%)        170Mi (0%)     12m
	  kube-system                 etcd-addons-315216                       100m (1%)     0 (0%)      100Mi (0%)       0 (0%)         12m
	  kube-system                 kube-apiserver-addons-315216             250m (3%)     0 (0%)      0 (0%)           0 (0%)         12m
	  kube-system                 kube-controller-manager-addons-315216    200m (2%)     0 (0%)      0 (0%)           0 (0%)         12m
	  kube-system                 kube-proxy-tfw7l                         0 (0%)        0 (0%)      0 (0%)           0 (0%)         12m
	  kube-system                 kube-scheduler-addons-315216             100m (1%)     0 (0%)      0 (0%)           0 (0%)         12m
	  kube-system                 storage-provisioner                      0 (0%)        0 (0%)      0 (0%)           0 (0%)         12m
	Allocated resources:
	  (Total limits may be over 100 percent, i.e., overcommitted.)
	  Resource           Requests    Limits
	  --------           --------    ------
	  cpu                750m (9%)   0 (0%)
	  memory             170Mi (0%)  170Mi (0%)
	  ephemeral-storage  0 (0%)      0 (0%)
	  hugepages-1Gi      0 (0%)      0 (0%)
	  hugepages-2Mi      0 (0%)      0 (0%)
	Events:
	  Type     Reason                   Age   From             Message
	  ----     ------                   ----  ----             -------
	  Normal   Starting                 12m   kube-proxy       
	  Normal   Starting                 12m   kubelet          Starting kubelet.
	  Warning  CgroupV1                 12m   kubelet          Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
	  Normal   NodeAllocatableEnforced  12m   kubelet          Updated Node Allocatable limit across pods
	  Normal   NodeHasSufficientMemory  12m   kubelet          Node addons-315216 status is now: NodeHasSufficientMemory
	  Normal   NodeHasNoDiskPressure    12m   kubelet          Node addons-315216 status is now: NodeHasNoDiskPressure
	  Normal   NodeHasSufficientPID     12m   kubelet          Node addons-315216 status is now: NodeHasSufficientPID
	  Normal   RegisteredNode           12m   node-controller  Node addons-315216 event: Registered Node addons-315216 in Controller
	
	
	==> dmesg <==
	[  +0.000005] ll header: 00000000: ff ff ff ff ff ff 62 d0 3e db cf d6 08 06
	[  +1.429097] IPv4: martian source 10.244.0.1 from 10.244.0.17, on dev eth0
	[  +0.000006] ll header: 00000000: ff ff ff ff ff ff 92 89 50 b4 21 9e 08 06
	[  +5.154670] IPv4: martian source 10.244.0.1 from 10.244.0.20, on dev eth0
	[  +0.000005] ll header: 00000000: ff ff ff ff ff ff 26 ad 20 80 b0 8d 08 06
	[  +0.513317] IPv4: martian source 10.244.0.1 from 10.244.0.18, on dev eth0
	[  +0.000006] ll header: 00000000: ff ff ff ff ff ff 2e ba 8c be bd 7c 08 06
	[  +0.384664] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
	[  +0.000007] ll header: 00000000: ff ff ff ff ff ff ee e9 68 d1 4a cb 08 06
	[ +24.413459] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
	[  +0.000005] ll header: 00000000: ff ff ff ff ff ff 56 42 49 f1 83 41 08 06
	[  +1.014545] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
	[  +0.000006] ll header: 00000000: ff ff ff ff ff ff 82 e9 6c 93 06 21 08 06
	[Sep13 18:23] IPv4: martian source 10.244.0.1 from 10.244.0.23, on dev eth0
	[  +0.000006] ll header: 00000000: ff ff ff ff ff ff 0a bb 85 00 22 a5 08 06
	[  +0.288265] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
	[  +0.000005] ll header: 00000000: ff ff ff ff ff ff f6 b2 49 14 0c e1 08 06
	[Sep13 18:24] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
	[  +0.000010] ll header: 00000000: ff ff ff ff ff ff 86 13 49 36 5b 1b 08 06
	[  +0.000479] IPv4: martian source 10.244.0.25 from 10.244.0.2, on dev eth0
	[  +0.000004] ll header: 00000000: ff ff ff ff ff ff ea c0 87 ff 69 0f 08 06
	[Sep13 18:33] IPv4: martian source 10.244.0.30 from 10.244.0.21, on dev eth0
	[  +0.000005] ll header: 00000000: ff ff ff ff ff ff 56 42 49 f1 83 41 08 06
	[  +1.693899] IPv4: martian source 10.244.0.21 from 10.244.0.2, on dev eth0
	[  +0.000007] ll header: 00000000: ff ff ff ff ff ff ea c0 87 ff 69 0f 08 06
	
	
	==> etcd [67ee7bebe548] <==
	{"level":"info","ts":"2024-09-13T18:21:10.924151Z","caller":"embed/etcd.go:279","msg":"now serving peer/client/metrics","local-member-id":"aec36adc501070cc","initial-advertise-peer-urls":["https://192.168.49.2:2380"],"listen-peer-urls":["https://192.168.49.2:2380"],"advertise-client-urls":["https://192.168.49.2:2379"],"listen-client-urls":["https://127.0.0.1:2379","https://192.168.49.2:2379"],"listen-metrics-urls":["http://127.0.0.1:2381"]}
	{"level":"info","ts":"2024-09-13T18:21:10.924180Z","caller":"embed/etcd.go:870","msg":"serving metrics","address":"http://127.0.0.1:2381"}
	{"level":"info","ts":"2024-09-13T18:21:11.214394Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc is starting a new election at term 1"}
	{"level":"info","ts":"2024-09-13T18:21:11.214430Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became pre-candidate at term 1"}
	{"level":"info","ts":"2024-09-13T18:21:11.214444Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgPreVoteResp from aec36adc501070cc at term 1"}
	{"level":"info","ts":"2024-09-13T18:21:11.214469Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became candidate at term 2"}
	{"level":"info","ts":"2024-09-13T18:21:11.214477Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2"}
	{"level":"info","ts":"2024-09-13T18:21:11.214489Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
	{"level":"info","ts":"2024-09-13T18:21:11.214507Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
	{"level":"info","ts":"2024-09-13T18:21:11.215628Z","caller":"etcdserver/server.go:2118","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-315216 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
	{"level":"info","ts":"2024-09-13T18:21:11.215694Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
	{"level":"info","ts":"2024-09-13T18:21:11.215710Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
	{"level":"info","ts":"2024-09-13T18:21:11.215815Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
	{"level":"info","ts":"2024-09-13T18:21:11.215860Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
	{"level":"info","ts":"2024-09-13T18:21:11.215826Z","caller":"etcdserver/server.go:2629","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
	{"level":"info","ts":"2024-09-13T18:21:11.216648Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
	{"level":"info","ts":"2024-09-13T18:21:11.216723Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
	{"level":"info","ts":"2024-09-13T18:21:11.216742Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
	{"level":"info","ts":"2024-09-13T18:21:11.216776Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
	{"level":"info","ts":"2024-09-13T18:21:11.216885Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
	{"level":"info","ts":"2024-09-13T18:21:11.217730Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
	{"level":"info","ts":"2024-09-13T18:21:11.218046Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
	{"level":"info","ts":"2024-09-13T18:31:11.622452Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1863}
	{"level":"info","ts":"2024-09-13T18:31:11.644836Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1863,"took":"21.868643ms","hash":687279061,"current-db-size-bytes":9109504,"current-db-size":"9.1 MB","current-db-size-in-use-bytes":4927488,"current-db-size-in-use":"4.9 MB"}
	{"level":"info","ts":"2024-09-13T18:31:11.644875Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":687279061,"revision":1863,"compact-revision":-1}
	
	
	==> gcp-auth [8a279ca7ee38] <==
	2024/09/13 18:24:42 Ready to write response ...
	2024/09/13 18:24:42 Ready to marshal response ...
	2024/09/13 18:24:42 Ready to write response ...
	2024/09/13 18:32:45 Ready to marshal response ...
	2024/09/13 18:32:45 Ready to write response ...
	2024/09/13 18:32:45 Ready to marshal response ...
	2024/09/13 18:32:45 Ready to write response ...
	2024/09/13 18:32:45 Ready to marshal response ...
	2024/09/13 18:32:45 Ready to write response ...
	2024/09/13 18:32:55 Ready to marshal response ...
	2024/09/13 18:32:55 Ready to write response ...
	2024/09/13 18:33:01 Ready to marshal response ...
	2024/09/13 18:33:01 Ready to write response ...
	2024/09/13 18:33:02 Ready to marshal response ...
	2024/09/13 18:33:02 Ready to write response ...
	2024/09/13 18:33:10 Ready to marshal response ...
	2024/09/13 18:33:10 Ready to write response ...
	2024/09/13 18:33:11 Ready to marshal response ...
	2024/09/13 18:33:11 Ready to write response ...
	2024/09/13 18:33:11 Ready to marshal response ...
	2024/09/13 18:33:11 Ready to write response ...
	2024/09/13 18:33:20 Ready to marshal response ...
	2024/09/13 18:33:20 Ready to write response ...
	2024/09/13 18:33:32 Ready to marshal response ...
	2024/09/13 18:33:32 Ready to write response ...
	
	
	==> kernel <==
	 18:33:56 up 16 min,  0 users,  load average: 1.78, 0.70, 0.46
	Linux addons-315216 5.15.0-1068-gcp #76~20.04.1-Ubuntu SMP Tue Aug 20 15:52:45 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
	PRETTY_NAME="Ubuntu 22.04.4 LTS"
	
	
	==> kube-apiserver [1513e7b25742] <==
	W0913 18:24:34.716093       1 cacher.go:171] Terminating all watchers from cacher jobs.batch.volcano.sh
	W0913 18:24:35.007862       1 cacher.go:171] Terminating all watchers from cacher jobflows.flow.volcano.sh
	W0913 18:24:35.300810       1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
	I0913 18:32:45.440187       1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.107.207.45"}
	I0913 18:32:50.241279       1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
	W0913 18:32:51.295851       1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
	I0913 18:33:01.357605       1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
	I0913 18:33:01.517874       1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.103.74.166"}
	I0913 18:33:03.132340       1 controller.go:129] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Nothing (removed from the queue).
	I0913 18:33:10.983415       1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.100.6.12"}
	I0913 18:33:10.992098       1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
	E0913 18:33:36.126991       1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
	I0913 18:33:47.284129       1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
	I0913 18:33:47.284184       1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
	I0913 18:33:47.296075       1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
	I0913 18:33:47.296122       1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
	I0913 18:33:47.297704       1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
	I0913 18:33:47.297746       1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
	I0913 18:33:47.307940       1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
	I0913 18:33:47.307985       1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
	I0913 18:33:47.318327       1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
	I0913 18:33:47.318364       1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
	W0913 18:33:48.298691       1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
	W0913 18:33:48.319299       1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
	W0913 18:33:48.428542       1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
	
	
	==> kube-controller-manager [40d4f39434b2] <==
	E0913 18:33:48.299997       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
	E0913 18:33:48.320381       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
	E0913 18:33:48.429768       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
	W0913 18:33:49.256808       1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
	E0913 18:33:49.256849       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
	W0913 18:33:49.525874       1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
	E0913 18:33:49.525922       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
	I0913 18:33:49.555479       1 range_allocator.go:241] "Successfully synced" logger="node-ipam-controller" key="addons-315216"
	W0913 18:33:49.862563       1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
	E0913 18:33:49.862598       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
	I0913 18:33:50.508427       1 shared_informer.go:313] Waiting for caches to sync for resource quota
	I0913 18:33:50.508468       1 shared_informer.go:320] Caches are synced for resource quota
	W0913 18:33:50.530337       1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
	E0913 18:33:50.530372       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
	I0913 18:33:50.701794       1 shared_informer.go:313] Waiting for caches to sync for garbage collector
	I0913 18:33:50.701842       1 shared_informer.go:320] Caches are synced for garbage collector
	W0913 18:33:52.015861       1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
	E0913 18:33:52.015899       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
	W0913 18:33:52.076499       1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
	E0913 18:33:52.076533       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
	W0913 18:33:52.100825       1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
	E0913 18:33:52.100854       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
	W0913 18:33:52.110975       1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
	E0913 18:33:52.111002       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
	I0913 18:33:55.550003       1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-66c9cd494c" duration="7.918µs"
	
	
	==> kube-proxy [0308e52ea005] <==
	I0913 18:21:24.217785       1 server_linux.go:66] "Using iptables proxy"
	I0913 18:21:24.706405       1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
	E0913 18:21:24.706483       1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
	I0913 18:21:24.909031       1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
	I0913 18:21:24.909089       1 server_linux.go:169] "Using iptables Proxier"
	I0913 18:21:24.999429       1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
	I0913 18:21:24.999841       1 server.go:483] "Version info" version="v1.31.1"
	I0913 18:21:24.999866       1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
	I0913 18:21:25.003344       1 config.go:199] "Starting service config controller"
	I0913 18:21:25.003382       1 shared_informer.go:313] Waiting for caches to sync for service config
	I0913 18:21:25.003415       1 config.go:105] "Starting endpoint slice config controller"
	I0913 18:21:25.003421       1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
	I0913 18:21:25.003892       1 config.go:328] "Starting node config controller"
	I0913 18:21:25.003903       1 shared_informer.go:313] Waiting for caches to sync for node config
	I0913 18:21:25.105363       1 shared_informer.go:320] Caches are synced for node config
	I0913 18:21:25.105413       1 shared_informer.go:320] Caches are synced for service config
	I0913 18:21:25.105444       1 shared_informer.go:320] Caches are synced for endpoint slice config
	
	
	==> kube-scheduler [8a77b7004be6] <==
	E0913 18:21:13.112176       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
	W0913 18:21:13.111157       1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
	E0913 18:21:13.112194       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
	E0913 18:21:13.112278       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
	W0913 18:21:13.111251       1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
	E0913 18:21:13.112312       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
	W0913 18:21:13.111270       1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
	E0913 18:21:13.112349       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
	W0913 18:21:13.111282       1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
	E0913 18:21:13.112226       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
	E0913 18:21:13.112385       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
	E0913 18:21:13.112282       1 reflector.go:158] "Unhandled Error" err="runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
	W0913 18:21:13.925807       1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
	E0913 18:21:13.925845       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
	W0913 18:21:14.042099       1 reflector.go:561] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
	E0913 18:21:14.042348       1 reflector.go:158] "Unhandled Error" err="runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
	W0913 18:21:14.093144       1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
	E0913 18:21:14.093187       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
	W0913 18:21:14.137282       1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
	E0913 18:21:14.137316       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
	W0913 18:21:14.144344       1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
	E0913 18:21:14.144384       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
	W0913 18:21:14.251764       1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
	E0913 18:21:14.251812       1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
	I0913 18:21:16.509959       1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
	
	
	==> kubelet <==
	Sep 13 18:33:49 addons-315216 kubelet[2446]: I0913 18:33:49.225520    2446 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af53978d-78a7-4c1a-960b-e203fade05e4" path="/var/lib/kubelet/pods/af53978d-78a7-4c1a-960b-e203fade05e4/volumes"
	Sep 13 18:33:50 addons-315216 kubelet[2446]: E0913 18:33:50.219972    2446 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="cc20ad18-6fce-408c-8205-eb67a03c9a44"
	Sep 13 18:33:50 addons-315216 kubelet[2446]: I0913 18:33:50.959534    2446 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd053b82-587a-4124-a7ff-31feeca967dc-config-volume\") pod \"dd053b82-587a-4124-a7ff-31feeca967dc\" (UID: \"dd053b82-587a-4124-a7ff-31feeca967dc\") "
	Sep 13 18:33:50 addons-315216 kubelet[2446]: I0913 18:33:50.959585    2446 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79fxs\" (UniqueName: \"kubernetes.io/projected/dd053b82-587a-4124-a7ff-31feeca967dc-kube-api-access-79fxs\") pod \"dd053b82-587a-4124-a7ff-31feeca967dc\" (UID: \"dd053b82-587a-4124-a7ff-31feeca967dc\") "
	Sep 13 18:33:50 addons-315216 kubelet[2446]: I0913 18:33:50.959953    2446 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd053b82-587a-4124-a7ff-31feeca967dc-config-volume" (OuterVolumeSpecName: "config-volume") pod "dd053b82-587a-4124-a7ff-31feeca967dc" (UID: "dd053b82-587a-4124-a7ff-31feeca967dc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue ""
	Sep 13 18:33:50 addons-315216 kubelet[2446]: I0913 18:33:50.961271    2446 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd053b82-587a-4124-a7ff-31feeca967dc-kube-api-access-79fxs" (OuterVolumeSpecName: "kube-api-access-79fxs") pod "dd053b82-587a-4124-a7ff-31feeca967dc" (UID: "dd053b82-587a-4124-a7ff-31feeca967dc"). InnerVolumeSpecName "kube-api-access-79fxs". PluginName "kubernetes.io/projected", VolumeGidValue ""
	Sep 13 18:33:51 addons-315216 kubelet[2446]: I0913 18:33:51.060630    2446 reconciler_common.go:288] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd053b82-587a-4124-a7ff-31feeca967dc-config-volume\") on node \"addons-315216\" DevicePath \"\""
	Sep 13 18:33:51 addons-315216 kubelet[2446]: I0913 18:33:51.060663    2446 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-79fxs\" (UniqueName: \"kubernetes.io/projected/dd053b82-587a-4124-a7ff-31feeca967dc-kube-api-access-79fxs\") on node \"addons-315216\" DevicePath \"\""
	Sep 13 18:33:51 addons-315216 kubelet[2446]: I0913 18:33:51.648876    2446 scope.go:117] "RemoveContainer" containerID="3005df670852e05dfec0a608f9fb5e529fe3817074331f2268b2b0b1020d4dab"
	Sep 13 18:33:51 addons-315216 kubelet[2446]: I0913 18:33:51.662575    2446 scope.go:117] "RemoveContainer" containerID="3005df670852e05dfec0a608f9fb5e529fe3817074331f2268b2b0b1020d4dab"
	Sep 13 18:33:51 addons-315216 kubelet[2446]: E0913 18:33:51.663209    2446 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 3005df670852e05dfec0a608f9fb5e529fe3817074331f2268b2b0b1020d4dab" containerID="3005df670852e05dfec0a608f9fb5e529fe3817074331f2268b2b0b1020d4dab"
	Sep 13 18:33:51 addons-315216 kubelet[2446]: I0913 18:33:51.663241    2446 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"3005df670852e05dfec0a608f9fb5e529fe3817074331f2268b2b0b1020d4dab"} err="failed to get container status \"3005df670852e05dfec0a608f9fb5e529fe3817074331f2268b2b0b1020d4dab\": rpc error: code = Unknown desc = Error response from daemon: No such container: 3005df670852e05dfec0a608f9fb5e529fe3817074331f2268b2b0b1020d4dab"
	Sep 13 18:33:53 addons-315216 kubelet[2446]: I0913 18:33:53.224596    2446 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd053b82-587a-4124-a7ff-31feeca967dc" path="/var/lib/kubelet/pods/dd053b82-587a-4124-a7ff-31feeca967dc/volumes"
	Sep 13 18:33:55 addons-315216 kubelet[2446]: I0913 18:33:55.285434    2446 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8cpc\" (UniqueName: \"kubernetes.io/projected/e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5-kube-api-access-m8cpc\") pod \"e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5\" (UID: \"e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5\") "
	Sep 13 18:33:55 addons-315216 kubelet[2446]: I0913 18:33:55.285483    2446 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5-gcp-creds\") pod \"e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5\" (UID: \"e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5\") "
	Sep 13 18:33:55 addons-315216 kubelet[2446]: I0913 18:33:55.285555    2446 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5" (UID: "e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
	Sep 13 18:33:55 addons-315216 kubelet[2446]: I0913 18:33:55.287088    2446 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5-kube-api-access-m8cpc" (OuterVolumeSpecName: "kube-api-access-m8cpc") pod "e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5" (UID: "e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5"). InnerVolumeSpecName "kube-api-access-m8cpc". PluginName "kubernetes.io/projected", VolumeGidValue ""
	Sep 13 18:33:55 addons-315216 kubelet[2446]: I0913 18:33:55.386331    2446 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-m8cpc\" (UniqueName: \"kubernetes.io/projected/e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5-kube-api-access-m8cpc\") on node \"addons-315216\" DevicePath \"\""
	Sep 13 18:33:55 addons-315216 kubelet[2446]: I0913 18:33:55.386364    2446 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/e8a2ab5b-f9f4-47b1-9797-7023fd89a3d5-gcp-creds\") on node \"addons-315216\" DevicePath \"\""
	Sep 13 18:33:55 addons-315216 kubelet[2446]: I0913 18:33:55.995829    2446 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nl4r\" (UniqueName: \"kubernetes.io/projected/58047cca-a75a-41df-bc9f-a91dc5a547ca-kube-api-access-9nl4r\") pod \"58047cca-a75a-41df-bc9f-a91dc5a547ca\" (UID: \"58047cca-a75a-41df-bc9f-a91dc5a547ca\") "
	Sep 13 18:33:55 addons-315216 kubelet[2446]: I0913 18:33:55.995893    2446 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xddr8\" (UniqueName: \"kubernetes.io/projected/7dfd72e6-65ac-495b-9832-0e900d22d7e6-kube-api-access-xddr8\") pod \"7dfd72e6-65ac-495b-9832-0e900d22d7e6\" (UID: \"7dfd72e6-65ac-495b-9832-0e900d22d7e6\") "
	Sep 13 18:33:55 addons-315216 kubelet[2446]: I0913 18:33:55.997944    2446 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dfd72e6-65ac-495b-9832-0e900d22d7e6-kube-api-access-xddr8" (OuterVolumeSpecName: "kube-api-access-xddr8") pod "7dfd72e6-65ac-495b-9832-0e900d22d7e6" (UID: "7dfd72e6-65ac-495b-9832-0e900d22d7e6"). InnerVolumeSpecName "kube-api-access-xddr8". PluginName "kubernetes.io/projected", VolumeGidValue ""
	Sep 13 18:33:55 addons-315216 kubelet[2446]: I0913 18:33:55.997939    2446 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58047cca-a75a-41df-bc9f-a91dc5a547ca-kube-api-access-9nl4r" (OuterVolumeSpecName: "kube-api-access-9nl4r") pod "58047cca-a75a-41df-bc9f-a91dc5a547ca" (UID: "58047cca-a75a-41df-bc9f-a91dc5a547ca"). InnerVolumeSpecName "kube-api-access-9nl4r". PluginName "kubernetes.io/projected", VolumeGidValue ""
	Sep 13 18:33:56 addons-315216 kubelet[2446]: I0913 18:33:56.096483    2446 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-9nl4r\" (UniqueName: \"kubernetes.io/projected/58047cca-a75a-41df-bc9f-a91dc5a547ca-kube-api-access-9nl4r\") on node \"addons-315216\" DevicePath \"\""
	Sep 13 18:33:56 addons-315216 kubelet[2446]: I0913 18:33:56.096533    2446 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-xddr8\" (UniqueName: \"kubernetes.io/projected/7dfd72e6-65ac-495b-9832-0e900d22d7e6-kube-api-access-xddr8\") on node \"addons-315216\" DevicePath \"\""
	
	
	==> storage-provisioner [a9ce8589ddea] <==
	I0913 18:21:28.807206       1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
	I0913 18:21:28.894610       1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
	I0913 18:21:28.894661       1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
	I0913 18:21:28.907515       1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
	I0913 18:21:28.908782       1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"9573cd56-4073-4047-8229-80fa85ec9d93", APIVersion:"v1", ResourceVersion:"619", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-315216_7ed8fcb8-4ad4-4bb5-a9dc-d3c6fa05c16a became leader
	I0913 18:21:28.908976       1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-315216_7ed8fcb8-4ad4-4bb5-a9dc-d3c6fa05c16a!
	I0913 18:21:29.010094       1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-315216_7ed8fcb8-4ad4-4bb5-a9dc-d3c6fa05c16a!
	

                                                
                                                
-- /stdout --
helpers_test.go:254: (dbg) Run:  out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-315216 -n addons-315216
helpers_test.go:261: (dbg) Run:  kubectl --context addons-315216 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run:  kubectl --context addons-315216 describe pod busybox
helpers_test.go:282: (dbg) kubectl --context addons-315216 describe pod busybox:

                                                
                                                
-- stdout --
	Name:             busybox
	Namespace:        default
	Priority:         0
	Service Account:  default
	Node:             addons-315216/192.168.49.2
	Start Time:       Fri, 13 Sep 2024 18:24:42 +0000
	Labels:           integration-test=busybox
	Annotations:      <none>
	Status:           Pending
	IP:               10.244.0.27
	IPs:
	  IP:  10.244.0.27
	Containers:
	  busybox:
	    Container ID:  
	    Image:         gcr.io/k8s-minikube/busybox:1.28.4-glibc
	    Image ID:      
	    Port:          <none>
	    Host Port:     <none>
	    Command:
	      sleep
	      3600
	    State:          Waiting
	      Reason:       ImagePullBackOff
	    Ready:          False
	    Restart Count:  0
	    Environment:
	      GOOGLE_APPLICATION_CREDENTIALS:  /google-app-creds.json
	      PROJECT_ID:                      this_is_fake
	      GCP_PROJECT:                     this_is_fake
	      GCLOUD_PROJECT:                  this_is_fake
	      GOOGLE_CLOUD_PROJECT:            this_is_fake
	      CLOUDSDK_CORE_PROJECT:           this_is_fake
	    Mounts:
	      /google-app-creds.json from gcp-creds (ro)
	      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-xwmf7 (ro)
	Conditions:
	  Type                        Status
	  PodReadyToStartContainers   True 
	  Initialized                 True 
	  Ready                       False 
	  ContainersReady             False 
	  PodScheduled                True 
	Volumes:
	  kube-api-access-xwmf7:
	    Type:                    Projected (a volume that contains injected data from multiple sources)
	    TokenExpirationSeconds:  3607
	    ConfigMapName:           kube-root-ca.crt
	    ConfigMapOptional:       <nil>
	    DownwardAPI:             true
	  gcp-creds:
	    Type:          HostPath (bare host directory volume)
	    Path:          /var/lib/minikube/google_application_credentials.json
	    HostPathType:  File
	QoS Class:         BestEffort
	Node-Selectors:    <none>
	Tolerations:       node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
	                   node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
	Events:
	  Type     Reason     Age                    From               Message
	  ----     ------     ----                   ----               -------
	  Normal   Scheduled  9m15s                  default-scheduler  Successfully assigned default/busybox to addons-315216
	  Normal   Pulling    7m46s (x4 over 9m14s)  kubelet            Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
	  Warning  Failed     7m46s (x4 over 9m14s)  kubelet            Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
	  Warning  Failed     7m46s (x4 over 9m14s)  kubelet            Error: ErrImagePull
	  Warning  Failed     7m31s (x6 over 9m13s)  kubelet            Error: ImagePullBackOff
	  Normal   BackOff    4m2s (x21 over 9m13s)  kubelet            Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"

                                                
                                                
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (72.37s)

                                                
                                    

Test pass (321/342)

Order passed test Duration
3 TestDownloadOnly/v1.20.0/json-events 4.53
4 TestDownloadOnly/v1.20.0/preload-exists 0
8 TestDownloadOnly/v1.20.0/LogsDuration 0.05
9 TestDownloadOnly/v1.20.0/DeleteAll 0.18
10 TestDownloadOnly/v1.20.0/DeleteAlwaysSucceeds 0.12
12 TestDownloadOnly/v1.31.1/json-events 4.26
13 TestDownloadOnly/v1.31.1/preload-exists 0
17 TestDownloadOnly/v1.31.1/LogsDuration 0.05
18 TestDownloadOnly/v1.31.1/DeleteAll 0.18
19 TestDownloadOnly/v1.31.1/DeleteAlwaysSucceeds 0.11
20 TestDownloadOnlyKic 0.94
21 TestBinaryMirror 0.72
22 TestOffline 55.25
25 TestAddons/PreSetup/EnablingAddonOnNonExistingCluster 0.05
26 TestAddons/PreSetup/DisablingAddonOnNonExistingCluster 0.05
27 TestAddons/Setup 206.41
29 TestAddons/serial/Volcano 38.56
31 TestAddons/serial/GCPAuth/Namespaces 0.11
34 TestAddons/parallel/Ingress 19.4
35 TestAddons/parallel/InspektorGadget 10.6
36 TestAddons/parallel/MetricsServer 5.77
38 TestAddons/parallel/CSI 57.02
39 TestAddons/parallel/Headlamp 16.26
40 TestAddons/parallel/CloudSpanner 5.4
41 TestAddons/parallel/LocalPath 51.61
42 TestAddons/parallel/NvidiaDevicePlugin 5.37
43 TestAddons/parallel/Yakd 10.6
44 TestAddons/StoppedEnableDisable 10.91
45 TestCertOptions 31.02
46 TestCertExpiration 232.22
47 TestDockerFlags 26.8
48 TestForceSystemdFlag 40.87
49 TestForceSystemdEnv 28.16
51 TestKVMDriverInstallOrUpdate 1.17
55 TestErrorSpam/setup 20.01
56 TestErrorSpam/start 0.52
57 TestErrorSpam/status 0.82
58 TestErrorSpam/pause 1.1
59 TestErrorSpam/unpause 1.16
60 TestErrorSpam/stop 1.86
63 TestFunctional/serial/CopySyncFile 0
64 TestFunctional/serial/StartWithProxy 67.27
65 TestFunctional/serial/AuditLog 0
66 TestFunctional/serial/SoftStart 32.63
67 TestFunctional/serial/KubeContext 0.04
68 TestFunctional/serial/KubectlGetPods 0.06
71 TestFunctional/serial/CacheCmd/cache/add_remote 2.29
72 TestFunctional/serial/CacheCmd/cache/add_local 0.65
73 TestFunctional/serial/CacheCmd/cache/CacheDelete 0.04
74 TestFunctional/serial/CacheCmd/cache/list 0.04
75 TestFunctional/serial/CacheCmd/cache/verify_cache_inside_node 0.26
76 TestFunctional/serial/CacheCmd/cache/cache_reload 1.18
77 TestFunctional/serial/CacheCmd/cache/delete 0.08
78 TestFunctional/serial/MinikubeKubectlCmd 0.1
79 TestFunctional/serial/MinikubeKubectlCmdDirectly 0.1
80 TestFunctional/serial/ExtraConfig 36.75
81 TestFunctional/serial/ComponentHealth 0.06
82 TestFunctional/serial/LogsCmd 0.86
83 TestFunctional/serial/LogsFileCmd 0.9
84 TestFunctional/serial/InvalidService 4.43
86 TestFunctional/parallel/ConfigCmd 0.34
87 TestFunctional/parallel/DashboardCmd 8.55
88 TestFunctional/parallel/DryRun 0.43
89 TestFunctional/parallel/InternationalLanguage 0.17
90 TestFunctional/parallel/StatusCmd 1.13
94 TestFunctional/parallel/ServiceCmdConnect 8.71
95 TestFunctional/parallel/AddonsCmd 0.13
96 TestFunctional/parallel/PersistentVolumeClaim 29.55
98 TestFunctional/parallel/SSHCmd 0.58
99 TestFunctional/parallel/CpCmd 1.77
100 TestFunctional/parallel/MySQL 20.79
101 TestFunctional/parallel/FileSync 0.28
102 TestFunctional/parallel/CertSync 1.76
106 TestFunctional/parallel/NodeLabels 0.08
108 TestFunctional/parallel/NonActiveRuntimeDisabled 0.28
110 TestFunctional/parallel/License 0.17
111 TestFunctional/parallel/ServiceCmd/DeployApp 10.19
112 TestFunctional/parallel/ProfileCmd/profile_not_create 0.43
113 TestFunctional/parallel/MountCmd/any-port 7.82
114 TestFunctional/parallel/ProfileCmd/profile_list 0.4
115 TestFunctional/parallel/ProfileCmd/profile_json_output 0.36
116 TestFunctional/parallel/MountCmd/specific-port 1.79
117 TestFunctional/parallel/MountCmd/VerifyCleanup 2.12
118 TestFunctional/parallel/ServiceCmd/List 0.61
119 TestFunctional/parallel/ServiceCmd/JSONOutput 0.51
120 TestFunctional/parallel/ServiceCmd/HTTPS 0.37
122 TestFunctional/parallel/TunnelCmd/serial/RunSecondTunnel 0.46
123 TestFunctional/parallel/ServiceCmd/Format 0.41
124 TestFunctional/parallel/TunnelCmd/serial/StartTunnel 0
126 TestFunctional/parallel/TunnelCmd/serial/WaitService/Setup 9.26
127 TestFunctional/parallel/ServiceCmd/URL 0.56
128 TestFunctional/parallel/DockerEnv/bash 1.15
129 TestFunctional/parallel/UpdateContextCmd/no_changes 0.15
130 TestFunctional/parallel/UpdateContextCmd/no_minikube_cluster 0.14
131 TestFunctional/parallel/UpdateContextCmd/no_clusters 0.14
132 TestFunctional/parallel/Version/short 0.06
133 TestFunctional/parallel/Version/components 0.72
134 TestFunctional/parallel/TunnelCmd/serial/WaitService/IngressIP 0.06
135 TestFunctional/parallel/TunnelCmd/serial/AccessDirect 0
139 TestFunctional/parallel/TunnelCmd/serial/DeleteTunnel 0.11
140 TestFunctional/parallel/ImageCommands/ImageListShort 0.2
141 TestFunctional/parallel/ImageCommands/ImageListTable 0.19
142 TestFunctional/parallel/ImageCommands/ImageListJson 0.19
143 TestFunctional/parallel/ImageCommands/ImageListYaml 0.2
144 TestFunctional/parallel/ImageCommands/ImageBuild 2.34
145 TestFunctional/parallel/ImageCommands/Setup 0.96
146 TestFunctional/parallel/ImageCommands/ImageLoadDaemon 1.07
147 TestFunctional/parallel/ImageCommands/ImageReloadDaemon 0.91
148 TestFunctional/parallel/ImageCommands/ImageTagAndLoadDaemon 0.96
149 TestFunctional/parallel/ImageCommands/ImageSaveToFile 0.28
150 TestFunctional/parallel/ImageCommands/ImageRemove 0.45
151 TestFunctional/parallel/ImageCommands/ImageLoadFromFile 0.53
152 TestFunctional/parallel/ImageCommands/ImageSaveDaemon 0.33
153 TestFunctional/delete_echo-server_images 0.03
154 TestFunctional/delete_my-image_image 0.01
155 TestFunctional/delete_minikube_cached_images 0.01
159 TestMultiControlPlane/serial/StartCluster 100.37
160 TestMultiControlPlane/serial/DeployApp 39.24
161 TestMultiControlPlane/serial/PingHostFromPods 1.02
162 TestMultiControlPlane/serial/AddWorkerNode 20.56
163 TestMultiControlPlane/serial/NodeLabels 0.06
164 TestMultiControlPlane/serial/HAppyAfterClusterStart 0.58
165 TestMultiControlPlane/serial/CopyFile 14.93
166 TestMultiControlPlane/serial/StopSecondaryNode 11.23
167 TestMultiControlPlane/serial/DegradedAfterControlPlaneNodeStop 0.46
168 TestMultiControlPlane/serial/RestartSecondaryNode 18.29
169 TestMultiControlPlane/serial/HAppyAfterSecondaryNodeRestart 2.85
170 TestMultiControlPlane/serial/RestartClusterKeepsNodes 253.71
171 TestMultiControlPlane/serial/DeleteSecondaryNode 9.11
172 TestMultiControlPlane/serial/DegradedAfterSecondaryNodeDelete 0.44
173 TestMultiControlPlane/serial/StopCluster 32.47
174 TestMultiControlPlane/serial/RestartCluster 81.09
175 TestMultiControlPlane/serial/DegradedAfterClusterRestart 0.44
176 TestMultiControlPlane/serial/AddSecondaryNode 39.53
177 TestMultiControlPlane/serial/HAppyAfterSecondaryNodeAdd 0.6
180 TestImageBuild/serial/Setup 19.35
181 TestImageBuild/serial/NormalBuild 1.13
182 TestImageBuild/serial/BuildWithBuildArg 0.7
183 TestImageBuild/serial/BuildWithDockerIgnore 0.49
184 TestImageBuild/serial/BuildWithSpecifiedDockerfile 0.51
188 TestJSONOutput/start/Command 65.12
189 TestJSONOutput/start/Audit 0
191 TestJSONOutput/start/parallel/DistinctCurrentSteps 0
192 TestJSONOutput/start/parallel/IncreasingCurrentSteps 0
194 TestJSONOutput/pause/Command 0.52
195 TestJSONOutput/pause/Audit 0
197 TestJSONOutput/pause/parallel/DistinctCurrentSteps 0
198 TestJSONOutput/pause/parallel/IncreasingCurrentSteps 0
200 TestJSONOutput/unpause/Command 0.41
201 TestJSONOutput/unpause/Audit 0
203 TestJSONOutput/unpause/parallel/DistinctCurrentSteps 0
204 TestJSONOutput/unpause/parallel/IncreasingCurrentSteps 0
206 TestJSONOutput/stop/Command 10.78
207 TestJSONOutput/stop/Audit 0
209 TestJSONOutput/stop/parallel/DistinctCurrentSteps 0
210 TestJSONOutput/stop/parallel/IncreasingCurrentSteps 0
211 TestErrorJSONOutput 0.19
213 TestKicCustomNetwork/create_custom_network 22.58
214 TestKicCustomNetwork/use_default_bridge_network 24.79
215 TestKicExistingNetwork 22.02
216 TestKicCustomSubnet 25.34
217 TestKicStaticIP 25.46
218 TestMainNoArgs 0.04
219 TestMinikubeProfile 52.08
222 TestMountStart/serial/StartWithMountFirst 6.09
223 TestMountStart/serial/VerifyMountFirst 0.23
224 TestMountStart/serial/StartWithMountSecond 9.07
225 TestMountStart/serial/VerifyMountSecond 0.23
226 TestMountStart/serial/DeleteFirst 1.43
227 TestMountStart/serial/VerifyMountPostDelete 0.23
228 TestMountStart/serial/Stop 1.16
229 TestMountStart/serial/RestartStopped 7.44
230 TestMountStart/serial/VerifyMountPostStop 0.23
233 TestMultiNode/serial/FreshStart2Nodes 68.98
234 TestMultiNode/serial/DeployApp2Nodes 38.37
235 TestMultiNode/serial/PingHostFrom2Pods 0.69
236 TestMultiNode/serial/AddNode 18.43
237 TestMultiNode/serial/MultiNodeLabels 0.06
238 TestMultiNode/serial/ProfileList 0.27
239 TestMultiNode/serial/CopyFile 8.6
240 TestMultiNode/serial/StopNode 2.03
241 TestMultiNode/serial/StartAfterStop 9.53
242 TestMultiNode/serial/RestartKeepsNodes 97.9
243 TestMultiNode/serial/DeleteNode 5.04
244 TestMultiNode/serial/StopMultiNode 21.3
245 TestMultiNode/serial/RestartMultiNode 52.67
246 TestMultiNode/serial/ValidateNameConflict 22.82
251 TestPreload 85.82
253 TestScheduledStopUnix 96.46
254 TestSkaffold 93.61
256 TestInsufficientStorage 12.51
257 TestRunningBinaryUpgrade 67.12
259 TestKubernetesUpgrade 332.95
260 TestMissingContainerUpgrade 93.48
261 TestStoppedBinaryUpgrade/Setup 0.61
263 TestNoKubernetes/serial/StartNoK8sWithVersion 0.07
264 TestNoKubernetes/serial/StartWithK8s 38.82
265 TestStoppedBinaryUpgrade/Upgrade 109.87
266 TestNoKubernetes/serial/StartWithStopK8s 9.87
278 TestNoKubernetes/serial/Start 10.58
279 TestNoKubernetes/serial/VerifyK8sNotRunning 0.27
280 TestNoKubernetes/serial/ProfileList 1.27
281 TestNoKubernetes/serial/Stop 1.19
282 TestNoKubernetes/serial/StartNoArgs 7.29
283 TestNoKubernetes/serial/VerifyK8sNotRunningSecond 0.26
291 TestStoppedBinaryUpgrade/MinikubeLogs 1.13
293 TestPause/serial/Start 65.57
294 TestPause/serial/SecondStartNoReconfiguration 34.86
295 TestNetworkPlugins/group/auto/Start 64.17
296 TestPause/serial/Pause 0.53
297 TestPause/serial/VerifyStatus 0.28
298 TestPause/serial/Unpause 0.43
299 TestPause/serial/PauseAgain 0.6
300 TestPause/serial/DeletePaused 2.01
301 TestPause/serial/VerifyDeletedResources 14.76
302 TestNetworkPlugins/group/kindnet/Start 40.86
303 TestNetworkPlugins/group/auto/KubeletFlags 0.31
304 TestNetworkPlugins/group/auto/NetCatPod 9.23
305 TestNetworkPlugins/group/auto/DNS 0.18
306 TestNetworkPlugins/group/auto/Localhost 0.23
307 TestNetworkPlugins/group/auto/HairPin 0.14
308 TestNetworkPlugins/group/calico/Start 55.96
309 TestNetworkPlugins/group/kindnet/ControllerPod 6.01
310 TestNetworkPlugins/group/kindnet/KubeletFlags 0.29
311 TestNetworkPlugins/group/kindnet/NetCatPod 9.23
312 TestNetworkPlugins/group/custom-flannel/Start 47.17
313 TestNetworkPlugins/group/kindnet/DNS 0.14
314 TestNetworkPlugins/group/kindnet/Localhost 0.13
315 TestNetworkPlugins/group/kindnet/HairPin 0.12
316 TestNetworkPlugins/group/false/Start 66.04
317 TestNetworkPlugins/group/calico/ControllerPod 6.01
318 TestNetworkPlugins/group/custom-flannel/KubeletFlags 0.26
319 TestNetworkPlugins/group/custom-flannel/NetCatPod 9.19
320 TestNetworkPlugins/group/calico/KubeletFlags 0.3
321 TestNetworkPlugins/group/calico/NetCatPod 10.2
322 TestNetworkPlugins/group/custom-flannel/DNS 0.14
323 TestNetworkPlugins/group/custom-flannel/Localhost 0.11
324 TestNetworkPlugins/group/custom-flannel/HairPin 0.12
325 TestNetworkPlugins/group/calico/DNS 0.12
326 TestNetworkPlugins/group/calico/Localhost 0.11
327 TestNetworkPlugins/group/calico/HairPin 0.1
328 TestNetworkPlugins/group/enable-default-cni/Start 40.86
329 TestNetworkPlugins/group/flannel/Start 42.15
330 TestNetworkPlugins/group/false/KubeletFlags 0.26
331 TestNetworkPlugins/group/false/NetCatPod 10.2
332 TestNetworkPlugins/group/false/DNS 0.17
333 TestNetworkPlugins/group/false/Localhost 0.11
334 TestNetworkPlugins/group/false/HairPin 0.11
335 TestNetworkPlugins/group/enable-default-cni/KubeletFlags 0.31
336 TestNetworkPlugins/group/enable-default-cni/NetCatPod 10.27
337 TestNetworkPlugins/group/bridge/Start 66.67
338 TestNetworkPlugins/group/flannel/ControllerPod 6.01
339 TestNetworkPlugins/group/enable-default-cni/DNS 0.13
340 TestNetworkPlugins/group/enable-default-cni/Localhost 0.1
341 TestNetworkPlugins/group/enable-default-cni/HairPin 0.11
342 TestNetworkPlugins/group/flannel/KubeletFlags 0.24
343 TestNetworkPlugins/group/flannel/NetCatPod 9.17
344 TestNetworkPlugins/group/flannel/DNS 0.21
345 TestNetworkPlugins/group/flannel/Localhost 0.18
346 TestNetworkPlugins/group/flannel/HairPin 0.15
347 TestNetworkPlugins/group/kubenet/Start 65.63
349 TestStartStop/group/old-k8s-version/serial/FirstStart 151.1
351 TestStartStop/group/no-preload/serial/FirstStart 70.09
352 TestNetworkPlugins/group/bridge/KubeletFlags 0.29
353 TestNetworkPlugins/group/bridge/NetCatPod 11.21
354 TestNetworkPlugins/group/bridge/DNS 0.13
355 TestNetworkPlugins/group/bridge/Localhost 0.1
356 TestNetworkPlugins/group/bridge/HairPin 0.11
357 TestNetworkPlugins/group/kubenet/KubeletFlags 0.27
358 TestNetworkPlugins/group/kubenet/NetCatPod 10.22
360 TestStartStop/group/embed-certs/serial/FirstStart 68.73
361 TestNetworkPlugins/group/kubenet/DNS 0.13
362 TestNetworkPlugins/group/kubenet/Localhost 0.1
363 TestNetworkPlugins/group/kubenet/HairPin 0.1
365 TestStartStop/group/default-k8s-diff-port/serial/FirstStart 64.28
366 TestStartStop/group/no-preload/serial/DeployApp 9.3
367 TestStartStop/group/no-preload/serial/EnableAddonWhileActive 0.8
368 TestStartStop/group/no-preload/serial/Stop 10.72
369 TestStartStop/group/no-preload/serial/EnableAddonAfterStop 0.23
370 TestStartStop/group/no-preload/serial/SecondStart 263.04
371 TestStartStop/group/embed-certs/serial/DeployApp 9.25
372 TestStartStop/group/embed-certs/serial/EnableAddonWhileActive 0.82
373 TestStartStop/group/embed-certs/serial/Stop 10.77
374 TestStartStop/group/default-k8s-diff-port/serial/DeployApp 9.24
375 TestStartStop/group/embed-certs/serial/EnableAddonAfterStop 0.15
376 TestStartStop/group/embed-certs/serial/SecondStart 262.1
377 TestStartStop/group/old-k8s-version/serial/DeployApp 8.36
378 TestStartStop/group/default-k8s-diff-port/serial/EnableAddonWhileActive 0.93
379 TestStartStop/group/default-k8s-diff-port/serial/Stop 10.91
380 TestStartStop/group/old-k8s-version/serial/EnableAddonWhileActive 0.8
381 TestStartStop/group/old-k8s-version/serial/Stop 10.8
382 TestStartStop/group/default-k8s-diff-port/serial/EnableAddonAfterStop 0.22
383 TestStartStop/group/default-k8s-diff-port/serial/SecondStart 263.3
384 TestStartStop/group/old-k8s-version/serial/EnableAddonAfterStop 0.16
385 TestStartStop/group/old-k8s-version/serial/SecondStart 119.41
386 TestStartStop/group/old-k8s-version/serial/UserAppExistsAfterStop 6.01
387 TestStartStop/group/old-k8s-version/serial/AddonExistsAfterStop 5.07
388 TestStartStop/group/old-k8s-version/serial/VerifyKubernetesImages 0.19
389 TestStartStop/group/old-k8s-version/serial/Pause 2.3
391 TestStartStop/group/newest-cni/serial/FirstStart 29.21
392 TestStartStop/group/newest-cni/serial/DeployApp 0
393 TestStartStop/group/newest-cni/serial/EnableAddonWhileActive 1.08
394 TestStartStop/group/newest-cni/serial/Stop 10.78
395 TestStartStop/group/newest-cni/serial/EnableAddonAfterStop 0.16
396 TestStartStop/group/newest-cni/serial/SecondStart 14.63
397 TestStartStop/group/newest-cni/serial/UserAppExistsAfterStop 0
398 TestStartStop/group/newest-cni/serial/AddonExistsAfterStop 0
399 TestStartStop/group/newest-cni/serial/VerifyKubernetesImages 0.21
400 TestStartStop/group/newest-cni/serial/Pause 2.48
401 TestStartStop/group/no-preload/serial/UserAppExistsAfterStop 6.01
402 TestStartStop/group/no-preload/serial/AddonExistsAfterStop 5.07
403 TestStartStop/group/no-preload/serial/VerifyKubernetesImages 0.19
404 TestStartStop/group/no-preload/serial/Pause 2.24
405 TestStartStop/group/embed-certs/serial/UserAppExistsAfterStop 6
406 TestStartStop/group/embed-certs/serial/AddonExistsAfterStop 5.07
407 TestStartStop/group/embed-certs/serial/VerifyKubernetesImages 0.19
408 TestStartStop/group/embed-certs/serial/Pause 2.23
409 TestStartStop/group/default-k8s-diff-port/serial/UserAppExistsAfterStop 6.01
410 TestStartStop/group/default-k8s-diff-port/serial/AddonExistsAfterStop 6.07
411 TestStartStop/group/default-k8s-diff-port/serial/VerifyKubernetesImages 0.19
412 TestStartStop/group/default-k8s-diff-port/serial/Pause 2.21
x
+
TestDownloadOnly/v1.20.0/json-events (4.53s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.20.0/json-events
aaa_download_only_test.go:81: (dbg) Run:  out/minikube-linux-amd64 start -o=json --download-only -p download-only-613757 --force --alsologtostderr --kubernetes-version=v1.20.0 --container-runtime=docker --driver=docker  --container-runtime=docker
aaa_download_only_test.go:81: (dbg) Done: out/minikube-linux-amd64 start -o=json --download-only -p download-only-613757 --force --alsologtostderr --kubernetes-version=v1.20.0 --container-runtime=docker --driver=docker  --container-runtime=docker: (4.527899465s)
--- PASS: TestDownloadOnly/v1.20.0/json-events (4.53s)

                                                
                                    
x
+
TestDownloadOnly/v1.20.0/preload-exists (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.20.0/preload-exists
--- PASS: TestDownloadOnly/v1.20.0/preload-exists (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.20.0/LogsDuration (0.05s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.20.0/LogsDuration
aaa_download_only_test.go:184: (dbg) Run:  out/minikube-linux-amd64 logs -p download-only-613757
aaa_download_only_test.go:184: (dbg) Non-zero exit: out/minikube-linux-amd64 logs -p download-only-613757: exit status 85 (51.612179ms)

                                                
                                                
-- stdout --
	
	==> Audit <==
	|---------|--------------------------------|----------------------|---------|---------|---------------------|----------|
	| Command |              Args              |       Profile        |  User   | Version |     Start Time      | End Time |
	|---------|--------------------------------|----------------------|---------|---------|---------------------|----------|
	| start   | -o=json --download-only        | download-only-613757 | jenkins | v1.34.0 | 13 Sep 24 18:20 UTC |          |
	|         | -p download-only-613757        |                      |         |         |                     |          |
	|         | --force --alsologtostderr      |                      |         |         |                     |          |
	|         | --kubernetes-version=v1.20.0   |                      |         |         |                     |          |
	|         | --container-runtime=docker     |                      |         |         |                     |          |
	|         | --driver=docker                |                      |         |         |                     |          |
	|         | --container-runtime=docker     |                      |         |         |                     |          |
	|---------|--------------------------------|----------------------|---------|---------|---------------------|----------|
	
	
	==> Last Start <==
	Log file created at: 2024/09/13 18:20:26
	Running on machine: ubuntu-20-agent-14
	Binary: Built with gc go1.23.0 for linux/amd64
	Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
	I0913 18:20:26.221281   10497 out.go:345] Setting OutFile to fd 1 ...
	I0913 18:20:26.221533   10497 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:20:26.221544   10497 out.go:358] Setting ErrFile to fd 2...
	I0913 18:20:26.221551   10497 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:20:26.221784   10497 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
	W0913 18:20:26.221924   10497 root.go:314] Error reading config file at /home/jenkins/minikube-integration/19636-3724/.minikube/config/config.json: open /home/jenkins/minikube-integration/19636-3724/.minikube/config/config.json: no such file or directory
	I0913 18:20:26.222513   10497 out.go:352] Setting JSON to true
	I0913 18:20:26.223380   10497 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-14","uptime":177,"bootTime":1726251449,"procs":178,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1068-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
	I0913 18:20:26.223468   10497 start.go:139] virtualization: kvm guest
	I0913 18:20:26.225814   10497 out.go:97] [download-only-613757] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
	W0913 18:20:26.225898   10497 preload.go:293] Failed to list preload files: open /home/jenkins/minikube-integration/19636-3724/.minikube/cache/preloaded-tarball: no such file or directory
	I0913 18:20:26.225928   10497 notify.go:220] Checking for updates...
	I0913 18:20:26.227284   10497 out.go:169] MINIKUBE_LOCATION=19636
	I0913 18:20:26.228624   10497 out.go:169] MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0913 18:20:26.229922   10497 out.go:169] KUBECONFIG=/home/jenkins/minikube-integration/19636-3724/kubeconfig
	I0913 18:20:26.231216   10497 out.go:169] MINIKUBE_HOME=/home/jenkins/minikube-integration/19636-3724/.minikube
	I0913 18:20:26.232516   10497 out.go:169] MINIKUBE_BIN=out/minikube-linux-amd64
	W0913 18:20:26.234649   10497 out.go:321] minikube skips various validations when --force is supplied; this may lead to unexpected behavior
	I0913 18:20:26.234840   10497 driver.go:394] Setting default libvirt URI to qemu:///system
	I0913 18:20:26.255813   10497 docker.go:123] docker version: linux-27.2.1:Docker Engine - Community
	I0913 18:20:26.255885   10497 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0913 18:20:26.578102   10497 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:30 OomKillDisable:true NGoroutines:53 SystemTime:2024-09-13 18:20:26.568901567 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
	I0913 18:20:26.578213   10497 docker.go:318] overlay module found
	I0913 18:20:26.579837   10497 out.go:97] Using the docker driver based on user configuration
	I0913 18:20:26.579864   10497 start.go:297] selected driver: docker
	I0913 18:20:26.579870   10497 start.go:901] validating driver "docker" against <nil>
	I0913 18:20:26.579952   10497 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0913 18:20:26.625242   10497 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:30 OomKillDisable:true NGoroutines:53 SystemTime:2024-09-13 18:20:26.617501622 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
	I0913 18:20:26.625383   10497 start_flags.go:310] no existing cluster config was found, will generate one from the flags 
	I0913 18:20:26.625905   10497 start_flags.go:393] Using suggested 8000MB memory alloc based on sys=32089MB, container=32089MB
	I0913 18:20:26.626055   10497 start_flags.go:929] Wait components to verify : map[apiserver:true system_pods:true]
	I0913 18:20:26.627934   10497 out.go:169] Using Docker driver with root privileges
	
	
	* The control-plane node download-only-613757 host does not exist
	  To start a cluster, run: "minikube start -p download-only-613757"

                                                
                                                
-- /stdout --
aaa_download_only_test.go:185: minikube logs failed with error: exit status 85
--- PASS: TestDownloadOnly/v1.20.0/LogsDuration (0.05s)

                                                
                                    
x
+
TestDownloadOnly/v1.20.0/DeleteAll (0.18s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.20.0/DeleteAll
aaa_download_only_test.go:197: (dbg) Run:  out/minikube-linux-amd64 delete --all
--- PASS: TestDownloadOnly/v1.20.0/DeleteAll (0.18s)

                                                
                                    
x
+
TestDownloadOnly/v1.20.0/DeleteAlwaysSucceeds (0.12s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.20.0/DeleteAlwaysSucceeds
aaa_download_only_test.go:208: (dbg) Run:  out/minikube-linux-amd64 delete -p download-only-613757
--- PASS: TestDownloadOnly/v1.20.0/DeleteAlwaysSucceeds (0.12s)

                                                
                                    
x
+
TestDownloadOnly/v1.31.1/json-events (4.26s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.31.1/json-events
aaa_download_only_test.go:81: (dbg) Run:  out/minikube-linux-amd64 start -o=json --download-only -p download-only-951288 --force --alsologtostderr --kubernetes-version=v1.31.1 --container-runtime=docker --driver=docker  --container-runtime=docker
aaa_download_only_test.go:81: (dbg) Done: out/minikube-linux-amd64 start -o=json --download-only -p download-only-951288 --force --alsologtostderr --kubernetes-version=v1.31.1 --container-runtime=docker --driver=docker  --container-runtime=docker: (4.262966848s)
--- PASS: TestDownloadOnly/v1.31.1/json-events (4.26s)

                                                
                                    
x
+
TestDownloadOnly/v1.31.1/preload-exists (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.31.1/preload-exists
--- PASS: TestDownloadOnly/v1.31.1/preload-exists (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.31.1/LogsDuration (0.05s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.31.1/LogsDuration
aaa_download_only_test.go:184: (dbg) Run:  out/minikube-linux-amd64 logs -p download-only-951288
aaa_download_only_test.go:184: (dbg) Non-zero exit: out/minikube-linux-amd64 logs -p download-only-951288: exit status 85 (51.655583ms)

                                                
                                                
-- stdout --
	
	==> Audit <==
	|---------|--------------------------------|----------------------|---------|---------|---------------------|---------------------|
	| Command |              Args              |       Profile        |  User   | Version |     Start Time      |      End Time       |
	|---------|--------------------------------|----------------------|---------|---------|---------------------|---------------------|
	| start   | -o=json --download-only        | download-only-613757 | jenkins | v1.34.0 | 13 Sep 24 18:20 UTC |                     |
	|         | -p download-only-613757        |                      |         |         |                     |                     |
	|         | --force --alsologtostderr      |                      |         |         |                     |                     |
	|         | --kubernetes-version=v1.20.0   |                      |         |         |                     |                     |
	|         | --container-runtime=docker     |                      |         |         |                     |                     |
	|         | --driver=docker                |                      |         |         |                     |                     |
	|         | --container-runtime=docker     |                      |         |         |                     |                     |
	| delete  | --all                          | minikube             | jenkins | v1.34.0 | 13 Sep 24 18:20 UTC | 13 Sep 24 18:20 UTC |
	| delete  | -p download-only-613757        | download-only-613757 | jenkins | v1.34.0 | 13 Sep 24 18:20 UTC | 13 Sep 24 18:20 UTC |
	| start   | -o=json --download-only        | download-only-951288 | jenkins | v1.34.0 | 13 Sep 24 18:20 UTC |                     |
	|         | -p download-only-951288        |                      |         |         |                     |                     |
	|         | --force --alsologtostderr      |                      |         |         |                     |                     |
	|         | --kubernetes-version=v1.31.1   |                      |         |         |                     |                     |
	|         | --container-runtime=docker     |                      |         |         |                     |                     |
	|         | --driver=docker                |                      |         |         |                     |                     |
	|         | --container-runtime=docker     |                      |         |         |                     |                     |
	|---------|--------------------------------|----------------------|---------|---------|---------------------|---------------------|
	
	
	==> Last Start <==
	Log file created at: 2024/09/13 18:20:31
	Running on machine: ubuntu-20-agent-14
	Binary: Built with gc go1.23.0 for linux/amd64
	Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
	I0913 18:20:31.095554   10825 out.go:345] Setting OutFile to fd 1 ...
	I0913 18:20:31.095656   10825 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:20:31.095665   10825 out.go:358] Setting ErrFile to fd 2...
	I0913 18:20:31.095670   10825 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:20:31.095843   10825 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
	I0913 18:20:31.096371   10825 out.go:352] Setting JSON to true
	I0913 18:20:31.097318   10825 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-14","uptime":182,"bootTime":1726251449,"procs":176,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1068-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
	I0913 18:20:31.097402   10825 start.go:139] virtualization: kvm guest
	I0913 18:20:31.099381   10825 out.go:97] [download-only-951288] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
	I0913 18:20:31.099501   10825 notify.go:220] Checking for updates...
	I0913 18:20:31.100666   10825 out.go:169] MINIKUBE_LOCATION=19636
	I0913 18:20:31.101801   10825 out.go:169] MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0913 18:20:31.102940   10825 out.go:169] KUBECONFIG=/home/jenkins/minikube-integration/19636-3724/kubeconfig
	I0913 18:20:31.103942   10825 out.go:169] MINIKUBE_HOME=/home/jenkins/minikube-integration/19636-3724/.minikube
	I0913 18:20:31.105135   10825 out.go:169] MINIKUBE_BIN=out/minikube-linux-amd64
	W0913 18:20:31.107134   10825 out.go:321] minikube skips various validations when --force is supplied; this may lead to unexpected behavior
	I0913 18:20:31.107396   10825 driver.go:394] Setting default libvirt URI to qemu:///system
	I0913 18:20:31.130543   10825 docker.go:123] docker version: linux-27.2.1:Docker Engine - Community
	I0913 18:20:31.130610   10825 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0913 18:20:31.179136   10825 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:30 OomKillDisable:true NGoroutines:46 SystemTime:2024-09-13 18:20:31.170879712 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
	I0913 18:20:31.179232   10825 docker.go:318] overlay module found
	I0913 18:20:31.180765   10825 out.go:97] Using the docker driver based on user configuration
	I0913 18:20:31.180783   10825 start.go:297] selected driver: docker
	I0913 18:20:31.180788   10825 start.go:901] validating driver "docker" against <nil>
	I0913 18:20:31.180857   10825 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0913 18:20:31.224266   10825 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:30 OomKillDisable:true NGoroutines:46 SystemTime:2024-09-13 18:20:31.215929299 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
	I0913 18:20:31.224490   10825 start_flags.go:310] no existing cluster config was found, will generate one from the flags 
	I0913 18:20:31.225173   10825 start_flags.go:393] Using suggested 8000MB memory alloc based on sys=32089MB, container=32089MB
	I0913 18:20:31.225364   10825 start_flags.go:929] Wait components to verify : map[apiserver:true system_pods:true]
	I0913 18:20:31.227004   10825 out.go:169] Using Docker driver with root privileges
	
	
	* The control-plane node download-only-951288 host does not exist
	  To start a cluster, run: "minikube start -p download-only-951288"

                                                
                                                
-- /stdout --
aaa_download_only_test.go:185: minikube logs failed with error: exit status 85
--- PASS: TestDownloadOnly/v1.31.1/LogsDuration (0.05s)

                                                
                                    
x
+
TestDownloadOnly/v1.31.1/DeleteAll (0.18s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.31.1/DeleteAll
aaa_download_only_test.go:197: (dbg) Run:  out/minikube-linux-amd64 delete --all
--- PASS: TestDownloadOnly/v1.31.1/DeleteAll (0.18s)

                                                
                                    
x
+
TestDownloadOnly/v1.31.1/DeleteAlwaysSucceeds (0.11s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.31.1/DeleteAlwaysSucceeds
aaa_download_only_test.go:208: (dbg) Run:  out/minikube-linux-amd64 delete -p download-only-951288
--- PASS: TestDownloadOnly/v1.31.1/DeleteAlwaysSucceeds (0.11s)

                                                
                                    
x
+
TestDownloadOnlyKic (0.94s)

                                                
                                                
=== RUN   TestDownloadOnlyKic
aaa_download_only_test.go:232: (dbg) Run:  out/minikube-linux-amd64 start --download-only -p download-docker-782708 --alsologtostderr --driver=docker  --container-runtime=docker
helpers_test.go:175: Cleaning up "download-docker-782708" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p download-docker-782708
--- PASS: TestDownloadOnlyKic (0.94s)

                                                
                                    
x
+
TestBinaryMirror (0.72s)

                                                
                                                
=== RUN   TestBinaryMirror
aaa_download_only_test.go:314: (dbg) Run:  out/minikube-linux-amd64 start --download-only -p binary-mirror-486689 --alsologtostderr --binary-mirror http://127.0.0.1:41715 --driver=docker  --container-runtime=docker
helpers_test.go:175: Cleaning up "binary-mirror-486689" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p binary-mirror-486689
--- PASS: TestBinaryMirror (0.72s)

                                                
                                    
x
+
TestOffline (55.25s)

                                                
                                                
=== RUN   TestOffline
=== PAUSE TestOffline

                                                
                                                

                                                
                                                
=== CONT  TestOffline
aab_offline_test.go:55: (dbg) Run:  out/minikube-linux-amd64 start -p offline-docker-363399 --alsologtostderr -v=1 --memory=2048 --wait=true --driver=docker  --container-runtime=docker
aab_offline_test.go:55: (dbg) Done: out/minikube-linux-amd64 start -p offline-docker-363399 --alsologtostderr -v=1 --memory=2048 --wait=true --driver=docker  --container-runtime=docker: (53.203299369s)
helpers_test.go:175: Cleaning up "offline-docker-363399" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p offline-docker-363399
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p offline-docker-363399: (2.050152243s)
--- PASS: TestOffline (55.25s)

                                                
                                    
x
+
TestAddons/PreSetup/EnablingAddonOnNonExistingCluster (0.05s)

                                                
                                                
=== RUN   TestAddons/PreSetup/EnablingAddonOnNonExistingCluster
=== PAUSE TestAddons/PreSetup/EnablingAddonOnNonExistingCluster

                                                
                                                

                                                
                                                
=== CONT  TestAddons/PreSetup/EnablingAddonOnNonExistingCluster
addons_test.go:975: (dbg) Run:  out/minikube-linux-amd64 addons enable dashboard -p addons-315216
addons_test.go:975: (dbg) Non-zero exit: out/minikube-linux-amd64 addons enable dashboard -p addons-315216: exit status 85 (45.685537ms)

                                                
                                                
-- stdout --
	* Profile "addons-315216" not found. Run "minikube profile list" to view all profiles.
	  To start a cluster, run: "minikube start -p addons-315216"

                                                
                                                
-- /stdout --
--- PASS: TestAddons/PreSetup/EnablingAddonOnNonExistingCluster (0.05s)

                                                
                                    
x
+
TestAddons/PreSetup/DisablingAddonOnNonExistingCluster (0.05s)

                                                
                                                
=== RUN   TestAddons/PreSetup/DisablingAddonOnNonExistingCluster
=== PAUSE TestAddons/PreSetup/DisablingAddonOnNonExistingCluster

                                                
                                                

                                                
                                                
=== CONT  TestAddons/PreSetup/DisablingAddonOnNonExistingCluster
addons_test.go:986: (dbg) Run:  out/minikube-linux-amd64 addons disable dashboard -p addons-315216
addons_test.go:986: (dbg) Non-zero exit: out/minikube-linux-amd64 addons disable dashboard -p addons-315216: exit status 85 (45.131022ms)

                                                
                                                
-- stdout --
	* Profile "addons-315216" not found. Run "minikube profile list" to view all profiles.
	  To start a cluster, run: "minikube start -p addons-315216"

                                                
                                                
-- /stdout --
--- PASS: TestAddons/PreSetup/DisablingAddonOnNonExistingCluster (0.05s)

                                                
                                    
x
+
TestAddons/Setup (206.41s)

                                                
                                                
=== RUN   TestAddons/Setup
addons_test.go:107: (dbg) Run:  out/minikube-linux-amd64 start -p addons-315216 --wait=true --memory=4000 --alsologtostderr --addons=registry --addons=metrics-server --addons=volumesnapshots --addons=csi-hostpath-driver --addons=gcp-auth --addons=cloud-spanner --addons=inspektor-gadget --addons=storage-provisioner-rancher --addons=nvidia-device-plugin --addons=yakd --addons=volcano --driver=docker  --container-runtime=docker --addons=ingress --addons=ingress-dns
addons_test.go:107: (dbg) Done: out/minikube-linux-amd64 start -p addons-315216 --wait=true --memory=4000 --alsologtostderr --addons=registry --addons=metrics-server --addons=volumesnapshots --addons=csi-hostpath-driver --addons=gcp-auth --addons=cloud-spanner --addons=inspektor-gadget --addons=storage-provisioner-rancher --addons=nvidia-device-plugin --addons=yakd --addons=volcano --driver=docker  --container-runtime=docker --addons=ingress --addons=ingress-dns: (3m26.408187637s)
--- PASS: TestAddons/Setup (206.41s)

                                                
                                    
x
+
TestAddons/serial/Volcano (38.56s)

                                                
                                                
=== RUN   TestAddons/serial/Volcano
addons_test.go:851: volcano-controller stabilized in 14.466124ms
addons_test.go:843: volcano-admission stabilized in 14.527756ms
addons_test.go:835: volcano-scheduler stabilized in 14.689894ms
addons_test.go:857: (dbg) TestAddons/serial/Volcano: waiting 6m0s for pods matching "app=volcano-scheduler" in namespace "volcano-system" ...
helpers_test.go:344: "volcano-scheduler-576bc46687-vls4s" [cfb259e4-4f12-4219-9909-27fdf299d486] Running
addons_test.go:857: (dbg) TestAddons/serial/Volcano: app=volcano-scheduler healthy within 6.003072022s
addons_test.go:861: (dbg) TestAddons/serial/Volcano: waiting 6m0s for pods matching "app=volcano-admission" in namespace "volcano-system" ...
helpers_test.go:344: "volcano-admission-77d7d48b68-n67j7" [8bffa595-851d-435f-bb6b-19163c1d31dd] Running
addons_test.go:861: (dbg) TestAddons/serial/Volcano: app=volcano-admission healthy within 5.00336s
addons_test.go:865: (dbg) TestAddons/serial/Volcano: waiting 6m0s for pods matching "app=volcano-controller" in namespace "volcano-system" ...
helpers_test.go:344: "volcano-controllers-56675bb4d5-vx95g" [7a207fe5-7eee-4d95-b1e7-0f389c603d5f] Running
addons_test.go:865: (dbg) TestAddons/serial/Volcano: app=volcano-controller healthy within 5.003041066s
addons_test.go:870: (dbg) Run:  kubectl --context addons-315216 delete -n volcano-system job volcano-admission-init
addons_test.go:876: (dbg) Run:  kubectl --context addons-315216 create -f testdata/vcjob.yaml
addons_test.go:884: (dbg) Run:  kubectl --context addons-315216 get vcjob -n my-volcano
addons_test.go:902: (dbg) TestAddons/serial/Volcano: waiting 3m0s for pods matching "volcano.sh/job-name=test-job" in namespace "my-volcano" ...
helpers_test.go:344: "test-job-nginx-0" [6fb7ff2f-26af-46a2-a81d-fcd2200114d5] Pending
helpers_test.go:344: "test-job-nginx-0" [6fb7ff2f-26af-46a2-a81d-fcd2200114d5] Pending / Ready:ContainersNotReady (containers with unready status: [nginx]) / ContainersReady:ContainersNotReady (containers with unready status: [nginx])
helpers_test.go:344: "test-job-nginx-0" [6fb7ff2f-26af-46a2-a81d-fcd2200114d5] Running
addons_test.go:902: (dbg) TestAddons/serial/Volcano: volcano.sh/job-name=test-job healthy within 12.003473761s
addons_test.go:906: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 addons disable volcano --alsologtostderr -v=1
addons_test.go:906: (dbg) Done: out/minikube-linux-amd64 -p addons-315216 addons disable volcano --alsologtostderr -v=1: (10.238092134s)
--- PASS: TestAddons/serial/Volcano (38.56s)

                                                
                                    
x
+
TestAddons/serial/GCPAuth/Namespaces (0.11s)

                                                
                                                
=== RUN   TestAddons/serial/GCPAuth/Namespaces
addons_test.go:594: (dbg) Run:  kubectl --context addons-315216 create ns new-namespace
addons_test.go:608: (dbg) Run:  kubectl --context addons-315216 get secret gcp-auth -n new-namespace
--- PASS: TestAddons/serial/GCPAuth/Namespaces (0.11s)

                                                
                                    
x
+
TestAddons/parallel/Ingress (19.4s)

                                                
                                                
=== RUN   TestAddons/parallel/Ingress
=== PAUSE TestAddons/parallel/Ingress

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/Ingress
addons_test.go:205: (dbg) Run:  kubectl --context addons-315216 wait --for=condition=ready --namespace=ingress-nginx pod --selector=app.kubernetes.io/component=controller --timeout=90s
addons_test.go:230: (dbg) Run:  kubectl --context addons-315216 replace --force -f testdata/nginx-ingress-v1.yaml
addons_test.go:243: (dbg) Run:  kubectl --context addons-315216 replace --force -f testdata/nginx-pod-svc.yaml
addons_test.go:248: (dbg) TestAddons/parallel/Ingress: waiting 8m0s for pods matching "run=nginx" in namespace "default" ...
helpers_test.go:344: "nginx" [ecba2f4a-f586-4ed0-9be7-7e2240df903b] Pending / Ready:ContainersNotReady (containers with unready status: [nginx]) / ContainersReady:ContainersNotReady (containers with unready status: [nginx])
helpers_test.go:344: "nginx" [ecba2f4a-f586-4ed0-9be7-7e2240df903b] Running
addons_test.go:248: (dbg) TestAddons/parallel/Ingress: run=nginx healthy within 9.003405619s
addons_test.go:260: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 ssh "curl -s http://127.0.0.1/ -H 'Host: nginx.example.com'"
addons_test.go:284: (dbg) Run:  kubectl --context addons-315216 replace --force -f testdata/ingress-dns-example-v1.yaml
addons_test.go:289: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 ip
addons_test.go:295: (dbg) Run:  nslookup hello-john.test 192.168.49.2
addons_test.go:304: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 addons disable ingress-dns --alsologtostderr -v=1
addons_test.go:304: (dbg) Done: out/minikube-linux-amd64 -p addons-315216 addons disable ingress-dns --alsologtostderr -v=1: (1.422061534s)
addons_test.go:309: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 addons disable ingress --alsologtostderr -v=1
addons_test.go:309: (dbg) Done: out/minikube-linux-amd64 -p addons-315216 addons disable ingress --alsologtostderr -v=1: (7.783652291s)
--- PASS: TestAddons/parallel/Ingress (19.40s)

                                                
                                    
x
+
TestAddons/parallel/InspektorGadget (10.6s)

                                                
                                                
=== RUN   TestAddons/parallel/InspektorGadget
=== PAUSE TestAddons/parallel/InspektorGadget

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/InspektorGadget
addons_test.go:786: (dbg) TestAddons/parallel/InspektorGadget: waiting 8m0s for pods matching "k8s-app=gadget" in namespace "gadget" ...
helpers_test.go:344: "gadget-h2qzx" [2e6c7913-4292-4f3d-85cf-a2844935dd58] Running / Ready:ContainersNotReady (containers with unready status: [gadget]) / ContainersReady:ContainersNotReady (containers with unready status: [gadget])
addons_test.go:786: (dbg) TestAddons/parallel/InspektorGadget: k8s-app=gadget healthy within 5.003284571s
addons_test.go:789: (dbg) Run:  out/minikube-linux-amd64 addons disable inspektor-gadget -p addons-315216
addons_test.go:789: (dbg) Done: out/minikube-linux-amd64 addons disable inspektor-gadget -p addons-315216: (5.595430684s)
--- PASS: TestAddons/parallel/InspektorGadget (10.60s)

                                                
                                    
x
+
TestAddons/parallel/MetricsServer (5.77s)

                                                
                                                
=== RUN   TestAddons/parallel/MetricsServer
=== PAUSE TestAddons/parallel/MetricsServer

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/MetricsServer
addons_test.go:405: metrics-server stabilized in 2.264759ms
addons_test.go:407: (dbg) TestAddons/parallel/MetricsServer: waiting 6m0s for pods matching "k8s-app=metrics-server" in namespace "kube-system" ...
helpers_test.go:344: "metrics-server-84c5f94fbc-5r2xh" [9fd33765-d89a-4aca-9d83-b2ef709128c4] Running
addons_test.go:407: (dbg) TestAddons/parallel/MetricsServer: k8s-app=metrics-server healthy within 5.002093545s
addons_test.go:413: (dbg) Run:  kubectl --context addons-315216 top pods -n kube-system
addons_test.go:430: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 addons disable metrics-server --alsologtostderr -v=1
--- PASS: TestAddons/parallel/MetricsServer (5.77s)

                                                
                                    
x
+
TestAddons/parallel/CSI (57.02s)

                                                
                                                
=== RUN   TestAddons/parallel/CSI
=== PAUSE TestAddons/parallel/CSI

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/CSI
addons_test.go:505: csi-hostpath-driver pods stabilized in 3.809755ms
addons_test.go:508: (dbg) Run:  kubectl --context addons-315216 create -f testdata/csi-hostpath-driver/pvc.yaml
addons_test.go:513: (dbg) TestAddons/parallel/CSI: waiting 6m0s for pvc "hpvc" in namespace "default" ...
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc -o jsonpath={.status.phase} -n default
addons_test.go:518: (dbg) Run:  kubectl --context addons-315216 create -f testdata/csi-hostpath-driver/pv-pod.yaml
addons_test.go:523: (dbg) TestAddons/parallel/CSI: waiting 6m0s for pods matching "app=task-pv-pod" in namespace "default" ...
helpers_test.go:344: "task-pv-pod" [d4d8b7c2-2469-4431-aeea-dc83db734bf6] Pending
helpers_test.go:344: "task-pv-pod" [d4d8b7c2-2469-4431-aeea-dc83db734bf6] Pending / Ready:ContainersNotReady (containers with unready status: [task-pv-container]) / ContainersReady:ContainersNotReady (containers with unready status: [task-pv-container])
helpers_test.go:344: "task-pv-pod" [d4d8b7c2-2469-4431-aeea-dc83db734bf6] Running
addons_test.go:523: (dbg) TestAddons/parallel/CSI: app=task-pv-pod healthy within 8.003549304s
addons_test.go:528: (dbg) Run:  kubectl --context addons-315216 create -f testdata/csi-hostpath-driver/snapshot.yaml
addons_test.go:533: (dbg) TestAddons/parallel/CSI: waiting 6m0s for volume snapshot "new-snapshot-demo" in namespace "default" ...
helpers_test.go:419: (dbg) Run:  kubectl --context addons-315216 get volumesnapshot new-snapshot-demo -o jsonpath={.status.readyToUse} -n default
helpers_test.go:419: (dbg) Run:  kubectl --context addons-315216 get volumesnapshot new-snapshot-demo -o jsonpath={.status.readyToUse} -n default
addons_test.go:538: (dbg) Run:  kubectl --context addons-315216 delete pod task-pv-pod
addons_test.go:544: (dbg) Run:  kubectl --context addons-315216 delete pvc hpvc
addons_test.go:550: (dbg) Run:  kubectl --context addons-315216 create -f testdata/csi-hostpath-driver/pvc-restore.yaml
addons_test.go:555: (dbg) TestAddons/parallel/CSI: waiting 6m0s for pvc "hpvc-restore" in namespace "default" ...
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
addons_test.go:560: (dbg) Run:  kubectl --context addons-315216 create -f testdata/csi-hostpath-driver/pv-pod-restore.yaml
addons_test.go:565: (dbg) TestAddons/parallel/CSI: waiting 6m0s for pods matching "app=task-pv-pod-restore" in namespace "default" ...
helpers_test.go:344: "task-pv-pod-restore" [4a173bf3-f286-475b-b6bc-aaffb735d9bd] Pending
helpers_test.go:344: "task-pv-pod-restore" [4a173bf3-f286-475b-b6bc-aaffb735d9bd] Pending / Ready:ContainersNotReady (containers with unready status: [task-pv-container]) / ContainersReady:ContainersNotReady (containers with unready status: [task-pv-container])
helpers_test.go:344: "task-pv-pod-restore" [4a173bf3-f286-475b-b6bc-aaffb735d9bd] Running
addons_test.go:565: (dbg) TestAddons/parallel/CSI: app=task-pv-pod-restore healthy within 7.003424244s
addons_test.go:570: (dbg) Run:  kubectl --context addons-315216 delete pod task-pv-pod-restore
addons_test.go:574: (dbg) Run:  kubectl --context addons-315216 delete pvc hpvc-restore
addons_test.go:578: (dbg) Run:  kubectl --context addons-315216 delete volumesnapshot new-snapshot-demo
addons_test.go:582: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 addons disable csi-hostpath-driver --alsologtostderr -v=1
addons_test.go:582: (dbg) Done: out/minikube-linux-amd64 -p addons-315216 addons disable csi-hostpath-driver --alsologtostderr -v=1: (6.438701659s)
addons_test.go:586: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 addons disable volumesnapshots --alsologtostderr -v=1
--- PASS: TestAddons/parallel/CSI (57.02s)

                                                
                                    
x
+
TestAddons/parallel/Headlamp (16.26s)

                                                
                                                
=== RUN   TestAddons/parallel/Headlamp
=== PAUSE TestAddons/parallel/Headlamp

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/Headlamp
addons_test.go:768: (dbg) Run:  out/minikube-linux-amd64 addons enable headlamp -p addons-315216 --alsologtostderr -v=1
addons_test.go:773: (dbg) TestAddons/parallel/Headlamp: waiting 8m0s for pods matching "app.kubernetes.io/name=headlamp" in namespace "headlamp" ...
helpers_test.go:344: "headlamp-57fb76fcdb-2hwsw" [60d19a6d-483f-46b2-a8f9-e02da8b05474] Pending / Ready:ContainersNotReady (containers with unready status: [headlamp]) / ContainersReady:ContainersNotReady (containers with unready status: [headlamp])
helpers_test.go:344: "headlamp-57fb76fcdb-2hwsw" [60d19a6d-483f-46b2-a8f9-e02da8b05474] Running
addons_test.go:773: (dbg) TestAddons/parallel/Headlamp: app.kubernetes.io/name=headlamp healthy within 10.004247871s
addons_test.go:777: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 addons disable headlamp --alsologtostderr -v=1
addons_test.go:777: (dbg) Done: out/minikube-linux-amd64 -p addons-315216 addons disable headlamp --alsologtostderr -v=1: (5.594740057s)
--- PASS: TestAddons/parallel/Headlamp (16.26s)

                                                
                                    
x
+
TestAddons/parallel/CloudSpanner (5.4s)

                                                
                                                
=== RUN   TestAddons/parallel/CloudSpanner
=== PAUSE TestAddons/parallel/CloudSpanner

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/CloudSpanner
addons_test.go:805: (dbg) TestAddons/parallel/CloudSpanner: waiting 6m0s for pods matching "app=cloud-spanner-emulator" in namespace "default" ...
helpers_test.go:344: "cloud-spanner-emulator-769b77f747-vtrdf" [631aa0f3-6d9b-48ee-97ce-d079a413cb41] Running
addons_test.go:805: (dbg) TestAddons/parallel/CloudSpanner: app=cloud-spanner-emulator healthy within 5.003670419s
addons_test.go:808: (dbg) Run:  out/minikube-linux-amd64 addons disable cloud-spanner -p addons-315216
--- PASS: TestAddons/parallel/CloudSpanner (5.40s)

                                                
                                    
x
+
TestAddons/parallel/LocalPath (51.61s)

                                                
                                                
=== RUN   TestAddons/parallel/LocalPath
=== PAUSE TestAddons/parallel/LocalPath

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/LocalPath
addons_test.go:920: (dbg) Run:  kubectl --context addons-315216 apply -f testdata/storage-provisioner-rancher/pvc.yaml
addons_test.go:926: (dbg) Run:  kubectl --context addons-315216 apply -f testdata/storage-provisioner-rancher/pod.yaml
addons_test.go:930: (dbg) TestAddons/parallel/LocalPath: waiting 5m0s for pvc "test-pvc" in namespace "default" ...
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc test-pvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc test-pvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc test-pvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc test-pvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc test-pvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run:  kubectl --context addons-315216 get pvc test-pvc -o jsonpath={.status.phase} -n default
addons_test.go:933: (dbg) TestAddons/parallel/LocalPath: waiting 3m0s for pods matching "run=test-local-path" in namespace "default" ...
helpers_test.go:344: "test-local-path" [6c7a518e-c5ed-45c1-99e8-0d1a4b57c3af] Pending / Ready:ContainersNotReady (containers with unready status: [busybox]) / ContainersReady:ContainersNotReady (containers with unready status: [busybox])
helpers_test.go:344: "test-local-path" [6c7a518e-c5ed-45c1-99e8-0d1a4b57c3af] Pending / Initialized:PodCompleted / Ready:PodCompleted / ContainersReady:PodCompleted
helpers_test.go:344: "test-local-path" [6c7a518e-c5ed-45c1-99e8-0d1a4b57c3af] Succeeded / Initialized:PodCompleted / Ready:PodCompleted / ContainersReady:PodCompleted
addons_test.go:933: (dbg) TestAddons/parallel/LocalPath: run=test-local-path healthy within 3.002934594s
addons_test.go:938: (dbg) Run:  kubectl --context addons-315216 get pvc test-pvc -o=json
addons_test.go:947: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 ssh "cat /opt/local-path-provisioner/pvc-2743739b-b850-4f4a-8324-d1e883dc17e5_default_test-pvc/file1"
addons_test.go:959: (dbg) Run:  kubectl --context addons-315216 delete pod test-local-path
addons_test.go:963: (dbg) Run:  kubectl --context addons-315216 delete pvc test-pvc
addons_test.go:967: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 addons disable storage-provisioner-rancher --alsologtostderr -v=1
addons_test.go:967: (dbg) Done: out/minikube-linux-amd64 -p addons-315216 addons disable storage-provisioner-rancher --alsologtostderr -v=1: (42.823304462s)
--- PASS: TestAddons/parallel/LocalPath (51.61s)

                                                
                                    
x
+
TestAddons/parallel/NvidiaDevicePlugin (5.37s)

                                                
                                                
=== RUN   TestAddons/parallel/NvidiaDevicePlugin
=== PAUSE TestAddons/parallel/NvidiaDevicePlugin

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/NvidiaDevicePlugin
addons_test.go:999: (dbg) TestAddons/parallel/NvidiaDevicePlugin: waiting 6m0s for pods matching "name=nvidia-device-plugin-ds" in namespace "kube-system" ...
helpers_test.go:344: "nvidia-device-plugin-daemonset-tvp28" [845bb4f2-4781-45e0-a654-5b57584b3a70] Running
addons_test.go:999: (dbg) TestAddons/parallel/NvidiaDevicePlugin: name=nvidia-device-plugin-ds healthy within 5.003550179s
addons_test.go:1002: (dbg) Run:  out/minikube-linux-amd64 addons disable nvidia-device-plugin -p addons-315216
--- PASS: TestAddons/parallel/NvidiaDevicePlugin (5.37s)

                                                
                                    
x
+
TestAddons/parallel/Yakd (10.6s)

                                                
                                                
=== RUN   TestAddons/parallel/Yakd
=== PAUSE TestAddons/parallel/Yakd

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/Yakd
addons_test.go:1010: (dbg) TestAddons/parallel/Yakd: waiting 2m0s for pods matching "app.kubernetes.io/name=yakd-dashboard" in namespace "yakd-dashboard" ...
helpers_test.go:344: "yakd-dashboard-67d98fc6b-hhmz8" [ecbb7e48-6475-4065-b3c4-de33906e09db] Running
addons_test.go:1010: (dbg) TestAddons/parallel/Yakd: app.kubernetes.io/name=yakd-dashboard healthy within 5.003063299s
addons_test.go:1014: (dbg) Run:  out/minikube-linux-amd64 -p addons-315216 addons disable yakd --alsologtostderr -v=1
addons_test.go:1014: (dbg) Done: out/minikube-linux-amd64 -p addons-315216 addons disable yakd --alsologtostderr -v=1: (5.598522892s)
--- PASS: TestAddons/parallel/Yakd (10.60s)

                                                
                                    
x
+
TestAddons/StoppedEnableDisable (10.91s)

                                                
                                                
=== RUN   TestAddons/StoppedEnableDisable
addons_test.go:170: (dbg) Run:  out/minikube-linux-amd64 stop -p addons-315216
addons_test.go:170: (dbg) Done: out/minikube-linux-amd64 stop -p addons-315216: (10.692342364s)
addons_test.go:174: (dbg) Run:  out/minikube-linux-amd64 addons enable dashboard -p addons-315216
addons_test.go:178: (dbg) Run:  out/minikube-linux-amd64 addons disable dashboard -p addons-315216
addons_test.go:183: (dbg) Run:  out/minikube-linux-amd64 addons disable gvisor -p addons-315216
--- PASS: TestAddons/StoppedEnableDisable (10.91s)

                                                
                                    
x
+
TestCertOptions (31.02s)

                                                
                                                
=== RUN   TestCertOptions
=== PAUSE TestCertOptions

                                                
                                                

                                                
                                                
=== CONT  TestCertOptions
cert_options_test.go:49: (dbg) Run:  out/minikube-linux-amd64 start -p cert-options-392074 --memory=2048 --apiserver-ips=127.0.0.1 --apiserver-ips=192.168.15.15 --apiserver-names=localhost --apiserver-names=www.google.com --apiserver-port=8555 --driver=docker  --container-runtime=docker
cert_options_test.go:49: (dbg) Done: out/minikube-linux-amd64 start -p cert-options-392074 --memory=2048 --apiserver-ips=127.0.0.1 --apiserver-ips=192.168.15.15 --apiserver-names=localhost --apiserver-names=www.google.com --apiserver-port=8555 --driver=docker  --container-runtime=docker: (25.589313992s)
cert_options_test.go:60: (dbg) Run:  out/minikube-linux-amd64 -p cert-options-392074 ssh "openssl x509 -text -noout -in /var/lib/minikube/certs/apiserver.crt"
cert_options_test.go:88: (dbg) Run:  kubectl --context cert-options-392074 config view
cert_options_test.go:100: (dbg) Run:  out/minikube-linux-amd64 ssh -p cert-options-392074 -- "sudo cat /etc/kubernetes/admin.conf"
helpers_test.go:175: Cleaning up "cert-options-392074" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p cert-options-392074
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p cert-options-392074: (4.701735317s)
--- PASS: TestCertOptions (31.02s)

                                                
                                    
x
+
TestCertExpiration (232.22s)

                                                
                                                
=== RUN   TestCertExpiration
=== PAUSE TestCertExpiration

                                                
                                                

                                                
                                                
=== CONT  TestCertExpiration
cert_options_test.go:123: (dbg) Run:  out/minikube-linux-amd64 start -p cert-expiration-174035 --memory=2048 --cert-expiration=3m --driver=docker  --container-runtime=docker
cert_options_test.go:123: (dbg) Done: out/minikube-linux-amd64 start -p cert-expiration-174035 --memory=2048 --cert-expiration=3m --driver=docker  --container-runtime=docker: (28.559610934s)
cert_options_test.go:131: (dbg) Run:  out/minikube-linux-amd64 start -p cert-expiration-174035 --memory=2048 --cert-expiration=8760h --driver=docker  --container-runtime=docker
cert_options_test.go:131: (dbg) Done: out/minikube-linux-amd64 start -p cert-expiration-174035 --memory=2048 --cert-expiration=8760h --driver=docker  --container-runtime=docker: (21.453461284s)
helpers_test.go:175: Cleaning up "cert-expiration-174035" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p cert-expiration-174035
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p cert-expiration-174035: (2.20342145s)
--- PASS: TestCertExpiration (232.22s)

                                                
                                    
x
+
TestDockerFlags (26.8s)

                                                
                                                
=== RUN   TestDockerFlags
=== PAUSE TestDockerFlags

                                                
                                                

                                                
                                                
=== CONT  TestDockerFlags
docker_test.go:51: (dbg) Run:  out/minikube-linux-amd64 start -p docker-flags-392384 --cache-images=false --memory=2048 --install-addons=false --wait=false --docker-env=FOO=BAR --docker-env=BAZ=BAT --docker-opt=debug --docker-opt=icc=true --alsologtostderr -v=5 --driver=docker  --container-runtime=docker
docker_test.go:51: (dbg) Done: out/minikube-linux-amd64 start -p docker-flags-392384 --cache-images=false --memory=2048 --install-addons=false --wait=false --docker-env=FOO=BAR --docker-env=BAZ=BAT --docker-opt=debug --docker-opt=icc=true --alsologtostderr -v=5 --driver=docker  --container-runtime=docker: (23.717923472s)
docker_test.go:56: (dbg) Run:  out/minikube-linux-amd64 -p docker-flags-392384 ssh "sudo systemctl show docker --property=Environment --no-pager"
docker_test.go:67: (dbg) Run:  out/minikube-linux-amd64 -p docker-flags-392384 ssh "sudo systemctl show docker --property=ExecStart --no-pager"
helpers_test.go:175: Cleaning up "docker-flags-392384" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p docker-flags-392384
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p docker-flags-392384: (2.402433853s)
--- PASS: TestDockerFlags (26.80s)

                                                
                                    
x
+
TestForceSystemdFlag (40.87s)

                                                
                                                
=== RUN   TestForceSystemdFlag
=== PAUSE TestForceSystemdFlag

                                                
                                                

                                                
                                                
=== CONT  TestForceSystemdFlag
docker_test.go:91: (dbg) Run:  out/minikube-linux-amd64 start -p force-systemd-flag-425143 --memory=2048 --force-systemd --alsologtostderr -v=5 --driver=docker  --container-runtime=docker
docker_test.go:91: (dbg) Done: out/minikube-linux-amd64 start -p force-systemd-flag-425143 --memory=2048 --force-systemd --alsologtostderr -v=5 --driver=docker  --container-runtime=docker: (38.427695878s)
docker_test.go:110: (dbg) Run:  out/minikube-linux-amd64 -p force-systemd-flag-425143 ssh "docker info --format {{.CgroupDriver}}"
helpers_test.go:175: Cleaning up "force-systemd-flag-425143" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p force-systemd-flag-425143
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p force-systemd-flag-425143: (2.136257899s)
--- PASS: TestForceSystemdFlag (40.87s)

                                                
                                    
x
+
TestForceSystemdEnv (28.16s)

                                                
                                                
=== RUN   TestForceSystemdEnv
=== PAUSE TestForceSystemdEnv

                                                
                                                

                                                
                                                
=== CONT  TestForceSystemdEnv
docker_test.go:155: (dbg) Run:  out/minikube-linux-amd64 start -p force-systemd-env-499716 --memory=2048 --alsologtostderr -v=5 --driver=docker  --container-runtime=docker
docker_test.go:155: (dbg) Done: out/minikube-linux-amd64 start -p force-systemd-env-499716 --memory=2048 --alsologtostderr -v=5 --driver=docker  --container-runtime=docker: (23.991084428s)
docker_test.go:110: (dbg) Run:  out/minikube-linux-amd64 -p force-systemd-env-499716 ssh "docker info --format {{.CgroupDriver}}"
helpers_test.go:175: Cleaning up "force-systemd-env-499716" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p force-systemd-env-499716
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p force-systemd-env-499716: (3.799713587s)
--- PASS: TestForceSystemdEnv (28.16s)

                                                
                                    
x
+
TestKVMDriverInstallOrUpdate (1.17s)

                                                
                                                
=== RUN   TestKVMDriverInstallOrUpdate
=== PAUSE TestKVMDriverInstallOrUpdate

                                                
                                                

                                                
                                                
=== CONT  TestKVMDriverInstallOrUpdate
--- PASS: TestKVMDriverInstallOrUpdate (1.17s)

                                                
                                    
x
+
TestErrorSpam/setup (20.01s)

                                                
                                                
=== RUN   TestErrorSpam/setup
error_spam_test.go:81: (dbg) Run:  out/minikube-linux-amd64 start -p nospam-621543 -n=1 --memory=2250 --wait=false --log_dir=/tmp/nospam-621543 --driver=docker  --container-runtime=docker
error_spam_test.go:81: (dbg) Done: out/minikube-linux-amd64 start -p nospam-621543 -n=1 --memory=2250 --wait=false --log_dir=/tmp/nospam-621543 --driver=docker  --container-runtime=docker: (20.01032655s)
--- PASS: TestErrorSpam/setup (20.01s)

                                                
                                    
x
+
TestErrorSpam/start (0.52s)

                                                
                                                
=== RUN   TestErrorSpam/start
error_spam_test.go:216: Cleaning up 1 logfile(s) ...
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 start --dry-run
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 start --dry-run
error_spam_test.go:182: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 start --dry-run
--- PASS: TestErrorSpam/start (0.52s)

                                                
                                    
x
+
TestErrorSpam/status (0.82s)

                                                
                                                
=== RUN   TestErrorSpam/status
error_spam_test.go:216: Cleaning up 0 logfile(s) ...
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 status
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 status
error_spam_test.go:182: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 status
--- PASS: TestErrorSpam/status (0.82s)

                                                
                                    
x
+
TestErrorSpam/pause (1.1s)

                                                
                                                
=== RUN   TestErrorSpam/pause
error_spam_test.go:216: Cleaning up 0 logfile(s) ...
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 pause
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 pause
error_spam_test.go:182: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 pause
--- PASS: TestErrorSpam/pause (1.10s)

                                                
                                    
x
+
TestErrorSpam/unpause (1.16s)

                                                
                                                
=== RUN   TestErrorSpam/unpause
error_spam_test.go:216: Cleaning up 0 logfile(s) ...
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 unpause
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 unpause
error_spam_test.go:182: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 unpause
--- PASS: TestErrorSpam/unpause (1.16s)

                                                
                                    
x
+
TestErrorSpam/stop (1.86s)

                                                
                                                
=== RUN   TestErrorSpam/stop
error_spam_test.go:216: Cleaning up 0 logfile(s) ...
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 stop
error_spam_test.go:159: (dbg) Done: out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 stop: (1.705561795s)
error_spam_test.go:159: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 stop
error_spam_test.go:182: (dbg) Run:  out/minikube-linux-amd64 -p nospam-621543 --log_dir /tmp/nospam-621543 stop
--- PASS: TestErrorSpam/stop (1.86s)

                                                
                                    
x
+
TestFunctional/serial/CopySyncFile (0s)

                                                
                                                
=== RUN   TestFunctional/serial/CopySyncFile
functional_test.go:1855: local sync path: /home/jenkins/minikube-integration/19636-3724/.minikube/files/etc/test/nested/copy/10475/hosts
--- PASS: TestFunctional/serial/CopySyncFile (0.00s)

                                                
                                    
x
+
TestFunctional/serial/StartWithProxy (67.27s)

                                                
                                                
=== RUN   TestFunctional/serial/StartWithProxy
functional_test.go:2234: (dbg) Run:  out/minikube-linux-amd64 start -p functional-463208 --memory=4000 --apiserver-port=8441 --wait=all --driver=docker  --container-runtime=docker
functional_test.go:2234: (dbg) Done: out/minikube-linux-amd64 start -p functional-463208 --memory=4000 --apiserver-port=8441 --wait=all --driver=docker  --container-runtime=docker: (1m7.26806832s)
--- PASS: TestFunctional/serial/StartWithProxy (67.27s)

                                                
                                    
x
+
TestFunctional/serial/AuditLog (0s)

                                                
                                                
=== RUN   TestFunctional/serial/AuditLog
--- PASS: TestFunctional/serial/AuditLog (0.00s)

                                                
                                    
x
+
TestFunctional/serial/SoftStart (32.63s)

                                                
                                                
=== RUN   TestFunctional/serial/SoftStart
functional_test.go:659: (dbg) Run:  out/minikube-linux-amd64 start -p functional-463208 --alsologtostderr -v=8
functional_test.go:659: (dbg) Done: out/minikube-linux-amd64 start -p functional-463208 --alsologtostderr -v=8: (32.628093134s)
functional_test.go:663: soft start took 32.629076302s for "functional-463208" cluster.
--- PASS: TestFunctional/serial/SoftStart (32.63s)

                                                
                                    
x
+
TestFunctional/serial/KubeContext (0.04s)

                                                
                                                
=== RUN   TestFunctional/serial/KubeContext
functional_test.go:681: (dbg) Run:  kubectl config current-context
--- PASS: TestFunctional/serial/KubeContext (0.04s)

                                                
                                    
x
+
TestFunctional/serial/KubectlGetPods (0.06s)

                                                
                                                
=== RUN   TestFunctional/serial/KubectlGetPods
functional_test.go:696: (dbg) Run:  kubectl --context functional-463208 get po -A
--- PASS: TestFunctional/serial/KubectlGetPods (0.06s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/add_remote (2.29s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/add_remote
functional_test.go:1049: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 cache add registry.k8s.io/pause:3.1
functional_test.go:1049: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 cache add registry.k8s.io/pause:3.3
functional_test.go:1049: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 cache add registry.k8s.io/pause:latest
--- PASS: TestFunctional/serial/CacheCmd/cache/add_remote (2.29s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/add_local (0.65s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/add_local
functional_test.go:1077: (dbg) Run:  docker build -t minikube-local-cache-test:functional-463208 /tmp/TestFunctionalserialCacheCmdcacheadd_local476022821/001
functional_test.go:1089: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 cache add minikube-local-cache-test:functional-463208
functional_test.go:1094: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 cache delete minikube-local-cache-test:functional-463208
functional_test.go:1083: (dbg) Run:  docker rmi minikube-local-cache-test:functional-463208
--- PASS: TestFunctional/serial/CacheCmd/cache/add_local (0.65s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/CacheDelete (0.04s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/CacheDelete
functional_test.go:1102: (dbg) Run:  out/minikube-linux-amd64 cache delete registry.k8s.io/pause:3.3
--- PASS: TestFunctional/serial/CacheCmd/cache/CacheDelete (0.04s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/list (0.04s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/list
functional_test.go:1110: (dbg) Run:  out/minikube-linux-amd64 cache list
--- PASS: TestFunctional/serial/CacheCmd/cache/list (0.04s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/verify_cache_inside_node (0.26s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/verify_cache_inside_node
functional_test.go:1124: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh sudo crictl images
--- PASS: TestFunctional/serial/CacheCmd/cache/verify_cache_inside_node (0.26s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/cache_reload (1.18s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/cache_reload
functional_test.go:1147: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh sudo docker rmi registry.k8s.io/pause:latest
functional_test.go:1153: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh sudo crictl inspecti registry.k8s.io/pause:latest
functional_test.go:1153: (dbg) Non-zero exit: out/minikube-linux-amd64 -p functional-463208 ssh sudo crictl inspecti registry.k8s.io/pause:latest: exit status 1 (249.699976ms)

                                                
                                                
-- stdout --
	FATA[0000] no such image "registry.k8s.io/pause:latest" present 

                                                
                                                
-- /stdout --
** stderr ** 
	ssh: Process exited with status 1

                                                
                                                
** /stderr **
functional_test.go:1158: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 cache reload
functional_test.go:1163: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh sudo crictl inspecti registry.k8s.io/pause:latest
--- PASS: TestFunctional/serial/CacheCmd/cache/cache_reload (1.18s)

                                                
                                    
x
+
TestFunctional/serial/CacheCmd/cache/delete (0.08s)

                                                
                                                
=== RUN   TestFunctional/serial/CacheCmd/cache/delete
functional_test.go:1172: (dbg) Run:  out/minikube-linux-amd64 cache delete registry.k8s.io/pause:3.1
functional_test.go:1172: (dbg) Run:  out/minikube-linux-amd64 cache delete registry.k8s.io/pause:latest
--- PASS: TestFunctional/serial/CacheCmd/cache/delete (0.08s)

                                                
                                    
x
+
TestFunctional/serial/MinikubeKubectlCmd (0.1s)

                                                
                                                
=== RUN   TestFunctional/serial/MinikubeKubectlCmd
functional_test.go:716: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 kubectl -- --context functional-463208 get pods
--- PASS: TestFunctional/serial/MinikubeKubectlCmd (0.10s)

                                                
                                    
x
+
TestFunctional/serial/MinikubeKubectlCmdDirectly (0.1s)

                                                
                                                
=== RUN   TestFunctional/serial/MinikubeKubectlCmdDirectly
functional_test.go:741: (dbg) Run:  out/kubectl --context functional-463208 get pods
--- PASS: TestFunctional/serial/MinikubeKubectlCmdDirectly (0.10s)

                                                
                                    
x
+
TestFunctional/serial/ExtraConfig (36.75s)

                                                
                                                
=== RUN   TestFunctional/serial/ExtraConfig
functional_test.go:757: (dbg) Run:  out/minikube-linux-amd64 start -p functional-463208 --extra-config=apiserver.enable-admission-plugins=NamespaceAutoProvision --wait=all
functional_test.go:757: (dbg) Done: out/minikube-linux-amd64 start -p functional-463208 --extra-config=apiserver.enable-admission-plugins=NamespaceAutoProvision --wait=all: (36.748507392s)
functional_test.go:761: restart took 36.748653492s for "functional-463208" cluster.
--- PASS: TestFunctional/serial/ExtraConfig (36.75s)

                                                
                                    
x
+
TestFunctional/serial/ComponentHealth (0.06s)

                                                
                                                
=== RUN   TestFunctional/serial/ComponentHealth
functional_test.go:810: (dbg) Run:  kubectl --context functional-463208 get po -l tier=control-plane -n kube-system -o=json
functional_test.go:825: etcd phase: Running
functional_test.go:835: etcd status: Ready
functional_test.go:825: kube-apiserver phase: Running
functional_test.go:835: kube-apiserver status: Ready
functional_test.go:825: kube-controller-manager phase: Running
functional_test.go:835: kube-controller-manager status: Ready
functional_test.go:825: kube-scheduler phase: Running
functional_test.go:835: kube-scheduler status: Ready
--- PASS: TestFunctional/serial/ComponentHealth (0.06s)

                                                
                                    
x
+
TestFunctional/serial/LogsCmd (0.86s)

                                                
                                                
=== RUN   TestFunctional/serial/LogsCmd
functional_test.go:1236: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 logs
--- PASS: TestFunctional/serial/LogsCmd (0.86s)

                                                
                                    
x
+
TestFunctional/serial/LogsFileCmd (0.9s)

                                                
                                                
=== RUN   TestFunctional/serial/LogsFileCmd
functional_test.go:1250: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 logs --file /tmp/TestFunctionalserialLogsFileCmd2843815083/001/logs.txt
--- PASS: TestFunctional/serial/LogsFileCmd (0.90s)

                                                
                                    
x
+
TestFunctional/serial/InvalidService (4.43s)

                                                
                                                
=== RUN   TestFunctional/serial/InvalidService
functional_test.go:2321: (dbg) Run:  kubectl --context functional-463208 apply -f testdata/invalidsvc.yaml
functional_test.go:2335: (dbg) Run:  out/minikube-linux-amd64 service invalid-svc -p functional-463208
functional_test.go:2335: (dbg) Non-zero exit: out/minikube-linux-amd64 service invalid-svc -p functional-463208: exit status 115 (297.855228ms)

                                                
                                                
-- stdout --
	|-----------|-------------|-------------|---------------------------|
	| NAMESPACE |    NAME     | TARGET PORT |            URL            |
	|-----------|-------------|-------------|---------------------------|
	| default   | invalid-svc |          80 | http://192.168.49.2:32307 |
	|-----------|-------------|-------------|---------------------------|
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	X Exiting due to SVC_UNREACHABLE: service not available: no running pod for service invalid-svc found
	* 
	╭─────────────────────────────────────────────────────────────────────────────────────────────╮
	│                                                                                             │
	│    * If the above advice does not help, please let us know:                                 │
	│      https://github.com/kubernetes/minikube/issues/new/choose                               │
	│                                                                                             │
	│    * Please run `minikube logs --file=logs.txt` and attach logs.txt to the GitHub issue.    │
	│    * Please also attach the following file to the GitHub issue:                             │
	│    * - /tmp/minikube_service_96b204199e3191fa1740d4430b018a3c8028d52d_0.log                 │
	│                                                                                             │
	╰─────────────────────────────────────────────────────────────────────────────────────────────╯

                                                
                                                
** /stderr **
functional_test.go:2327: (dbg) Run:  kubectl --context functional-463208 delete -f testdata/invalidsvc.yaml
--- PASS: TestFunctional/serial/InvalidService (4.43s)

                                                
                                    
x
+
TestFunctional/parallel/ConfigCmd (0.34s)

                                                
                                                
=== RUN   TestFunctional/parallel/ConfigCmd
=== PAUSE TestFunctional/parallel/ConfigCmd

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ConfigCmd
functional_test.go:1199: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 config unset cpus
functional_test.go:1199: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 config get cpus
functional_test.go:1199: (dbg) Non-zero exit: out/minikube-linux-amd64 -p functional-463208 config get cpus: exit status 14 (68.46242ms)

                                                
                                                
** stderr ** 
	Error: specified key could not be found in config

                                                
                                                
** /stderr **
functional_test.go:1199: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 config set cpus 2
functional_test.go:1199: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 config get cpus
functional_test.go:1199: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 config unset cpus
functional_test.go:1199: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 config get cpus
functional_test.go:1199: (dbg) Non-zero exit: out/minikube-linux-amd64 -p functional-463208 config get cpus: exit status 14 (43.842687ms)

                                                
                                                
** stderr ** 
	Error: specified key could not be found in config

                                                
                                                
** /stderr **
--- PASS: TestFunctional/parallel/ConfigCmd (0.34s)

                                                
                                    
x
+
TestFunctional/parallel/DashboardCmd (8.55s)

                                                
                                                
=== RUN   TestFunctional/parallel/DashboardCmd
=== PAUSE TestFunctional/parallel/DashboardCmd

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/DashboardCmd
functional_test.go:905: (dbg) daemon: [out/minikube-linux-amd64 dashboard --url --port 36195 -p functional-463208 --alsologtostderr -v=1]
functional_test.go:910: (dbg) stopping [out/minikube-linux-amd64 dashboard --url --port 36195 -p functional-463208 --alsologtostderr -v=1] ...
helpers_test.go:508: unable to kill pid 60948: os: process already finished
--- PASS: TestFunctional/parallel/DashboardCmd (8.55s)

                                                
                                    
x
+
TestFunctional/parallel/DryRun (0.43s)

                                                
                                                
=== RUN   TestFunctional/parallel/DryRun
=== PAUSE TestFunctional/parallel/DryRun

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/DryRun
functional_test.go:974: (dbg) Run:  out/minikube-linux-amd64 start -p functional-463208 --dry-run --memory 250MB --alsologtostderr --driver=docker  --container-runtime=docker
functional_test.go:974: (dbg) Non-zero exit: out/minikube-linux-amd64 start -p functional-463208 --dry-run --memory 250MB --alsologtostderr --driver=docker  --container-runtime=docker: exit status 23 (196.038322ms)

                                                
                                                
-- stdout --
	* [functional-463208] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
	  - MINIKUBE_LOCATION=19636
	  - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	  - KUBECONFIG=/home/jenkins/minikube-integration/19636-3724/kubeconfig
	  - MINIKUBE_HOME=/home/jenkins/minikube-integration/19636-3724/.minikube
	  - MINIKUBE_BIN=out/minikube-linux-amd64
	  - MINIKUBE_FORCE_SYSTEMD=
	* Using the docker driver based on existing profile
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	I0913 18:37:13.350408   60357 out.go:345] Setting OutFile to fd 1 ...
	I0913 18:37:13.350509   60357 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:37:13.350549   60357 out.go:358] Setting ErrFile to fd 2...
	I0913 18:37:13.350568   60357 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:37:13.350812   60357 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
	I0913 18:37:13.351412   60357 out.go:352] Setting JSON to false
	I0913 18:37:13.361154   60357 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-14","uptime":1184,"bootTime":1726251449,"procs":292,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1068-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
	I0913 18:37:13.361309   60357 start.go:139] virtualization: kvm guest
	I0913 18:37:13.369810   60357 out.go:177] * [functional-463208] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
	I0913 18:37:13.371490   60357 notify.go:220] Checking for updates...
	I0913 18:37:13.371575   60357 out.go:177]   - MINIKUBE_LOCATION=19636
	I0913 18:37:13.373025   60357 out.go:177]   - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0913 18:37:13.374497   60357 out.go:177]   - KUBECONFIG=/home/jenkins/minikube-integration/19636-3724/kubeconfig
	I0913 18:37:13.377240   60357 out.go:177]   - MINIKUBE_HOME=/home/jenkins/minikube-integration/19636-3724/.minikube
	I0913 18:37:13.378521   60357 out.go:177]   - MINIKUBE_BIN=out/minikube-linux-amd64
	I0913 18:37:13.379928   60357 out.go:177]   - MINIKUBE_FORCE_SYSTEMD=
	I0913 18:37:13.381605   60357 config.go:182] Loaded profile config "functional-463208": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
	I0913 18:37:13.382131   60357 driver.go:394] Setting default libvirt URI to qemu:///system
	I0913 18:37:13.410821   60357 docker.go:123] docker version: linux-27.2.1:Docker Engine - Community
	I0913 18:37:13.410901   60357 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0913 18:37:13.480673   60357 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:1 ContainersRunning:1 ContainersPaused:0 ContainersStopped:0 Images:2 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:33 OomKillDisable:true NGoroutines:53 SystemTime:2024-09-13 18:37:13.47174022 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86_
64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErr
ors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
	I0913 18:37:13.480774   60357 docker.go:318] overlay module found
	I0913 18:37:13.482323   60357 out.go:177] * Using the docker driver based on existing profile
	I0913 18:37:13.483327   60357 start.go:297] selected driver: docker
	I0913 18:37:13.483338   60357 start.go:901] validating driver "docker" against &{Name:functional-463208 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8441 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:functional-463208 Namespace:default APIServerHAVIP: APIServerNa
me:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[{Component:apiserver Key:enable-admission-plugins Value:NamespaceAutoProvision}] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8441 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker Mou
ntIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
	I0913 18:37:13.483418   60357 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
	I0913 18:37:13.485174   60357 out.go:201] 
	W0913 18:37:13.486331   60357 out.go:270] X Exiting due to RSRC_INSUFFICIENT_REQ_MEMORY: Requested memory allocation 250MiB is less than the usable minimum of 1800MB
	X Exiting due to RSRC_INSUFFICIENT_REQ_MEMORY: Requested memory allocation 250MiB is less than the usable minimum of 1800MB
	I0913 18:37:13.487408   60357 out.go:201] 

                                                
                                                
** /stderr **
functional_test.go:991: (dbg) Run:  out/minikube-linux-amd64 start -p functional-463208 --dry-run --alsologtostderr -v=1 --driver=docker  --container-runtime=docker
--- PASS: TestFunctional/parallel/DryRun (0.43s)

                                                
                                    
x
+
TestFunctional/parallel/InternationalLanguage (0.17s)

                                                
                                                
=== RUN   TestFunctional/parallel/InternationalLanguage
=== PAUSE TestFunctional/parallel/InternationalLanguage

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/InternationalLanguage
functional_test.go:1020: (dbg) Run:  out/minikube-linux-amd64 start -p functional-463208 --dry-run --memory 250MB --alsologtostderr --driver=docker  --container-runtime=docker
functional_test.go:1020: (dbg) Non-zero exit: out/minikube-linux-amd64 start -p functional-463208 --dry-run --memory 250MB --alsologtostderr --driver=docker  --container-runtime=docker: exit status 23 (173.372728ms)

                                                
                                                
-- stdout --
	* [functional-463208] minikube v1.34.0 sur Ubuntu 20.04 (kvm/amd64)
	  - MINIKUBE_LOCATION=19636
	  - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	  - KUBECONFIG=/home/jenkins/minikube-integration/19636-3724/kubeconfig
	  - MINIKUBE_HOME=/home/jenkins/minikube-integration/19636-3724/.minikube
	  - MINIKUBE_BIN=out/minikube-linux-amd64
	  - MINIKUBE_FORCE_SYSTEMD=
	* Utilisation du pilote docker basé sur le profil existant
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	I0913 18:37:13.173957   60209 out.go:345] Setting OutFile to fd 1 ...
	I0913 18:37:13.174063   60209 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:37:13.174072   60209 out.go:358] Setting ErrFile to fd 2...
	I0913 18:37:13.174076   60209 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:37:13.174342   60209 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
	I0913 18:37:13.174844   60209 out.go:352] Setting JSON to false
	I0913 18:37:13.175914   60209 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-14","uptime":1184,"bootTime":1726251449,"procs":296,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1068-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
	I0913 18:37:13.176021   60209 start.go:139] virtualization: kvm guest
	I0913 18:37:13.178461   60209 out.go:177] * [functional-463208] minikube v1.34.0 sur Ubuntu 20.04 (kvm/amd64)
	I0913 18:37:13.179637   60209 notify.go:220] Checking for updates...
	I0913 18:37:13.179644   60209 out.go:177]   - MINIKUBE_LOCATION=19636
	I0913 18:37:13.180857   60209 out.go:177]   - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	I0913 18:37:13.181975   60209 out.go:177]   - KUBECONFIG=/home/jenkins/minikube-integration/19636-3724/kubeconfig
	I0913 18:37:13.183221   60209 out.go:177]   - MINIKUBE_HOME=/home/jenkins/minikube-integration/19636-3724/.minikube
	I0913 18:37:13.185874   60209 out.go:177]   - MINIKUBE_BIN=out/minikube-linux-amd64
	I0913 18:37:13.187208   60209 out.go:177]   - MINIKUBE_FORCE_SYSTEMD=
	I0913 18:37:13.188890   60209 config.go:182] Loaded profile config "functional-463208": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
	I0913 18:37:13.189601   60209 driver.go:394] Setting default libvirt URI to qemu:///system
	I0913 18:37:13.216273   60209 docker.go:123] docker version: linux-27.2.1:Docker Engine - Community
	I0913 18:37:13.216368   60209 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0913 18:37:13.282120   60209 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:1 ContainersRunning:1 ContainersPaused:0 ContainersStopped:0 Images:2 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:33 OomKillDisable:true NGoroutines:53 SystemTime:2024-09-13 18:37:13.2668776 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86_6
4 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErro
rs:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
	I0913 18:37:13.282276   60209 docker.go:318] overlay module found
	I0913 18:37:13.284674   60209 out.go:177] * Utilisation du pilote docker basé sur le profil existant
	I0913 18:37:13.285829   60209 start.go:297] selected driver: docker
	I0913 18:37:13.285842   60209 start.go:901] validating driver "docker" against &{Name:functional-463208 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726193793-19634@sha256:4434bf9c4c4590e602ea482d2337d9d858a3db898bec2a85c17f78c81593c44e Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8441 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:functional-463208 Namespace:default APIServerHAVIP: APIServerNa
me:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[{Component:apiserver Key:enable-admission-plugins Value:NamespaceAutoProvision}] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8441 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker Mou
ntIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
	I0913 18:37:13.285971   60209 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
	I0913 18:37:13.287992   60209 out.go:201] 
	W0913 18:37:13.289184   60209 out.go:270] X Fermeture en raison de RSRC_INSUFFICIENT_REQ_MEMORY : L'allocation de mémoire demandée 250 Mio est inférieure au minimum utilisable de 1800 Mo
	X Fermeture en raison de RSRC_INSUFFICIENT_REQ_MEMORY : L'allocation de mémoire demandée 250 Mio est inférieure au minimum utilisable de 1800 Mo
	I0913 18:37:13.290492   60209 out.go:201] 

                                                
                                                
** /stderr **
--- PASS: TestFunctional/parallel/InternationalLanguage (0.17s)

                                                
                                    
x
+
TestFunctional/parallel/StatusCmd (1.13s)

                                                
                                                
=== RUN   TestFunctional/parallel/StatusCmd
=== PAUSE TestFunctional/parallel/StatusCmd

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/StatusCmd
functional_test.go:854: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 status
functional_test.go:860: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 status -f host:{{.Host}},kublet:{{.Kubelet}},apiserver:{{.APIServer}},kubeconfig:{{.Kubeconfig}}
functional_test.go:872: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 status -o json
--- PASS: TestFunctional/parallel/StatusCmd (1.13s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmdConnect (8.71s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmdConnect
=== PAUSE TestFunctional/parallel/ServiceCmdConnect

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ServiceCmdConnect
functional_test.go:1629: (dbg) Run:  kubectl --context functional-463208 create deployment hello-node-connect --image=registry.k8s.io/echoserver:1.8
functional_test.go:1635: (dbg) Run:  kubectl --context functional-463208 expose deployment hello-node-connect --type=NodePort --port=8080
functional_test.go:1640: (dbg) TestFunctional/parallel/ServiceCmdConnect: waiting 10m0s for pods matching "app=hello-node-connect" in namespace "default" ...
helpers_test.go:344: "hello-node-connect-67bdd5bbb4-4lfl9" [70449cf7-f1c9-497e-9c59-71af84a32eec] Pending / Ready:ContainersNotReady (containers with unready status: [echoserver]) / ContainersReady:ContainersNotReady (containers with unready status: [echoserver])
helpers_test.go:344: "hello-node-connect-67bdd5bbb4-4lfl9" [70449cf7-f1c9-497e-9c59-71af84a32eec] Running
functional_test.go:1640: (dbg) TestFunctional/parallel/ServiceCmdConnect: app=hello-node-connect healthy within 8.003810193s
functional_test.go:1649: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 service hello-node-connect --url
functional_test.go:1655: found endpoint for hello-node-connect: http://192.168.49.2:31254
functional_test.go:1675: http://192.168.49.2:31254: success! body:

                                                
                                                

                                                
                                                
Hostname: hello-node-connect-67bdd5bbb4-4lfl9

                                                
                                                
Pod Information:
	-no pod information available-

                                                
                                                
Server values:
	server_version=nginx: 1.13.3 - lua: 10008

                                                
                                                
Request Information:
	client_address=10.244.0.1
	method=GET
	real path=/
	query=
	request_version=1.1
	request_uri=http://192.168.49.2:8080/

                                                
                                                
Request Headers:
	accept-encoding=gzip
	host=192.168.49.2:31254
	user-agent=Go-http-client/1.1

                                                
                                                
Request Body:
	-no body in request-

                                                
                                                
--- PASS: TestFunctional/parallel/ServiceCmdConnect (8.71s)

                                                
                                    
x
+
TestFunctional/parallel/AddonsCmd (0.13s)

                                                
                                                
=== RUN   TestFunctional/parallel/AddonsCmd
=== PAUSE TestFunctional/parallel/AddonsCmd

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/AddonsCmd
functional_test.go:1690: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 addons list
functional_test.go:1702: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 addons list -o json
--- PASS: TestFunctional/parallel/AddonsCmd (0.13s)

                                                
                                    
x
+
TestFunctional/parallel/PersistentVolumeClaim (29.55s)

                                                
                                                
=== RUN   TestFunctional/parallel/PersistentVolumeClaim
=== PAUSE TestFunctional/parallel/PersistentVolumeClaim

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/PersistentVolumeClaim
functional_test_pvc_test.go:44: (dbg) TestFunctional/parallel/PersistentVolumeClaim: waiting 4m0s for pods matching "integration-test=storage-provisioner" in namespace "kube-system" ...
helpers_test.go:344: "storage-provisioner" [0a50e946-0374-41f2-b85e-a64be5d8adbb] Running
functional_test_pvc_test.go:44: (dbg) TestFunctional/parallel/PersistentVolumeClaim: integration-test=storage-provisioner healthy within 5.003960581s
functional_test_pvc_test.go:49: (dbg) Run:  kubectl --context functional-463208 get storageclass -o=json
functional_test_pvc_test.go:69: (dbg) Run:  kubectl --context functional-463208 apply -f testdata/storage-provisioner/pvc.yaml
functional_test_pvc_test.go:76: (dbg) Run:  kubectl --context functional-463208 get pvc myclaim -o=json
functional_test_pvc_test.go:125: (dbg) Run:  kubectl --context functional-463208 apply -f testdata/storage-provisioner/pod.yaml
functional_test_pvc_test.go:130: (dbg) TestFunctional/parallel/PersistentVolumeClaim: waiting 3m0s for pods matching "test=storage-provisioner" in namespace "default" ...
helpers_test.go:344: "sp-pod" [a251c1f2-d90e-473e-b20e-e49f44a542ca] Pending
helpers_test.go:344: "sp-pod" [a251c1f2-d90e-473e-b20e-e49f44a542ca] Pending / Ready:ContainersNotReady (containers with unready status: [myfrontend]) / ContainersReady:ContainersNotReady (containers with unready status: [myfrontend])
helpers_test.go:344: "sp-pod" [a251c1f2-d90e-473e-b20e-e49f44a542ca] Running
functional_test_pvc_test.go:130: (dbg) TestFunctional/parallel/PersistentVolumeClaim: test=storage-provisioner healthy within 13.003428379s
functional_test_pvc_test.go:100: (dbg) Run:  kubectl --context functional-463208 exec sp-pod -- touch /tmp/mount/foo
functional_test_pvc_test.go:106: (dbg) Run:  kubectl --context functional-463208 delete -f testdata/storage-provisioner/pod.yaml
functional_test_pvc_test.go:106: (dbg) Done: kubectl --context functional-463208 delete -f testdata/storage-provisioner/pod.yaml: (2.735825292s)
functional_test_pvc_test.go:125: (dbg) Run:  kubectl --context functional-463208 apply -f testdata/storage-provisioner/pod.yaml
functional_test_pvc_test.go:130: (dbg) TestFunctional/parallel/PersistentVolumeClaim: waiting 3m0s for pods matching "test=storage-provisioner" in namespace "default" ...
helpers_test.go:344: "sp-pod" [b3238608-4c3c-41d7-a897-40c9c9c1992c] Pending
helpers_test.go:344: "sp-pod" [b3238608-4c3c-41d7-a897-40c9c9c1992c] Pending / Ready:ContainersNotReady (containers with unready status: [myfrontend]) / ContainersReady:ContainersNotReady (containers with unready status: [myfrontend])
helpers_test.go:344: "sp-pod" [b3238608-4c3c-41d7-a897-40c9c9c1992c] Running
functional_test_pvc_test.go:130: (dbg) TestFunctional/parallel/PersistentVolumeClaim: test=storage-provisioner healthy within 8.003710642s
functional_test_pvc_test.go:114: (dbg) Run:  kubectl --context functional-463208 exec sp-pod -- ls /tmp/mount
--- PASS: TestFunctional/parallel/PersistentVolumeClaim (29.55s)

                                                
                                    
x
+
TestFunctional/parallel/SSHCmd (0.58s)

                                                
                                                
=== RUN   TestFunctional/parallel/SSHCmd
=== PAUSE TestFunctional/parallel/SSHCmd

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/SSHCmd
functional_test.go:1725: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "echo hello"
functional_test.go:1742: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "cat /etc/hostname"
--- PASS: TestFunctional/parallel/SSHCmd (0.58s)

                                                
                                    
x
+
TestFunctional/parallel/CpCmd (1.77s)

                                                
                                                
=== RUN   TestFunctional/parallel/CpCmd
=== PAUSE TestFunctional/parallel/CpCmd

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/CpCmd
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 cp testdata/cp-test.txt /home/docker/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh -n functional-463208 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 cp functional-463208:/home/docker/cp-test.txt /tmp/TestFunctionalparallelCpCmd3719662857/001/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh -n functional-463208 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 cp testdata/cp-test.txt /tmp/does/not/exist/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh -n functional-463208 "sudo cat /tmp/does/not/exist/cp-test.txt"
--- PASS: TestFunctional/parallel/CpCmd (1.77s)

                                                
                                    
x
+
TestFunctional/parallel/MySQL (20.79s)

                                                
                                                
=== RUN   TestFunctional/parallel/MySQL
=== PAUSE TestFunctional/parallel/MySQL

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/MySQL
functional_test.go:1793: (dbg) Run:  kubectl --context functional-463208 replace --force -f testdata/mysql.yaml
functional_test.go:1799: (dbg) TestFunctional/parallel/MySQL: waiting 10m0s for pods matching "app=mysql" in namespace "default" ...
helpers_test.go:344: "mysql-6cdb49bbb-pqmcg" [69b33f20-ccfd-4c9b-a8d1-a3a1abd1c6f8] Pending / Ready:ContainersNotReady (containers with unready status: [mysql]) / ContainersReady:ContainersNotReady (containers with unready status: [mysql])
helpers_test.go:344: "mysql-6cdb49bbb-pqmcg" [69b33f20-ccfd-4c9b-a8d1-a3a1abd1c6f8] Running
functional_test.go:1799: (dbg) TestFunctional/parallel/MySQL: app=mysql healthy within 18.003717648s
functional_test.go:1807: (dbg) Run:  kubectl --context functional-463208 exec mysql-6cdb49bbb-pqmcg -- mysql -ppassword -e "show databases;"
functional_test.go:1807: (dbg) Non-zero exit: kubectl --context functional-463208 exec mysql-6cdb49bbb-pqmcg -- mysql -ppassword -e "show databases;": exit status 1 (105.343513ms)

                                                
                                                
** stderr ** 
	mysql: [Warning] Using a password on the command line interface can be insecure.
	ERROR 2002 (HY000): Can't connect to local MySQL server through socket '/var/run/mysqld/mysqld.sock' (2)
	command terminated with exit code 1

                                                
                                                
** /stderr **
functional_test.go:1807: (dbg) Run:  kubectl --context functional-463208 exec mysql-6cdb49bbb-pqmcg -- mysql -ppassword -e "show databases;"
functional_test.go:1807: (dbg) Non-zero exit: kubectl --context functional-463208 exec mysql-6cdb49bbb-pqmcg -- mysql -ppassword -e "show databases;": exit status 1 (157.252921ms)

                                                
                                                
** stderr ** 
	mysql: [Warning] Using a password on the command line interface can be insecure.
	ERROR 2002 (HY000): Can't connect to local MySQL server through socket '/var/run/mysqld/mysqld.sock' (2)
	command terminated with exit code 1

                                                
                                                
** /stderr **
functional_test.go:1807: (dbg) Run:  kubectl --context functional-463208 exec mysql-6cdb49bbb-pqmcg -- mysql -ppassword -e "show databases;"
--- PASS: TestFunctional/parallel/MySQL (20.79s)

                                                
                                    
x
+
TestFunctional/parallel/FileSync (0.28s)

                                                
                                                
=== RUN   TestFunctional/parallel/FileSync
=== PAUSE TestFunctional/parallel/FileSync

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/FileSync
functional_test.go:1929: Checking for existence of /etc/test/nested/copy/10475/hosts within VM
functional_test.go:1931: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "sudo cat /etc/test/nested/copy/10475/hosts"
functional_test.go:1936: file sync test content: Test file for checking file sync process
--- PASS: TestFunctional/parallel/FileSync (0.28s)

                                                
                                    
x
+
TestFunctional/parallel/CertSync (1.76s)

                                                
                                                
=== RUN   TestFunctional/parallel/CertSync
=== PAUSE TestFunctional/parallel/CertSync

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/CertSync
functional_test.go:1972: Checking for existence of /etc/ssl/certs/10475.pem within VM
functional_test.go:1973: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "sudo cat /etc/ssl/certs/10475.pem"
functional_test.go:1972: Checking for existence of /usr/share/ca-certificates/10475.pem within VM
functional_test.go:1973: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "sudo cat /usr/share/ca-certificates/10475.pem"
functional_test.go:1972: Checking for existence of /etc/ssl/certs/51391683.0 within VM
functional_test.go:1973: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "sudo cat /etc/ssl/certs/51391683.0"
functional_test.go:1999: Checking for existence of /etc/ssl/certs/104752.pem within VM
functional_test.go:2000: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "sudo cat /etc/ssl/certs/104752.pem"
functional_test.go:1999: Checking for existence of /usr/share/ca-certificates/104752.pem within VM
functional_test.go:2000: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "sudo cat /usr/share/ca-certificates/104752.pem"
functional_test.go:1999: Checking for existence of /etc/ssl/certs/3ec20f2e.0 within VM
functional_test.go:2000: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "sudo cat /etc/ssl/certs/3ec20f2e.0"
--- PASS: TestFunctional/parallel/CertSync (1.76s)

                                                
                                    
x
+
TestFunctional/parallel/NodeLabels (0.08s)

                                                
                                                
=== RUN   TestFunctional/parallel/NodeLabels
=== PAUSE TestFunctional/parallel/NodeLabels

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/NodeLabels
functional_test.go:219: (dbg) Run:  kubectl --context functional-463208 get nodes --output=go-template "--template='{{range $k, $v := (index .items 0).metadata.labels}}{{$k}} {{end}}'"
--- PASS: TestFunctional/parallel/NodeLabels (0.08s)

                                                
                                    
x
+
TestFunctional/parallel/NonActiveRuntimeDisabled (0.28s)

                                                
                                                
=== RUN   TestFunctional/parallel/NonActiveRuntimeDisabled
=== PAUSE TestFunctional/parallel/NonActiveRuntimeDisabled

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/NonActiveRuntimeDisabled
functional_test.go:2027: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "sudo systemctl is-active crio"
functional_test.go:2027: (dbg) Non-zero exit: out/minikube-linux-amd64 -p functional-463208 ssh "sudo systemctl is-active crio": exit status 1 (275.805641ms)

                                                
                                                
-- stdout --
	inactive

                                                
                                                
-- /stdout --
** stderr ** 
	ssh: Process exited with status 3

                                                
                                                
** /stderr **
--- PASS: TestFunctional/parallel/NonActiveRuntimeDisabled (0.28s)

                                                
                                    
x
+
TestFunctional/parallel/License (0.17s)

                                                
                                                
=== RUN   TestFunctional/parallel/License
=== PAUSE TestFunctional/parallel/License

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/License
functional_test.go:2288: (dbg) Run:  out/minikube-linux-amd64 license
--- PASS: TestFunctional/parallel/License (0.17s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmd/DeployApp (10.19s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmd/DeployApp
functional_test.go:1439: (dbg) Run:  kubectl --context functional-463208 create deployment hello-node --image=registry.k8s.io/echoserver:1.8
functional_test.go:1445: (dbg) Run:  kubectl --context functional-463208 expose deployment hello-node --type=NodePort --port=8080
functional_test.go:1450: (dbg) TestFunctional/parallel/ServiceCmd/DeployApp: waiting 10m0s for pods matching "app=hello-node" in namespace "default" ...
helpers_test.go:344: "hello-node-6b9f76b5c7-wzrg7" [54542e9d-eed3-4683-9c94-f5a25334345c] Pending / Ready:ContainersNotReady (containers with unready status: [echoserver]) / ContainersReady:ContainersNotReady (containers with unready status: [echoserver])
helpers_test.go:344: "hello-node-6b9f76b5c7-wzrg7" [54542e9d-eed3-4683-9c94-f5a25334345c] Running
functional_test.go:1450: (dbg) TestFunctional/parallel/ServiceCmd/DeployApp: app=hello-node healthy within 10.004544257s
--- PASS: TestFunctional/parallel/ServiceCmd/DeployApp (10.19s)

                                                
                                    
x
+
TestFunctional/parallel/ProfileCmd/profile_not_create (0.43s)

                                                
                                                
=== RUN   TestFunctional/parallel/ProfileCmd/profile_not_create
functional_test.go:1270: (dbg) Run:  out/minikube-linux-amd64 profile lis
functional_test.go:1275: (dbg) Run:  out/minikube-linux-amd64 profile list --output json
--- PASS: TestFunctional/parallel/ProfileCmd/profile_not_create (0.43s)

                                                
                                    
x
+
TestFunctional/parallel/MountCmd/any-port (7.82s)

                                                
                                                
=== RUN   TestFunctional/parallel/MountCmd/any-port
functional_test_mount_test.go:73: (dbg) daemon: [out/minikube-linux-amd64 mount -p functional-463208 /tmp/TestFunctionalparallelMountCmdany-port3607540149/001:/mount-9p --alsologtostderr -v=1]
functional_test_mount_test.go:107: wrote "test-1726252631692913111" to /tmp/TestFunctionalparallelMountCmdany-port3607540149/001/created-by-test
functional_test_mount_test.go:107: wrote "test-1726252631692913111" to /tmp/TestFunctionalparallelMountCmdany-port3607540149/001/created-by-test-removed-by-pod
functional_test_mount_test.go:107: wrote "test-1726252631692913111" to /tmp/TestFunctionalparallelMountCmdany-port3607540149/001/test-1726252631692913111
functional_test_mount_test.go:115: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "findmnt -T /mount-9p | grep 9p"
functional_test_mount_test.go:115: (dbg) Non-zero exit: out/minikube-linux-amd64 -p functional-463208 ssh "findmnt -T /mount-9p | grep 9p": exit status 1 (311.966676ms)

                                                
                                                
** stderr ** 
	ssh: Process exited with status 1

                                                
                                                
** /stderr **
functional_test_mount_test.go:115: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "findmnt -T /mount-9p | grep 9p"
functional_test_mount_test.go:129: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh -- ls -la /mount-9p
functional_test_mount_test.go:133: guest mount directory contents
total 2
-rw-r--r-- 1 docker docker 24 Sep 13 18:37 created-by-test
-rw-r--r-- 1 docker docker 24 Sep 13 18:37 created-by-test-removed-by-pod
-rw-r--r-- 1 docker docker 24 Sep 13 18:37 test-1726252631692913111
functional_test_mount_test.go:137: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh cat /mount-9p/test-1726252631692913111
functional_test_mount_test.go:148: (dbg) Run:  kubectl --context functional-463208 replace --force -f testdata/busybox-mount-test.yaml
functional_test_mount_test.go:153: (dbg) TestFunctional/parallel/MountCmd/any-port: waiting 4m0s for pods matching "integration-test=busybox-mount" in namespace "default" ...
helpers_test.go:344: "busybox-mount" [013a2c61-9ea5-47ba-bdb9-b0dbca8095a5] Pending / Ready:ContainersNotReady (containers with unready status: [mount-munger]) / ContainersReady:ContainersNotReady (containers with unready status: [mount-munger])
helpers_test.go:344: "busybox-mount" [013a2c61-9ea5-47ba-bdb9-b0dbca8095a5] Pending / Initialized:PodCompleted / Ready:PodCompleted / ContainersReady:PodCompleted
helpers_test.go:344: "busybox-mount" [013a2c61-9ea5-47ba-bdb9-b0dbca8095a5] Succeeded / Initialized:PodCompleted / Ready:PodCompleted / ContainersReady:PodCompleted
functional_test_mount_test.go:153: (dbg) TestFunctional/parallel/MountCmd/any-port: integration-test=busybox-mount healthy within 5.003889935s
functional_test_mount_test.go:169: (dbg) Run:  kubectl --context functional-463208 logs busybox-mount
functional_test_mount_test.go:181: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh stat /mount-9p/created-by-test
functional_test_mount_test.go:181: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh stat /mount-9p/created-by-pod
functional_test_mount_test.go:90: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "sudo umount -f /mount-9p"
functional_test_mount_test.go:94: (dbg) stopping [out/minikube-linux-amd64 mount -p functional-463208 /tmp/TestFunctionalparallelMountCmdany-port3607540149/001:/mount-9p --alsologtostderr -v=1] ...
--- PASS: TestFunctional/parallel/MountCmd/any-port (7.82s)

                                                
                                    
x
+
TestFunctional/parallel/ProfileCmd/profile_list (0.4s)

                                                
                                                
=== RUN   TestFunctional/parallel/ProfileCmd/profile_list
functional_test.go:1310: (dbg) Run:  out/minikube-linux-amd64 profile list
functional_test.go:1315: Took "358.237161ms" to run "out/minikube-linux-amd64 profile list"
functional_test.go:1324: (dbg) Run:  out/minikube-linux-amd64 profile list -l
functional_test.go:1329: Took "45.738686ms" to run "out/minikube-linux-amd64 profile list -l"
--- PASS: TestFunctional/parallel/ProfileCmd/profile_list (0.40s)

                                                
                                    
x
+
TestFunctional/parallel/ProfileCmd/profile_json_output (0.36s)

                                                
                                                
=== RUN   TestFunctional/parallel/ProfileCmd/profile_json_output
functional_test.go:1361: (dbg) Run:  out/minikube-linux-amd64 profile list -o json
functional_test.go:1366: Took "301.565245ms" to run "out/minikube-linux-amd64 profile list -o json"
functional_test.go:1374: (dbg) Run:  out/minikube-linux-amd64 profile list -o json --light
functional_test.go:1379: Took "57.700868ms" to run "out/minikube-linux-amd64 profile list -o json --light"
--- PASS: TestFunctional/parallel/ProfileCmd/profile_json_output (0.36s)

                                                
                                    
x
+
TestFunctional/parallel/MountCmd/specific-port (1.79s)

                                                
                                                
=== RUN   TestFunctional/parallel/MountCmd/specific-port
functional_test_mount_test.go:213: (dbg) daemon: [out/minikube-linux-amd64 mount -p functional-463208 /tmp/TestFunctionalparallelMountCmdspecific-port3271074210/001:/mount-9p --alsologtostderr -v=1 --port 46464]
functional_test_mount_test.go:243: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "findmnt -T /mount-9p | grep 9p"
functional_test_mount_test.go:243: (dbg) Non-zero exit: out/minikube-linux-amd64 -p functional-463208 ssh "findmnt -T /mount-9p | grep 9p": exit status 1 (278.69168ms)

                                                
                                                
** stderr ** 
	ssh: Process exited with status 1

                                                
                                                
** /stderr **
functional_test_mount_test.go:243: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "findmnt -T /mount-9p | grep 9p"
functional_test_mount_test.go:257: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh -- ls -la /mount-9p
functional_test_mount_test.go:261: guest mount directory contents
total 0
functional_test_mount_test.go:263: (dbg) stopping [out/minikube-linux-amd64 mount -p functional-463208 /tmp/TestFunctionalparallelMountCmdspecific-port3271074210/001:/mount-9p --alsologtostderr -v=1 --port 46464] ...
functional_test_mount_test.go:264: reading mount text
functional_test_mount_test.go:278: done reading mount text
functional_test_mount_test.go:230: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "sudo umount -f /mount-9p"
functional_test_mount_test.go:230: (dbg) Non-zero exit: out/minikube-linux-amd64 -p functional-463208 ssh "sudo umount -f /mount-9p": exit status 1 (253.891389ms)

                                                
                                                
-- stdout --
	umount: /mount-9p: not mounted.

                                                
                                                
-- /stdout --
** stderr ** 
	ssh: Process exited with status 32

                                                
                                                
** /stderr **
functional_test_mount_test.go:232: "out/minikube-linux-amd64 -p functional-463208 ssh \"sudo umount -f /mount-9p\"": exit status 1
functional_test_mount_test.go:234: (dbg) stopping [out/minikube-linux-amd64 mount -p functional-463208 /tmp/TestFunctionalparallelMountCmdspecific-port3271074210/001:/mount-9p --alsologtostderr -v=1 --port 46464] ...
--- PASS: TestFunctional/parallel/MountCmd/specific-port (1.79s)

                                                
                                    
x
+
TestFunctional/parallel/MountCmd/VerifyCleanup (2.12s)

                                                
                                                
=== RUN   TestFunctional/parallel/MountCmd/VerifyCleanup
functional_test_mount_test.go:298: (dbg) daemon: [out/minikube-linux-amd64 mount -p functional-463208 /tmp/TestFunctionalparallelMountCmdVerifyCleanup2346944980/001:/mount1 --alsologtostderr -v=1]
functional_test_mount_test.go:298: (dbg) daemon: [out/minikube-linux-amd64 mount -p functional-463208 /tmp/TestFunctionalparallelMountCmdVerifyCleanup2346944980/001:/mount2 --alsologtostderr -v=1]
functional_test_mount_test.go:298: (dbg) daemon: [out/minikube-linux-amd64 mount -p functional-463208 /tmp/TestFunctionalparallelMountCmdVerifyCleanup2346944980/001:/mount3 --alsologtostderr -v=1]
functional_test_mount_test.go:325: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "findmnt -T" /mount1
functional_test_mount_test.go:325: (dbg) Non-zero exit: out/minikube-linux-amd64 -p functional-463208 ssh "findmnt -T" /mount1: exit status 1 (402.903959ms)

                                                
                                                
** stderr ** 
	ssh: Process exited with status 1

                                                
                                                
** /stderr **
2024/09/13 18:37:22 [DEBUG] GET http://127.0.0.1:36195/api/v1/namespaces/kubernetes-dashboard/services/http:kubernetes-dashboard:/proxy/
functional_test_mount_test.go:325: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "findmnt -T" /mount1
functional_test_mount_test.go:325: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "findmnt -T" /mount2
functional_test_mount_test.go:325: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh "findmnt -T" /mount3
functional_test_mount_test.go:370: (dbg) Run:  out/minikube-linux-amd64 mount -p functional-463208 --kill=true
functional_test_mount_test.go:313: (dbg) stopping [out/minikube-linux-amd64 mount -p functional-463208 /tmp/TestFunctionalparallelMountCmdVerifyCleanup2346944980/001:/mount1 --alsologtostderr -v=1] ...
helpers_test.go:490: unable to find parent, assuming dead: process does not exist
functional_test_mount_test.go:313: (dbg) stopping [out/minikube-linux-amd64 mount -p functional-463208 /tmp/TestFunctionalparallelMountCmdVerifyCleanup2346944980/001:/mount2 --alsologtostderr -v=1] ...
helpers_test.go:490: unable to find parent, assuming dead: process does not exist
functional_test_mount_test.go:313: (dbg) stopping [out/minikube-linux-amd64 mount -p functional-463208 /tmp/TestFunctionalparallelMountCmdVerifyCleanup2346944980/001:/mount3 --alsologtostderr -v=1] ...
helpers_test.go:490: unable to find parent, assuming dead: process does not exist
--- PASS: TestFunctional/parallel/MountCmd/VerifyCleanup (2.12s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmd/List (0.61s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmd/List
functional_test.go:1459: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 service list
--- PASS: TestFunctional/parallel/ServiceCmd/List (0.61s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmd/JSONOutput (0.51s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmd/JSONOutput
functional_test.go:1489: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 service list -o json
functional_test.go:1494: Took "505.199937ms" to run "out/minikube-linux-amd64 -p functional-463208 service list -o json"
--- PASS: TestFunctional/parallel/ServiceCmd/JSONOutput (0.51s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmd/HTTPS (0.37s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmd/HTTPS
functional_test.go:1509: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 service --namespace=default --https --url hello-node
functional_test.go:1522: found endpoint: https://192.168.49.2:31253
--- PASS: TestFunctional/parallel/ServiceCmd/HTTPS (0.37s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/RunSecondTunnel (0.46s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/RunSecondTunnel
functional_test_tunnel_test.go:154: (dbg) daemon: [out/minikube-linux-amd64 -p functional-463208 tunnel --alsologtostderr]
functional_test_tunnel_test.go:154: (dbg) daemon: [out/minikube-linux-amd64 -p functional-463208 tunnel --alsologtostderr]
functional_test_tunnel_test.go:194: (dbg) stopping [out/minikube-linux-amd64 -p functional-463208 tunnel --alsologtostderr] ...
helpers_test.go:508: unable to kill pid 63260: os: process already finished
functional_test_tunnel_test.go:194: (dbg) stopping [out/minikube-linux-amd64 -p functional-463208 tunnel --alsologtostderr] ...
helpers_test.go:490: unable to find parent, assuming dead: process does not exist
--- PASS: TestFunctional/parallel/TunnelCmd/serial/RunSecondTunnel (0.46s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmd/Format (0.41s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmd/Format
functional_test.go:1540: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 service hello-node --url --format={{.IP}}
--- PASS: TestFunctional/parallel/ServiceCmd/Format (0.41s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/StartTunnel (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/StartTunnel
functional_test_tunnel_test.go:129: (dbg) daemon: [out/minikube-linux-amd64 -p functional-463208 tunnel --alsologtostderr]
--- PASS: TestFunctional/parallel/TunnelCmd/serial/StartTunnel (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/WaitService/Setup (9.26s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/WaitService/Setup
functional_test_tunnel_test.go:212: (dbg) Run:  kubectl --context functional-463208 apply -f testdata/testsvc.yaml
functional_test_tunnel_test.go:216: (dbg) TestFunctional/parallel/TunnelCmd/serial/WaitService/Setup: waiting 4m0s for pods matching "run=nginx-svc" in namespace "default" ...
helpers_test.go:344: "nginx-svc" [c01bad4f-4b30-4d60-8425-7e2adab15cfb] Pending / Ready:ContainersNotReady (containers with unready status: [nginx]) / ContainersReady:ContainersNotReady (containers with unready status: [nginx])
helpers_test.go:344: "nginx-svc" [c01bad4f-4b30-4d60-8425-7e2adab15cfb] Running
functional_test_tunnel_test.go:216: (dbg) TestFunctional/parallel/TunnelCmd/serial/WaitService/Setup: run=nginx-svc healthy within 9.00433163s
--- PASS: TestFunctional/parallel/TunnelCmd/serial/WaitService/Setup (9.26s)

                                                
                                    
x
+
TestFunctional/parallel/ServiceCmd/URL (0.56s)

                                                
                                                
=== RUN   TestFunctional/parallel/ServiceCmd/URL
functional_test.go:1559: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 service hello-node --url
functional_test.go:1565: found endpoint for hello-node: http://192.168.49.2:31253
--- PASS: TestFunctional/parallel/ServiceCmd/URL (0.56s)

                                                
                                    
x
+
TestFunctional/parallel/DockerEnv/bash (1.15s)

                                                
                                                
=== RUN   TestFunctional/parallel/DockerEnv/bash
functional_test.go:499: (dbg) Run:  /bin/bash -c "eval $(out/minikube-linux-amd64 -p functional-463208 docker-env) && out/minikube-linux-amd64 status -p functional-463208"
functional_test.go:522: (dbg) Run:  /bin/bash -c "eval $(out/minikube-linux-amd64 -p functional-463208 docker-env) && docker images"
--- PASS: TestFunctional/parallel/DockerEnv/bash (1.15s)

                                                
                                    
x
+
TestFunctional/parallel/UpdateContextCmd/no_changes (0.15s)

                                                
                                                
=== RUN   TestFunctional/parallel/UpdateContextCmd/no_changes
=== PAUSE TestFunctional/parallel/UpdateContextCmd/no_changes

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/UpdateContextCmd/no_changes
functional_test.go:2119: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 update-context --alsologtostderr -v=2
--- PASS: TestFunctional/parallel/UpdateContextCmd/no_changes (0.15s)

                                                
                                    
x
+
TestFunctional/parallel/UpdateContextCmd/no_minikube_cluster (0.14s)

                                                
                                                
=== RUN   TestFunctional/parallel/UpdateContextCmd/no_minikube_cluster
=== PAUSE TestFunctional/parallel/UpdateContextCmd/no_minikube_cluster

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/UpdateContextCmd/no_minikube_cluster
functional_test.go:2119: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 update-context --alsologtostderr -v=2
--- PASS: TestFunctional/parallel/UpdateContextCmd/no_minikube_cluster (0.14s)

                                                
                                    
x
+
TestFunctional/parallel/UpdateContextCmd/no_clusters (0.14s)

                                                
                                                
=== RUN   TestFunctional/parallel/UpdateContextCmd/no_clusters
=== PAUSE TestFunctional/parallel/UpdateContextCmd/no_clusters

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/UpdateContextCmd/no_clusters
functional_test.go:2119: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 update-context --alsologtostderr -v=2
--- PASS: TestFunctional/parallel/UpdateContextCmd/no_clusters (0.14s)

                                                
                                    
x
+
TestFunctional/parallel/Version/short (0.06s)

                                                
                                                
=== RUN   TestFunctional/parallel/Version/short
=== PAUSE TestFunctional/parallel/Version/short

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/Version/short
functional_test.go:2256: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 version --short
--- PASS: TestFunctional/parallel/Version/short (0.06s)

                                                
                                    
x
+
TestFunctional/parallel/Version/components (0.72s)

                                                
                                                
=== RUN   TestFunctional/parallel/Version/components
=== PAUSE TestFunctional/parallel/Version/components

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/Version/components
functional_test.go:2270: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 version -o=json --components
--- PASS: TestFunctional/parallel/Version/components (0.72s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/WaitService/IngressIP (0.06s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/WaitService/IngressIP
functional_test_tunnel_test.go:234: (dbg) Run:  kubectl --context functional-463208 get svc nginx-svc -o jsonpath={.status.loadBalancer.ingress[0].ip}
--- PASS: TestFunctional/parallel/TunnelCmd/serial/WaitService/IngressIP (0.06s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/AccessDirect (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/AccessDirect
functional_test_tunnel_test.go:299: tunnel at http://10.102.246.51 is working!
--- PASS: TestFunctional/parallel/TunnelCmd/serial/AccessDirect (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/DeleteTunnel (0.11s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/DeleteTunnel
functional_test_tunnel_test.go:434: (dbg) stopping [out/minikube-linux-amd64 -p functional-463208 tunnel --alsologtostderr] ...
--- PASS: TestFunctional/parallel/TunnelCmd/serial/DeleteTunnel (0.11s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageListShort (0.2s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageListShort
=== PAUSE TestFunctional/parallel/ImageCommands/ImageListShort

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ImageCommands/ImageListShort
functional_test.go:261: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image ls --format short --alsologtostderr
functional_test.go:266: (dbg) Stdout: out/minikube-linux-amd64 -p functional-463208 image ls --format short --alsologtostderr:
registry.k8s.io/pause:latest
registry.k8s.io/pause:3.3
registry.k8s.io/pause:3.10
registry.k8s.io/pause:3.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/echoserver:1.8
registry.k8s.io/coredns/coredns:v1.11.3
gcr.io/k8s-minikube/storage-provisioner:v5
gcr.io/k8s-minikube/busybox:1.28.4-glibc
docker.io/library/nginx:latest
docker.io/library/nginx:alpine
docker.io/library/mysql:5.7
docker.io/library/minikube-local-cache-test:functional-463208
docker.io/kubernetesui/metrics-scraper:<none>
docker.io/kubernetesui/dashboard:<none>
docker.io/kicbase/echo-server:functional-463208
functional_test.go:269: (dbg) Stderr: out/minikube-linux-amd64 -p functional-463208 image ls --format short --alsologtostderr:
I0913 18:37:38.228064   67482 out.go:345] Setting OutFile to fd 1 ...
I0913 18:37:38.228193   67482 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0913 18:37:38.228204   67482 out.go:358] Setting ErrFile to fd 2...
I0913 18:37:38.228210   67482 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0913 18:37:38.228422   67482 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
I0913 18:37:38.229076   67482 config.go:182] Loaded profile config "functional-463208": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0913 18:37:38.229252   67482 config.go:182] Loaded profile config "functional-463208": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0913 18:37:38.229879   67482 cli_runner.go:164] Run: docker container inspect functional-463208 --format={{.State.Status}}
I0913 18:37:38.248240   67482 ssh_runner.go:195] Run: systemctl --version
I0913 18:37:38.248282   67482 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" functional-463208
I0913 18:37:38.267401   67482 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32778 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/functional-463208/id_rsa Username:docker}
I0913 18:37:38.353913   67482 ssh_runner.go:195] Run: docker images --no-trunc --format "{{json .}}"
--- PASS: TestFunctional/parallel/ImageCommands/ImageListShort (0.20s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageListTable (0.19s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageListTable
=== PAUSE TestFunctional/parallel/ImageCommands/ImageListTable

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ImageCommands/ImageListTable
functional_test.go:261: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image ls --format table --alsologtostderr
functional_test.go:266: (dbg) Stdout: out/minikube-linux-amd64 -p functional-463208 image ls --format table --alsologtostderr:
|---------------------------------------------|-------------------|---------------|--------|
|                    Image                    |        Tag        |   Image ID    |  Size  |
|---------------------------------------------|-------------------|---------------|--------|
| registry.k8s.io/pause                       | 3.10              | 873ed75102791 | 736kB  |
| gcr.io/k8s-minikube/storage-provisioner     | v5                | 6e38f40d628db | 31.5MB |
| docker.io/library/nginx                     | latest            | 39286ab8a5e14 | 188MB  |
| registry.k8s.io/etcd                        | 3.5.15-0          | 2e96e5913fc06 | 148MB  |
| registry.k8s.io/pause                       | latest            | 350b164e7ae1d | 240kB  |
| docker.io/kubernetesui/dashboard            | <none>            | 07655ddf2eebe | 246MB  |
| docker.io/kubernetesui/metrics-scraper      | <none>            | 115053965e86b | 43.8MB |
| registry.k8s.io/kube-proxy                  | v1.31.1           | 60c005f310ff3 | 91.5MB |
| docker.io/library/nginx                     | alpine            | c7b4f26a7d93f | 43.2MB |
| registry.k8s.io/pause                       | 3.3               | 0184c1613d929 | 683kB  |
| gcr.io/k8s-minikube/busybox                 | 1.28.4-glibc      | 56cc512116c8f | 4.4MB  |
| docker.io/library/minikube-local-cache-test | functional-463208 | bef76f5f2e90f | 30B    |
| registry.k8s.io/kube-controller-manager     | v1.31.1           | 175ffd71cce3d | 88.4MB |
| registry.k8s.io/coredns/coredns             | v1.11.3           | c69fa2e9cbf5f | 61.8MB |
| docker.io/library/mysql                     | 5.7               | 5107333e08a87 | 501MB  |
| docker.io/kicbase/echo-server               | functional-463208 | 9056ab77afb8e | 4.94MB |
| registry.k8s.io/pause                       | 3.1               | da86e6ba6ca19 | 742kB  |
| registry.k8s.io/echoserver                  | 1.8               | 82e4c8a736a4f | 95.4MB |
| registry.k8s.io/kube-apiserver              | v1.31.1           | 6bab7719df100 | 94.2MB |
| registry.k8s.io/kube-scheduler              | v1.31.1           | 9aa1fad941575 | 67.4MB |
|---------------------------------------------|-------------------|---------------|--------|
functional_test.go:269: (dbg) Stderr: out/minikube-linux-amd64 -p functional-463208 image ls --format table --alsologtostderr:
I0913 18:37:38.618580   67720 out.go:345] Setting OutFile to fd 1 ...
I0913 18:37:38.618832   67720 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0913 18:37:38.618845   67720 out.go:358] Setting ErrFile to fd 2...
I0913 18:37:38.618852   67720 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0913 18:37:38.619031   67720 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
I0913 18:37:38.619575   67720 config.go:182] Loaded profile config "functional-463208": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0913 18:37:38.619664   67720 config.go:182] Loaded profile config "functional-463208": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0913 18:37:38.619984   67720 cli_runner.go:164] Run: docker container inspect functional-463208 --format={{.State.Status}}
I0913 18:37:38.638513   67720 ssh_runner.go:195] Run: systemctl --version
I0913 18:37:38.638564   67720 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" functional-463208
I0913 18:37:38.655163   67720 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32778 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/functional-463208/id_rsa Username:docker}
I0913 18:37:38.741492   67720 ssh_runner.go:195] Run: docker images --no-trunc --format "{{json .}}"
--- PASS: TestFunctional/parallel/ImageCommands/ImageListTable (0.19s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageListJson (0.19s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageListJson
=== PAUSE TestFunctional/parallel/ImageCommands/ImageListJson

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ImageCommands/ImageListJson
functional_test.go:261: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image ls --format json --alsologtostderr
functional_test.go:266: (dbg) Stdout: out/minikube-linux-amd64 -p functional-463208 image ls --format json --alsologtostderr:
[{"id":"82e4c8a736a4fcf22b5ef9f6a4ff6207064c7187d7694bf97bd561605a538410","repoDigests":[],"repoTags":["registry.k8s.io/echoserver:1.8"],"size":"95400000"},{"id":"60c005f310ff3ad6d131805170f07d2946095307063eaaa5eedcaf06a0a89561","repoDigests":[],"repoTags":["registry.k8s.io/kube-proxy:v1.31.1"],"size":"91500000"},{"id":"c7b4f26a7d93f4f1f276c51adb03ef0df54a82de89f254a9aec5c18bf0e45ee9","repoDigests":[],"repoTags":["docker.io/library/nginx:alpine"],"size":"43200000"},{"id":"39286ab8a5e14aeaf5fdd6e2fac76e0c8d31a0c07224f0ee5e6be502f12e93f3","repoDigests":[],"repoTags":["docker.io/library/nginx:latest"],"size":"188000000"},{"id":"07655ddf2eebe5d250f7a72c25f638b27126805d61779741b4e62e69ba080558","repoDigests":[],"repoTags":["docker.io/kubernetesui/dashboard:\u003cnone\u003e"],"size":"246000000"},{"id":"bef76f5f2e90fdd10bc17f1bc4c46a158ac5a69df01814720cb6a0f852b1d497","repoDigests":[],"repoTags":["docker.io/library/minikube-local-cache-test:functional-463208"],"size":"30"},{"id":"6bab7719df1001fdcc7e39f1decf
a1f73b7f3af2757a91c5bafa1aaea29d1aee","repoDigests":[],"repoTags":["registry.k8s.io/kube-apiserver:v1.31.1"],"size":"94200000"},{"id":"0184c1613d92931126feb4c548e5da11015513b9e4c104e7305ee8b53b50a9da","repoDigests":[],"repoTags":["registry.k8s.io/pause:3.3"],"size":"683000"},{"id":"da86e6ba6ca197bf6bc5e9d900febd906b133eaa4750e6bed647b0fbe50ed43e","repoDigests":[],"repoTags":["registry.k8s.io/pause:3.1"],"size":"742000"},{"id":"c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6","repoDigests":[],"repoTags":["registry.k8s.io/coredns/coredns:v1.11.3"],"size":"61800000"},{"id":"2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4","repoDigests":[],"repoTags":["registry.k8s.io/etcd:3.5.15-0"],"size":"148000000"},{"id":"873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136","repoDigests":[],"repoTags":["registry.k8s.io/pause:3.10"],"size":"736000"},{"id":"6e38f40d628db3002f5617342c8872c935de530d867d0f709a2fbda1a302a562","repoDigests":[],"repoTags":["gcr.io/k8s-minikube/storage-p
rovisioner:v5"],"size":"31500000"},{"id":"115053965e86b2df4d78af78d7951b8644839d20a03820c6df59a261103315f7","repoDigests":[],"repoTags":["docker.io/kubernetesui/metrics-scraper:\u003cnone\u003e"],"size":"43800000"},{"id":"56cc512116c8f894f11ce1995460aef1ee0972d48bc2a3bdb1faaac7c020289c","repoDigests":[],"repoTags":["gcr.io/k8s-minikube/busybox:1.28.4-glibc"],"size":"4400000"},{"id":"350b164e7ae1dcddeffadd65c76226c9b6dc5553f5179153fb0e36b78f2a5e06","repoDigests":[],"repoTags":["registry.k8s.io/pause:latest"],"size":"240000"},{"id":"9aa1fad941575eed91ab13d44f3e4cb5b1ff4e09cbbe954ea63002289416a13b","repoDigests":[],"repoTags":["registry.k8s.io/kube-scheduler:v1.31.1"],"size":"67400000"},{"id":"175ffd71cce3d90bae95904b55260db941b10007a4e5471a19f3135b30aa9cd1","repoDigests":[],"repoTags":["registry.k8s.io/kube-controller-manager:v1.31.1"],"size":"88400000"},{"id":"5107333e08a87b836d48ff7528b1e84b9c86781cc9f1748bbc1b8c42a870d933","repoDigests":[],"repoTags":["docker.io/library/mysql:5.7"],"size":"501000000"},{"id":
"9056ab77afb8e18e04303f11000a9d31b3f16b74c59475b899ae1b342d328d30","repoDigests":[],"repoTags":["docker.io/kicbase/echo-server:functional-463208"],"size":"4940000"}]
functional_test.go:269: (dbg) Stderr: out/minikube-linux-amd64 -p functional-463208 image ls --format json --alsologtostderr:
I0913 18:37:38.426760   67578 out.go:345] Setting OutFile to fd 1 ...
I0913 18:37:38.426907   67578 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0913 18:37:38.426919   67578 out.go:358] Setting ErrFile to fd 2...
I0913 18:37:38.426925   67578 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0913 18:37:38.427175   67578 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
I0913 18:37:38.427942   67578 config.go:182] Loaded profile config "functional-463208": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0913 18:37:38.428086   67578 config.go:182] Loaded profile config "functional-463208": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0913 18:37:38.428679   67578 cli_runner.go:164] Run: docker container inspect functional-463208 --format={{.State.Status}}
I0913 18:37:38.444605   67578 ssh_runner.go:195] Run: systemctl --version
I0913 18:37:38.444645   67578 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" functional-463208
I0913 18:37:38.462042   67578 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32778 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/functional-463208/id_rsa Username:docker}
I0913 18:37:38.549377   67578 ssh_runner.go:195] Run: docker images --no-trunc --format "{{json .}}"
--- PASS: TestFunctional/parallel/ImageCommands/ImageListJson (0.19s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageListYaml (0.2s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageListYaml
=== PAUSE TestFunctional/parallel/ImageCommands/ImageListYaml

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ImageCommands/ImageListYaml
functional_test.go:261: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image ls --format yaml --alsologtostderr
functional_test.go:266: (dbg) Stdout: out/minikube-linux-amd64 -p functional-463208 image ls --format yaml --alsologtostderr:
- id: 9056ab77afb8e18e04303f11000a9d31b3f16b74c59475b899ae1b342d328d30
repoDigests: []
repoTags:
- docker.io/kicbase/echo-server:functional-463208
size: "4940000"
- id: 115053965e86b2df4d78af78d7951b8644839d20a03820c6df59a261103315f7
repoDigests: []
repoTags:
- docker.io/kubernetesui/metrics-scraper:<none>
size: "43800000"
- id: 6e38f40d628db3002f5617342c8872c935de530d867d0f709a2fbda1a302a562
repoDigests: []
repoTags:
- gcr.io/k8s-minikube/storage-provisioner:v5
size: "31500000"
- id: 0184c1613d92931126feb4c548e5da11015513b9e4c104e7305ee8b53b50a9da
repoDigests: []
repoTags:
- registry.k8s.io/pause:3.3
size: "683000"
- id: 60c005f310ff3ad6d131805170f07d2946095307063eaaa5eedcaf06a0a89561
repoDigests: []
repoTags:
- registry.k8s.io/kube-proxy:v1.31.1
size: "91500000"
- id: 39286ab8a5e14aeaf5fdd6e2fac76e0c8d31a0c07224f0ee5e6be502f12e93f3
repoDigests: []
repoTags:
- docker.io/library/nginx:latest
size: "188000000"
- id: c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6
repoDigests: []
repoTags:
- registry.k8s.io/coredns/coredns:v1.11.3
size: "61800000"
- id: 873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136
repoDigests: []
repoTags:
- registry.k8s.io/pause:3.10
size: "736000"
- id: 56cc512116c8f894f11ce1995460aef1ee0972d48bc2a3bdb1faaac7c020289c
repoDigests: []
repoTags:
- gcr.io/k8s-minikube/busybox:1.28.4-glibc
size: "4400000"
- id: 350b164e7ae1dcddeffadd65c76226c9b6dc5553f5179153fb0e36b78f2a5e06
repoDigests: []
repoTags:
- registry.k8s.io/pause:latest
size: "240000"
- id: 6bab7719df1001fdcc7e39f1decfa1f73b7f3af2757a91c5bafa1aaea29d1aee
repoDigests: []
repoTags:
- registry.k8s.io/kube-apiserver:v1.31.1
size: "94200000"
- id: 175ffd71cce3d90bae95904b55260db941b10007a4e5471a19f3135b30aa9cd1
repoDigests: []
repoTags:
- registry.k8s.io/kube-controller-manager:v1.31.1
size: "88400000"
- id: 2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4
repoDigests: []
repoTags:
- registry.k8s.io/etcd:3.5.15-0
size: "148000000"
- id: da86e6ba6ca197bf6bc5e9d900febd906b133eaa4750e6bed647b0fbe50ed43e
repoDigests: []
repoTags:
- registry.k8s.io/pause:3.1
size: "742000"
- id: 82e4c8a736a4fcf22b5ef9f6a4ff6207064c7187d7694bf97bd561605a538410
repoDigests: []
repoTags:
- registry.k8s.io/echoserver:1.8
size: "95400000"
- id: bef76f5f2e90fdd10bc17f1bc4c46a158ac5a69df01814720cb6a0f852b1d497
repoDigests: []
repoTags:
- docker.io/library/minikube-local-cache-test:functional-463208
size: "30"
- id: c7b4f26a7d93f4f1f276c51adb03ef0df54a82de89f254a9aec5c18bf0e45ee9
repoDigests: []
repoTags:
- docker.io/library/nginx:alpine
size: "43200000"
- id: 5107333e08a87b836d48ff7528b1e84b9c86781cc9f1748bbc1b8c42a870d933
repoDigests: []
repoTags:
- docker.io/library/mysql:5.7
size: "501000000"
- id: 9aa1fad941575eed91ab13d44f3e4cb5b1ff4e09cbbe954ea63002289416a13b
repoDigests: []
repoTags:
- registry.k8s.io/kube-scheduler:v1.31.1
size: "67400000"
- id: 07655ddf2eebe5d250f7a72c25f638b27126805d61779741b4e62e69ba080558
repoDigests: []
repoTags:
- docker.io/kubernetesui/dashboard:<none>
size: "246000000"

                                                
                                                
functional_test.go:269: (dbg) Stderr: out/minikube-linux-amd64 -p functional-463208 image ls --format yaml --alsologtostderr:
I0913 18:37:38.228399   67481 out.go:345] Setting OutFile to fd 1 ...
I0913 18:37:38.228515   67481 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0913 18:37:38.228524   67481 out.go:358] Setting ErrFile to fd 2...
I0913 18:37:38.228528   67481 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0913 18:37:38.228689   67481 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
I0913 18:37:38.230576   67481 config.go:182] Loaded profile config "functional-463208": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0913 18:37:38.230879   67481 config.go:182] Loaded profile config "functional-463208": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0913 18:37:38.231525   67481 cli_runner.go:164] Run: docker container inspect functional-463208 --format={{.State.Status}}
I0913 18:37:38.250273   67481 ssh_runner.go:195] Run: systemctl --version
I0913 18:37:38.250317   67481 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" functional-463208
I0913 18:37:38.268342   67481 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32778 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/functional-463208/id_rsa Username:docker}
I0913 18:37:38.353868   67481 ssh_runner.go:195] Run: docker images --no-trunc --format "{{json .}}"
--- PASS: TestFunctional/parallel/ImageCommands/ImageListYaml (0.20s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageBuild (2.34s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageBuild
=== PAUSE TestFunctional/parallel/ImageCommands/ImageBuild

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/ImageCommands/ImageBuild
functional_test.go:308: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 ssh pgrep buildkitd
functional_test.go:308: (dbg) Non-zero exit: out/minikube-linux-amd64 -p functional-463208 ssh pgrep buildkitd: exit status 1 (229.100572ms)

                                                
                                                
** stderr ** 
	ssh: Process exited with status 1

                                                
                                                
** /stderr **
functional_test.go:315: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image build -t localhost/my-image:functional-463208 testdata/build --alsologtostderr
functional_test.go:315: (dbg) Done: out/minikube-linux-amd64 -p functional-463208 image build -t localhost/my-image:functional-463208 testdata/build --alsologtostderr: (1.923458923s)
functional_test.go:323: (dbg) Stderr: out/minikube-linux-amd64 -p functional-463208 image build -t localhost/my-image:functional-463208 testdata/build --alsologtostderr:
I0913 18:37:38.651892   67733 out.go:345] Setting OutFile to fd 1 ...
I0913 18:37:38.652059   67733 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0913 18:37:38.652073   67733 out.go:358] Setting ErrFile to fd 2...
I0913 18:37:38.652079   67733 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0913 18:37:38.652324   67733 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
I0913 18:37:38.652966   67733 config.go:182] Loaded profile config "functional-463208": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0913 18:37:38.653464   67733 config.go:182] Loaded profile config "functional-463208": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0913 18:37:38.653919   67733 cli_runner.go:164] Run: docker container inspect functional-463208 --format={{.State.Status}}
I0913 18:37:38.669999   67733 ssh_runner.go:195] Run: systemctl --version
I0913 18:37:38.670034   67733 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" functional-463208
I0913 18:37:38.685762   67733 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32778 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/functional-463208/id_rsa Username:docker}
I0913 18:37:38.773830   67733 build_images.go:161] Building image from path: /tmp/build.638566503.tar
I0913 18:37:38.773876   67733 ssh_runner.go:195] Run: sudo mkdir -p /var/lib/minikube/build
I0913 18:37:38.781644   67733 ssh_runner.go:195] Run: stat -c "%s %y" /var/lib/minikube/build/build.638566503.tar
I0913 18:37:38.784502   67733 ssh_runner.go:352] existence check for /var/lib/minikube/build/build.638566503.tar: stat -c "%s %y" /var/lib/minikube/build/build.638566503.tar: Process exited with status 1
stdout:

                                                
                                                
stderr:
stat: cannot statx '/var/lib/minikube/build/build.638566503.tar': No such file or directory
I0913 18:37:38.784535   67733 ssh_runner.go:362] scp /tmp/build.638566503.tar --> /var/lib/minikube/build/build.638566503.tar (3072 bytes)
I0913 18:37:38.805036   67733 ssh_runner.go:195] Run: sudo mkdir -p /var/lib/minikube/build/build.638566503
I0913 18:37:38.812444   67733 ssh_runner.go:195] Run: sudo tar -C /var/lib/minikube/build/build.638566503 -xf /var/lib/minikube/build/build.638566503.tar
I0913 18:37:38.819826   67733 docker.go:360] Building image: /var/lib/minikube/build/build.638566503
I0913 18:37:38.819882   67733 ssh_runner.go:195] Run: docker build -t localhost/my-image:functional-463208 /var/lib/minikube/build/build.638566503
#0 building with "default" instance using docker driver

                                                
                                                
#1 [internal] load build definition from Dockerfile
#1 transferring dockerfile: 97B done
#1 DONE 0.0s

                                                
                                                
#2 [internal] load metadata for gcr.io/k8s-minikube/busybox:latest
#2 DONE 1.0s

                                                
                                                
#3 [internal] load .dockerignore
#3 transferring context: 2B done
#3 DONE 0.0s

                                                
                                                
#4 [internal] load build context
#4 transferring context: 62B done
#4 DONE 0.0s

                                                
                                                
#5 [1/3] FROM gcr.io/k8s-minikube/busybox:latest@sha256:ca5ae90100d50772da31f3b5016209e25ad61972404e2ccd83d44f10dee7e79b
#5 resolve gcr.io/k8s-minikube/busybox:latest@sha256:ca5ae90100d50772da31f3b5016209e25ad61972404e2ccd83d44f10dee7e79b 0.0s done
#5 sha256:ca5ae90100d50772da31f3b5016209e25ad61972404e2ccd83d44f10dee7e79b 770B / 770B done
#5 sha256:62ffc2ed7554e4c6d360bce40bbcf196573dd27c4ce080641a2c59867e732dee 527B / 527B done
#5 sha256:beae173ccac6ad749f76713cf4440fe3d21d1043fe616dfbe30775815d1d0f6a 1.46kB / 1.46kB done
#5 sha256:5cc84ad355aaa64f46ea9c7bbcc319a9d808ab15088a27209c9e70ef86e5a2aa 0B / 772.79kB 0.1s
#5 extracting sha256:5cc84ad355aaa64f46ea9c7bbcc319a9d808ab15088a27209c9e70ef86e5a2aa
#5 sha256:5cc84ad355aaa64f46ea9c7bbcc319a9d808ab15088a27209c9e70ef86e5a2aa 772.79kB / 772.79kB 0.2s done
#5 extracting sha256:5cc84ad355aaa64f46ea9c7bbcc319a9d808ab15088a27209c9e70ef86e5a2aa done
#5 DONE 0.3s

                                                
                                                
#6 [2/3] RUN true
#6 DONE 0.2s

                                                
                                                
#7 [3/3] ADD content.txt /
#7 DONE 0.0s

                                                
                                                
#8 exporting to image
#8 exporting layers 0.0s done
#8 writing image sha256:400d8a700002e47a480520b62980ddbd45a6e8dc18e8e73b3d0868cff1279a1e done
#8 naming to localhost/my-image:functional-463208 done
#8 DONE 0.0s
I0913 18:37:40.511617   67733 ssh_runner.go:235] Completed: docker build -t localhost/my-image:functional-463208 /var/lib/minikube/build/build.638566503: (1.691713629s)
I0913 18:37:40.511668   67733 ssh_runner.go:195] Run: sudo rm -rf /var/lib/minikube/build/build.638566503
I0913 18:37:40.519775   67733 ssh_runner.go:195] Run: sudo rm -f /var/lib/minikube/build/build.638566503.tar
I0913 18:37:40.526956   67733 build_images.go:217] Built localhost/my-image:functional-463208 from /tmp/build.638566503.tar
I0913 18:37:40.526982   67733 build_images.go:133] succeeded building to: functional-463208
I0913 18:37:40.526988   67733 build_images.go:134] failed building to: 
functional_test.go:451: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image ls
--- PASS: TestFunctional/parallel/ImageCommands/ImageBuild (2.34s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/Setup (0.96s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/Setup
functional_test.go:342: (dbg) Run:  docker pull kicbase/echo-server:1.0
functional_test.go:347: (dbg) Run:  docker tag kicbase/echo-server:1.0 kicbase/echo-server:functional-463208
--- PASS: TestFunctional/parallel/ImageCommands/Setup (0.96s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageLoadDaemon (1.07s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageLoadDaemon
functional_test.go:355: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image load --daemon kicbase/echo-server:functional-463208 --alsologtostderr
functional_test.go:451: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image ls
--- PASS: TestFunctional/parallel/ImageCommands/ImageLoadDaemon (1.07s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageReloadDaemon (0.91s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageReloadDaemon
functional_test.go:365: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image load --daemon kicbase/echo-server:functional-463208 --alsologtostderr
functional_test.go:451: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image ls
--- PASS: TestFunctional/parallel/ImageCommands/ImageReloadDaemon (0.91s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageTagAndLoadDaemon (0.96s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageTagAndLoadDaemon
functional_test.go:235: (dbg) Run:  docker pull kicbase/echo-server:latest
functional_test.go:240: (dbg) Run:  docker tag kicbase/echo-server:latest kicbase/echo-server:functional-463208
functional_test.go:245: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image load --daemon kicbase/echo-server:functional-463208 --alsologtostderr
functional_test.go:451: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image ls
--- PASS: TestFunctional/parallel/ImageCommands/ImageTagAndLoadDaemon (0.96s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageSaveToFile (0.28s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageSaveToFile
functional_test.go:380: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image save kicbase/echo-server:functional-463208 /home/jenkins/workspace/Docker_Linux_integration/echo-server-save.tar --alsologtostderr
--- PASS: TestFunctional/parallel/ImageCommands/ImageSaveToFile (0.28s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageRemove (0.45s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageRemove
functional_test.go:392: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image rm kicbase/echo-server:functional-463208 --alsologtostderr
functional_test.go:451: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image ls
--- PASS: TestFunctional/parallel/ImageCommands/ImageRemove (0.45s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageLoadFromFile (0.53s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageLoadFromFile
functional_test.go:409: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image load /home/jenkins/workspace/Docker_Linux_integration/echo-server-save.tar --alsologtostderr
functional_test.go:451: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image ls
--- PASS: TestFunctional/parallel/ImageCommands/ImageLoadFromFile (0.53s)

                                                
                                    
x
+
TestFunctional/parallel/ImageCommands/ImageSaveDaemon (0.33s)

                                                
                                                
=== RUN   TestFunctional/parallel/ImageCommands/ImageSaveDaemon
functional_test.go:419: (dbg) Run:  docker rmi kicbase/echo-server:functional-463208
functional_test.go:424: (dbg) Run:  out/minikube-linux-amd64 -p functional-463208 image save --daemon kicbase/echo-server:functional-463208 --alsologtostderr
functional_test.go:432: (dbg) Run:  docker image inspect kicbase/echo-server:functional-463208
--- PASS: TestFunctional/parallel/ImageCommands/ImageSaveDaemon (0.33s)

                                                
                                    
x
+
TestFunctional/delete_echo-server_images (0.03s)

                                                
                                                
=== RUN   TestFunctional/delete_echo-server_images
functional_test.go:190: (dbg) Run:  docker rmi -f kicbase/echo-server:1.0
functional_test.go:190: (dbg) Run:  docker rmi -f kicbase/echo-server:functional-463208
--- PASS: TestFunctional/delete_echo-server_images (0.03s)

                                                
                                    
x
+
TestFunctional/delete_my-image_image (0.01s)

                                                
                                                
=== RUN   TestFunctional/delete_my-image_image
functional_test.go:198: (dbg) Run:  docker rmi -f localhost/my-image:functional-463208
--- PASS: TestFunctional/delete_my-image_image (0.01s)

                                                
                                    
x
+
TestFunctional/delete_minikube_cached_images (0.01s)

                                                
                                                
=== RUN   TestFunctional/delete_minikube_cached_images
functional_test.go:206: (dbg) Run:  docker rmi -f minikube-local-cache-test:functional-463208
--- PASS: TestFunctional/delete_minikube_cached_images (0.01s)

                                                
                                    
x
+
TestMultiControlPlane/serial/StartCluster (100.37s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/StartCluster
ha_test.go:101: (dbg) Run:  out/minikube-linux-amd64 start -p ha-243920 --wait=true --memory=2200 --ha -v=7 --alsologtostderr --driver=docker  --container-runtime=docker
E0913 18:39:03.999016   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:39:04.005828   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:39:04.017170   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:39:04.038484   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:39:04.079792   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:39:04.161172   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:39:04.322517   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:39:04.644134   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:39:05.285774   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:39:06.567088   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:39:09.129360   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:39:14.250884   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:39:24.493062   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
ha_test.go:101: (dbg) Done: out/minikube-linux-amd64 start -p ha-243920 --wait=true --memory=2200 --ha -v=7 --alsologtostderr --driver=docker  --container-runtime=docker: (1m39.747256868s)
ha_test.go:107: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 status -v=7 --alsologtostderr
--- PASS: TestMultiControlPlane/serial/StartCluster (100.37s)

                                                
                                    
x
+
TestMultiControlPlane/serial/DeployApp (39.24s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/DeployApp
ha_test.go:128: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- apply -f ./testdata/ha/ha-pod-dns-test.yaml
ha_test.go:133: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- rollout status deployment/busybox
ha_test.go:133: (dbg) Done: out/minikube-linux-amd64 kubectl -p ha-243920 -- rollout status deployment/busybox: (2.651571934s)
ha_test.go:140: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:149: expected 3 Pod IPs but got 2 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.4 10.244.1.2'\n\n-- /stdout --"
ha_test.go:140: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:149: expected 3 Pod IPs but got 2 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.4 10.244.1.2'\n\n-- /stdout --"
ha_test.go:140: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:149: expected 3 Pod IPs but got 2 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.4 10.244.1.2'\n\n-- /stdout --"
ha_test.go:140: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:149: expected 3 Pod IPs but got 2 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.4 10.244.1.2'\n\n-- /stdout --"
ha_test.go:140: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:149: expected 3 Pod IPs but got 2 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.4 10.244.1.2'\n\n-- /stdout --"
E0913 18:39:44.974412   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
ha_test.go:140: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:149: expected 3 Pod IPs but got 2 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.4 10.244.1.2'\n\n-- /stdout --"
ha_test.go:140: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:149: expected 3 Pod IPs but got 2 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.4 10.244.1.2'\n\n-- /stdout --"
ha_test.go:140: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:163: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- get pods -o jsonpath='{.items[*].metadata.name}'
ha_test.go:171: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-59nr6 -- nslookup kubernetes.io
ha_test.go:171: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-6cfgd -- nslookup kubernetes.io
ha_test.go:171: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-bzwg4 -- nslookup kubernetes.io
ha_test.go:181: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-59nr6 -- nslookup kubernetes.default
ha_test.go:181: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-6cfgd -- nslookup kubernetes.default
ha_test.go:181: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-bzwg4 -- nslookup kubernetes.default
ha_test.go:189: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-59nr6 -- nslookup kubernetes.default.svc.cluster.local
ha_test.go:189: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-6cfgd -- nslookup kubernetes.default.svc.cluster.local
ha_test.go:189: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-bzwg4 -- nslookup kubernetes.default.svc.cluster.local
--- PASS: TestMultiControlPlane/serial/DeployApp (39.24s)

                                                
                                    
x
+
TestMultiControlPlane/serial/PingHostFromPods (1.02s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/PingHostFromPods
ha_test.go:199: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- get pods -o jsonpath='{.items[*].metadata.name}'
ha_test.go:207: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-59nr6 -- sh -c "nslookup host.minikube.internal | awk 'NR==5' | cut -d' ' -f3"
ha_test.go:218: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-59nr6 -- sh -c "ping -c 1 192.168.49.1"
ha_test.go:207: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-6cfgd -- sh -c "nslookup host.minikube.internal | awk 'NR==5' | cut -d' ' -f3"
ha_test.go:218: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-6cfgd -- sh -c "ping -c 1 192.168.49.1"
ha_test.go:207: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-bzwg4 -- sh -c "nslookup host.minikube.internal | awk 'NR==5' | cut -d' ' -f3"
ha_test.go:218: (dbg) Run:  out/minikube-linux-amd64 kubectl -p ha-243920 -- exec busybox-7dff88458-bzwg4 -- sh -c "ping -c 1 192.168.49.1"
--- PASS: TestMultiControlPlane/serial/PingHostFromPods (1.02s)

                                                
                                    
x
+
TestMultiControlPlane/serial/AddWorkerNode (20.56s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/AddWorkerNode
ha_test.go:228: (dbg) Run:  out/minikube-linux-amd64 node add -p ha-243920 -v=7 --alsologtostderr
E0913 18:40:25.936274   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
ha_test.go:228: (dbg) Done: out/minikube-linux-amd64 node add -p ha-243920 -v=7 --alsologtostderr: (19.787486803s)
ha_test.go:234: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 status -v=7 --alsologtostderr
--- PASS: TestMultiControlPlane/serial/AddWorkerNode (20.56s)

                                                
                                    
x
+
TestMultiControlPlane/serial/NodeLabels (0.06s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/NodeLabels
ha_test.go:255: (dbg) Run:  kubectl --context ha-243920 get nodes -o "jsonpath=[{range .items[*]}{.metadata.labels},{end}]"
--- PASS: TestMultiControlPlane/serial/NodeLabels (0.06s)

                                                
                                    
x
+
TestMultiControlPlane/serial/HAppyAfterClusterStart (0.58s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/HAppyAfterClusterStart
ha_test.go:281: (dbg) Run:  out/minikube-linux-amd64 profile list --output json
--- PASS: TestMultiControlPlane/serial/HAppyAfterClusterStart (0.58s)

                                                
                                    
x
+
TestMultiControlPlane/serial/CopyFile (14.93s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/CopyFile
ha_test.go:326: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 status --output json -v=7 --alsologtostderr
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp testdata/cp-test.txt ha-243920:/home/docker/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920:/home/docker/cp-test.txt /tmp/TestMultiControlPlaneserialCopyFile3934711309/001/cp-test_ha-243920.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920:/home/docker/cp-test.txt ha-243920-m02:/home/docker/cp-test_ha-243920_ha-243920-m02.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m02 "sudo cat /home/docker/cp-test_ha-243920_ha-243920-m02.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920:/home/docker/cp-test.txt ha-243920-m03:/home/docker/cp-test_ha-243920_ha-243920-m03.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m03 "sudo cat /home/docker/cp-test_ha-243920_ha-243920-m03.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920:/home/docker/cp-test.txt ha-243920-m04:/home/docker/cp-test_ha-243920_ha-243920-m04.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m04 "sudo cat /home/docker/cp-test_ha-243920_ha-243920-m04.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp testdata/cp-test.txt ha-243920-m02:/home/docker/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m02 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920-m02:/home/docker/cp-test.txt /tmp/TestMultiControlPlaneserialCopyFile3934711309/001/cp-test_ha-243920-m02.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m02 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920-m02:/home/docker/cp-test.txt ha-243920:/home/docker/cp-test_ha-243920-m02_ha-243920.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m02 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920 "sudo cat /home/docker/cp-test_ha-243920-m02_ha-243920.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920-m02:/home/docker/cp-test.txt ha-243920-m03:/home/docker/cp-test_ha-243920-m02_ha-243920-m03.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m02 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m03 "sudo cat /home/docker/cp-test_ha-243920-m02_ha-243920-m03.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920-m02:/home/docker/cp-test.txt ha-243920-m04:/home/docker/cp-test_ha-243920-m02_ha-243920-m04.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m02 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m04 "sudo cat /home/docker/cp-test_ha-243920-m02_ha-243920-m04.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp testdata/cp-test.txt ha-243920-m03:/home/docker/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m03 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920-m03:/home/docker/cp-test.txt /tmp/TestMultiControlPlaneserialCopyFile3934711309/001/cp-test_ha-243920-m03.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m03 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920-m03:/home/docker/cp-test.txt ha-243920:/home/docker/cp-test_ha-243920-m03_ha-243920.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m03 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920 "sudo cat /home/docker/cp-test_ha-243920-m03_ha-243920.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920-m03:/home/docker/cp-test.txt ha-243920-m02:/home/docker/cp-test_ha-243920-m03_ha-243920-m02.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m03 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m02 "sudo cat /home/docker/cp-test_ha-243920-m03_ha-243920-m02.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920-m03:/home/docker/cp-test.txt ha-243920-m04:/home/docker/cp-test_ha-243920-m03_ha-243920-m04.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m03 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m04 "sudo cat /home/docker/cp-test_ha-243920-m03_ha-243920-m04.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp testdata/cp-test.txt ha-243920-m04:/home/docker/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m04 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920-m04:/home/docker/cp-test.txt /tmp/TestMultiControlPlaneserialCopyFile3934711309/001/cp-test_ha-243920-m04.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m04 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920-m04:/home/docker/cp-test.txt ha-243920:/home/docker/cp-test_ha-243920-m04_ha-243920.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m04 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920 "sudo cat /home/docker/cp-test_ha-243920-m04_ha-243920.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920-m04:/home/docker/cp-test.txt ha-243920-m02:/home/docker/cp-test_ha-243920-m04_ha-243920-m02.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m04 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m02 "sudo cat /home/docker/cp-test_ha-243920-m04_ha-243920-m02.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 cp ha-243920-m04:/home/docker/cp-test.txt ha-243920-m03:/home/docker/cp-test_ha-243920-m04_ha-243920-m03.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m04 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 ssh -n ha-243920-m03 "sudo cat /home/docker/cp-test_ha-243920-m04_ha-243920-m03.txt"
--- PASS: TestMultiControlPlane/serial/CopyFile (14.93s)

                                                
                                    
x
+
TestMultiControlPlane/serial/StopSecondaryNode (11.23s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/StopSecondaryNode
ha_test.go:363: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 node stop m02 -v=7 --alsologtostderr
ha_test.go:363: (dbg) Done: out/minikube-linux-amd64 -p ha-243920 node stop m02 -v=7 --alsologtostderr: (10.613209234s)
ha_test.go:369: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 status -v=7 --alsologtostderr
ha_test.go:369: (dbg) Non-zero exit: out/minikube-linux-amd64 -p ha-243920 status -v=7 --alsologtostderr: exit status 7 (611.302068ms)

                                                
                                                
-- stdout --
	ha-243920
	type: Control Plane
	host: Running
	kubelet: Running
	apiserver: Running
	kubeconfig: Configured
	
	ha-243920-m02
	type: Control Plane
	host: Stopped
	kubelet: Stopped
	apiserver: Stopped
	kubeconfig: Stopped
	
	ha-243920-m03
	type: Control Plane
	host: Running
	kubelet: Running
	apiserver: Running
	kubeconfig: Configured
	
	ha-243920-m04
	type: Worker
	host: Running
	kubelet: Running
	

                                                
                                                
-- /stdout --
** stderr ** 
	I0913 18:40:56.893412   95714 out.go:345] Setting OutFile to fd 1 ...
	I0913 18:40:56.893519   95714 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:40:56.893527   95714 out.go:358] Setting ErrFile to fd 2...
	I0913 18:40:56.893536   95714 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:40:56.893745   95714 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
	I0913 18:40:56.893919   95714 out.go:352] Setting JSON to false
	I0913 18:40:56.893945   95714 mustload.go:65] Loading cluster: ha-243920
	I0913 18:40:56.893996   95714 notify.go:220] Checking for updates...
	I0913 18:40:56.894417   95714 config.go:182] Loaded profile config "ha-243920": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
	I0913 18:40:56.894436   95714 status.go:255] checking status of ha-243920 ...
	I0913 18:40:56.894922   95714 cli_runner.go:164] Run: docker container inspect ha-243920 --format={{.State.Status}}
	I0913 18:40:56.912553   95714 status.go:330] ha-243920 host status = "Running" (err=<nil>)
	I0913 18:40:56.912579   95714 host.go:66] Checking if "ha-243920" exists ...
	I0913 18:40:56.912892   95714 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" ha-243920
	I0913 18:40:56.929185   95714 host.go:66] Checking if "ha-243920" exists ...
	I0913 18:40:56.929479   95714 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0913 18:40:56.929520   95714 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" ha-243920
	I0913 18:40:56.944737   95714 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32783 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/ha-243920/id_rsa Username:docker}
	I0913 18:40:57.030695   95714 ssh_runner.go:195] Run: systemctl --version
	I0913 18:40:57.034419   95714 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0913 18:40:57.044094   95714 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0913 18:40:57.093495   95714 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:4 ContainersRunning:3 ContainersPaused:0 ContainersStopped:1 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:54 OomKillDisable:true NGoroutines:72 SystemTime:2024-09-13 18:40:57.083318264 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
	I0913 18:40:57.094125   95714 kubeconfig.go:125] found "ha-243920" server: "https://192.168.49.254:8443"
	I0913 18:40:57.094155   95714 api_server.go:166] Checking apiserver status ...
	I0913 18:40:57.094201   95714 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0913 18:40:57.104905   95714 ssh_runner.go:195] Run: sudo egrep ^[0-9]+:freezer: /proc/2461/cgroup
	I0913 18:40:57.114529   95714 api_server.go:182] apiserver freezer: "7:freezer:/docker/285fe45474fca3ca69861fdb8067dc70a2289075156622116d5d33e044135706/kubepods/burstable/pod491ca927081ec40ebbbd7e613baf44bd/2cd3dbd7f091feeea9323579240fcc92d9aeb61de314bfe5a66b9d5568c38af8"
	I0913 18:40:57.114595   95714 ssh_runner.go:195] Run: sudo cat /sys/fs/cgroup/freezer/docker/285fe45474fca3ca69861fdb8067dc70a2289075156622116d5d33e044135706/kubepods/burstable/pod491ca927081ec40ebbbd7e613baf44bd/2cd3dbd7f091feeea9323579240fcc92d9aeb61de314bfe5a66b9d5568c38af8/freezer.state
	I0913 18:40:57.122214   95714 api_server.go:204] freezer state: "THAWED"
	I0913 18:40:57.122239   95714 api_server.go:253] Checking apiserver healthz at https://192.168.49.254:8443/healthz ...
	I0913 18:40:57.125776   95714 api_server.go:279] https://192.168.49.254:8443/healthz returned 200:
	ok
	I0913 18:40:57.125794   95714 status.go:422] ha-243920 apiserver status = Running (err=<nil>)
	I0913 18:40:57.125804   95714 status.go:257] ha-243920 status: &{Name:ha-243920 Host:Running Kubelet:Running APIServer:Running Kubeconfig:Configured Worker:false TimeToStop: DockerEnv: PodManEnv:}
	I0913 18:40:57.125828   95714 status.go:255] checking status of ha-243920-m02 ...
	I0913 18:40:57.126078   95714 cli_runner.go:164] Run: docker container inspect ha-243920-m02 --format={{.State.Status}}
	I0913 18:40:57.143264   95714 status.go:330] ha-243920-m02 host status = "Stopped" (err=<nil>)
	I0913 18:40:57.143284   95714 status.go:343] host is not running, skipping remaining checks
	I0913 18:40:57.143291   95714 status.go:257] ha-243920-m02 status: &{Name:ha-243920-m02 Host:Stopped Kubelet:Stopped APIServer:Stopped Kubeconfig:Stopped Worker:false TimeToStop: DockerEnv: PodManEnv:}
	I0913 18:40:57.143314   95714 status.go:255] checking status of ha-243920-m03 ...
	I0913 18:40:57.143627   95714 cli_runner.go:164] Run: docker container inspect ha-243920-m03 --format={{.State.Status}}
	I0913 18:40:57.159597   95714 status.go:330] ha-243920-m03 host status = "Running" (err=<nil>)
	I0913 18:40:57.159618   95714 host.go:66] Checking if "ha-243920-m03" exists ...
	I0913 18:40:57.159841   95714 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" ha-243920-m03
	I0913 18:40:57.175766   95714 host.go:66] Checking if "ha-243920-m03" exists ...
	I0913 18:40:57.176010   95714 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0913 18:40:57.176042   95714 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" ha-243920-m03
	I0913 18:40:57.191623   95714 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32793 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/ha-243920-m03/id_rsa Username:docker}
	I0913 18:40:57.278320   95714 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0913 18:40:57.288524   95714 kubeconfig.go:125] found "ha-243920" server: "https://192.168.49.254:8443"
	I0913 18:40:57.288548   95714 api_server.go:166] Checking apiserver status ...
	I0913 18:40:57.288579   95714 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0913 18:40:57.298407   95714 ssh_runner.go:195] Run: sudo egrep ^[0-9]+:freezer: /proc/2249/cgroup
	I0913 18:40:57.306588   95714 api_server.go:182] apiserver freezer: "7:freezer:/docker/48b027596c827bf3bc683ac658edde23821471a77b7275f273f5cda0040f4c7f/kubepods/burstable/pode0710e7a6a11e56faaaf39769deac076/27e14d13e213c6bae424a0143fc96a5c56aefa7821661b986f5535e98ade9638"
	I0913 18:40:57.306649   95714 ssh_runner.go:195] Run: sudo cat /sys/fs/cgroup/freezer/docker/48b027596c827bf3bc683ac658edde23821471a77b7275f273f5cda0040f4c7f/kubepods/burstable/pode0710e7a6a11e56faaaf39769deac076/27e14d13e213c6bae424a0143fc96a5c56aefa7821661b986f5535e98ade9638/freezer.state
	I0913 18:40:57.314045   95714 api_server.go:204] freezer state: "THAWED"
	I0913 18:40:57.314068   95714 api_server.go:253] Checking apiserver healthz at https://192.168.49.254:8443/healthz ...
	I0913 18:40:57.317575   95714 api_server.go:279] https://192.168.49.254:8443/healthz returned 200:
	ok
	I0913 18:40:57.317602   95714 status.go:422] ha-243920-m03 apiserver status = Running (err=<nil>)
	I0913 18:40:57.317612   95714 status.go:257] ha-243920-m03 status: &{Name:ha-243920-m03 Host:Running Kubelet:Running APIServer:Running Kubeconfig:Configured Worker:false TimeToStop: DockerEnv: PodManEnv:}
	I0913 18:40:57.317636   95714 status.go:255] checking status of ha-243920-m04 ...
	I0913 18:40:57.317921   95714 cli_runner.go:164] Run: docker container inspect ha-243920-m04 --format={{.State.Status}}
	I0913 18:40:57.335154   95714 status.go:330] ha-243920-m04 host status = "Running" (err=<nil>)
	I0913 18:40:57.335173   95714 host.go:66] Checking if "ha-243920-m04" exists ...
	I0913 18:40:57.335399   95714 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" ha-243920-m04
	I0913 18:40:57.352084   95714 host.go:66] Checking if "ha-243920-m04" exists ...
	I0913 18:40:57.352320   95714 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0913 18:40:57.352364   95714 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" ha-243920-m04
	I0913 18:40:57.369198   95714 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32798 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/ha-243920-m04/id_rsa Username:docker}
	I0913 18:40:57.453947   95714 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0913 18:40:57.463630   95714 status.go:257] ha-243920-m04 status: &{Name:ha-243920-m04 Host:Running Kubelet:Running APIServer:Irrelevant Kubeconfig:Irrelevant Worker:true TimeToStop: DockerEnv: PodManEnv:}

                                                
                                                
** /stderr **
--- PASS: TestMultiControlPlane/serial/StopSecondaryNode (11.23s)

                                                
                                    
x
+
TestMultiControlPlane/serial/DegradedAfterControlPlaneNodeStop (0.46s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/DegradedAfterControlPlaneNodeStop
ha_test.go:390: (dbg) Run:  out/minikube-linux-amd64 profile list --output json
--- PASS: TestMultiControlPlane/serial/DegradedAfterControlPlaneNodeStop (0.46s)

                                                
                                    
x
+
TestMultiControlPlane/serial/RestartSecondaryNode (18.29s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/RestartSecondaryNode
ha_test.go:420: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 node start m02 -v=7 --alsologtostderr
ha_test.go:420: (dbg) Done: out/minikube-linux-amd64 -p ha-243920 node start m02 -v=7 --alsologtostderr: (17.315357249s)
ha_test.go:428: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 status -v=7 --alsologtostderr
ha_test.go:448: (dbg) Run:  kubectl get nodes
--- PASS: TestMultiControlPlane/serial/RestartSecondaryNode (18.29s)

                                                
                                    
x
+
TestMultiControlPlane/serial/HAppyAfterSecondaryNodeRestart (2.85s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/HAppyAfterSecondaryNodeRestart
ha_test.go:281: (dbg) Run:  out/minikube-linux-amd64 profile list --output json
ha_test.go:281: (dbg) Done: out/minikube-linux-amd64 profile list --output json: (2.850535446s)
--- PASS: TestMultiControlPlane/serial/HAppyAfterSecondaryNodeRestart (2.85s)

                                                
                                    
x
+
TestMultiControlPlane/serial/RestartClusterKeepsNodes (253.71s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/RestartClusterKeepsNodes
ha_test.go:456: (dbg) Run:  out/minikube-linux-amd64 node list -p ha-243920 -v=7 --alsologtostderr
ha_test.go:462: (dbg) Run:  out/minikube-linux-amd64 stop -p ha-243920 -v=7 --alsologtostderr
E0913 18:41:47.859230   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
ha_test.go:462: (dbg) Done: out/minikube-linux-amd64 stop -p ha-243920 -v=7 --alsologtostderr: (33.608587612s)
ha_test.go:467: (dbg) Run:  out/minikube-linux-amd64 start -p ha-243920 --wait=true -v=7 --alsologtostderr
E0913 18:42:11.538852   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:42:11.545192   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:42:11.556541   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:42:11.577887   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:42:11.619240   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:42:11.700618   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:42:11.862076   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:42:12.183760   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:42:12.825840   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:42:14.107557   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:42:16.669153   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:42:21.791261   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:42:32.033316   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:42:52.514887   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:43:33.476321   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:44:03.998953   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:44:31.701106   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 18:44:55.397844   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
ha_test.go:467: (dbg) Done: out/minikube-linux-amd64 start -p ha-243920 --wait=true -v=7 --alsologtostderr: (3m40.007792152s)
ha_test.go:472: (dbg) Run:  out/minikube-linux-amd64 node list -p ha-243920
--- PASS: TestMultiControlPlane/serial/RestartClusterKeepsNodes (253.71s)

                                                
                                    
x
+
TestMultiControlPlane/serial/DeleteSecondaryNode (9.11s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/DeleteSecondaryNode
ha_test.go:487: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 node delete m03 -v=7 --alsologtostderr
ha_test.go:487: (dbg) Done: out/minikube-linux-amd64 -p ha-243920 node delete m03 -v=7 --alsologtostderr: (8.376160218s)
ha_test.go:493: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 status -v=7 --alsologtostderr
ha_test.go:511: (dbg) Run:  kubectl get nodes
ha_test.go:519: (dbg) Run:  kubectl get nodes -o "go-template='{{range .items}}{{range .status.conditions}}{{if eq .type "Ready"}} {{.status}}{{"\n"}}{{end}}{{end}}{{end}}'"
--- PASS: TestMultiControlPlane/serial/DeleteSecondaryNode (9.11s)

                                                
                                    
x
+
TestMultiControlPlane/serial/DegradedAfterSecondaryNodeDelete (0.44s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/DegradedAfterSecondaryNodeDelete
ha_test.go:390: (dbg) Run:  out/minikube-linux-amd64 profile list --output json
--- PASS: TestMultiControlPlane/serial/DegradedAfterSecondaryNodeDelete (0.44s)

                                                
                                    
x
+
TestMultiControlPlane/serial/StopCluster (32.47s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/StopCluster
ha_test.go:531: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 stop -v=7 --alsologtostderr
ha_test.go:531: (dbg) Done: out/minikube-linux-amd64 -p ha-243920 stop -v=7 --alsologtostderr: (32.374633004s)
ha_test.go:537: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 status -v=7 --alsologtostderr
ha_test.go:537: (dbg) Non-zero exit: out/minikube-linux-amd64 -p ha-243920 status -v=7 --alsologtostderr: exit status 7 (90.937607ms)

                                                
                                                
-- stdout --
	ha-243920
	type: Control Plane
	host: Stopped
	kubelet: Stopped
	apiserver: Stopped
	kubeconfig: Stopped
	
	ha-243920-m02
	type: Control Plane
	host: Stopped
	kubelet: Stopped
	apiserver: Stopped
	kubeconfig: Stopped
	
	ha-243920-m04
	type: Worker
	host: Stopped
	kubelet: Stopped
	

                                                
                                                
-- /stdout --
** stderr ** 
	I0913 18:46:14.731894  126218 out.go:345] Setting OutFile to fd 1 ...
	I0913 18:46:14.732000  126218 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:46:14.732008  126218 out.go:358] Setting ErrFile to fd 2...
	I0913 18:46:14.732013  126218 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:46:14.732231  126218 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
	I0913 18:46:14.732384  126218 out.go:352] Setting JSON to false
	I0913 18:46:14.732408  126218 mustload.go:65] Loading cluster: ha-243920
	I0913 18:46:14.732525  126218 notify.go:220] Checking for updates...
	I0913 18:46:14.732953  126218 config.go:182] Loaded profile config "ha-243920": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
	I0913 18:46:14.732973  126218 status.go:255] checking status of ha-243920 ...
	I0913 18:46:14.733526  126218 cli_runner.go:164] Run: docker container inspect ha-243920 --format={{.State.Status}}
	I0913 18:46:14.750028  126218 status.go:330] ha-243920 host status = "Stopped" (err=<nil>)
	I0913 18:46:14.750048  126218 status.go:343] host is not running, skipping remaining checks
	I0913 18:46:14.750053  126218 status.go:257] ha-243920 status: &{Name:ha-243920 Host:Stopped Kubelet:Stopped APIServer:Stopped Kubeconfig:Stopped Worker:false TimeToStop: DockerEnv: PodManEnv:}
	I0913 18:46:14.750072  126218 status.go:255] checking status of ha-243920-m02 ...
	I0913 18:46:14.750301  126218 cli_runner.go:164] Run: docker container inspect ha-243920-m02 --format={{.State.Status}}
	I0913 18:46:14.765721  126218 status.go:330] ha-243920-m02 host status = "Stopped" (err=<nil>)
	I0913 18:46:14.765740  126218 status.go:343] host is not running, skipping remaining checks
	I0913 18:46:14.765745  126218 status.go:257] ha-243920-m02 status: &{Name:ha-243920-m02 Host:Stopped Kubelet:Stopped APIServer:Stopped Kubeconfig:Stopped Worker:false TimeToStop: DockerEnv: PodManEnv:}
	I0913 18:46:14.765763  126218 status.go:255] checking status of ha-243920-m04 ...
	I0913 18:46:14.765994  126218 cli_runner.go:164] Run: docker container inspect ha-243920-m04 --format={{.State.Status}}
	I0913 18:46:14.781480  126218 status.go:330] ha-243920-m04 host status = "Stopped" (err=<nil>)
	I0913 18:46:14.781498  126218 status.go:343] host is not running, skipping remaining checks
	I0913 18:46:14.781504  126218 status.go:257] ha-243920-m04 status: &{Name:ha-243920-m04 Host:Stopped Kubelet:Stopped APIServer:Stopped Kubeconfig:Stopped Worker:true TimeToStop: DockerEnv: PodManEnv:}

                                                
                                                
** /stderr **
--- PASS: TestMultiControlPlane/serial/StopCluster (32.47s)

                                                
                                    
x
+
TestMultiControlPlane/serial/RestartCluster (81.09s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/RestartCluster
ha_test.go:560: (dbg) Run:  out/minikube-linux-amd64 start -p ha-243920 --wait=true -v=7 --alsologtostderr --driver=docker  --container-runtime=docker
E0913 18:47:11.538754   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
ha_test.go:560: (dbg) Done: out/minikube-linux-amd64 start -p ha-243920 --wait=true -v=7 --alsologtostderr --driver=docker  --container-runtime=docker: (1m20.383592864s)
ha_test.go:566: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 status -v=7 --alsologtostderr
ha_test.go:584: (dbg) Run:  kubectl get nodes
ha_test.go:592: (dbg) Run:  kubectl get nodes -o "go-template='{{range .items}}{{range .status.conditions}}{{if eq .type "Ready"}} {{.status}}{{"\n"}}{{end}}{{end}}{{end}}'"
--- PASS: TestMultiControlPlane/serial/RestartCluster (81.09s)

                                                
                                    
x
+
TestMultiControlPlane/serial/DegradedAfterClusterRestart (0.44s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/DegradedAfterClusterRestart
ha_test.go:390: (dbg) Run:  out/minikube-linux-amd64 profile list --output json
--- PASS: TestMultiControlPlane/serial/DegradedAfterClusterRestart (0.44s)

                                                
                                    
x
+
TestMultiControlPlane/serial/AddSecondaryNode (39.53s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/AddSecondaryNode
ha_test.go:605: (dbg) Run:  out/minikube-linux-amd64 node add -p ha-243920 --control-plane -v=7 --alsologtostderr
E0913 18:47:39.240651   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
ha_test.go:605: (dbg) Done: out/minikube-linux-amd64 node add -p ha-243920 --control-plane -v=7 --alsologtostderr: (38.767824096s)
ha_test.go:611: (dbg) Run:  out/minikube-linux-amd64 -p ha-243920 status -v=7 --alsologtostderr
--- PASS: TestMultiControlPlane/serial/AddSecondaryNode (39.53s)

                                                
                                    
x
+
TestMultiControlPlane/serial/HAppyAfterSecondaryNodeAdd (0.6s)

                                                
                                                
=== RUN   TestMultiControlPlane/serial/HAppyAfterSecondaryNodeAdd
ha_test.go:281: (dbg) Run:  out/minikube-linux-amd64 profile list --output json
--- PASS: TestMultiControlPlane/serial/HAppyAfterSecondaryNodeAdd (0.60s)

                                                
                                    
x
+
TestImageBuild/serial/Setup (19.35s)

                                                
                                                
=== RUN   TestImageBuild/serial/Setup
image_test.go:69: (dbg) Run:  out/minikube-linux-amd64 start -p image-365884 --driver=docker  --container-runtime=docker
image_test.go:69: (dbg) Done: out/minikube-linux-amd64 start -p image-365884 --driver=docker  --container-runtime=docker: (19.354253843s)
--- PASS: TestImageBuild/serial/Setup (19.35s)

                                                
                                    
x
+
TestImageBuild/serial/NormalBuild (1.13s)

                                                
                                                
=== RUN   TestImageBuild/serial/NormalBuild
image_test.go:78: (dbg) Run:  out/minikube-linux-amd64 image build -t aaa:latest ./testdata/image-build/test-normal -p image-365884
image_test.go:78: (dbg) Done: out/minikube-linux-amd64 image build -t aaa:latest ./testdata/image-build/test-normal -p image-365884: (1.130469148s)
--- PASS: TestImageBuild/serial/NormalBuild (1.13s)

                                                
                                    
x
+
TestImageBuild/serial/BuildWithBuildArg (0.7s)

                                                
                                                
=== RUN   TestImageBuild/serial/BuildWithBuildArg
image_test.go:99: (dbg) Run:  out/minikube-linux-amd64 image build -t aaa:latest --build-opt=build-arg=ENV_A=test_env_str --build-opt=no-cache ./testdata/image-build/test-arg -p image-365884
--- PASS: TestImageBuild/serial/BuildWithBuildArg (0.70s)

                                                
                                    
x
+
TestImageBuild/serial/BuildWithDockerIgnore (0.49s)

                                                
                                                
=== RUN   TestImageBuild/serial/BuildWithDockerIgnore
image_test.go:133: (dbg) Run:  out/minikube-linux-amd64 image build -t aaa:latest ./testdata/image-build/test-normal --build-opt=no-cache -p image-365884
--- PASS: TestImageBuild/serial/BuildWithDockerIgnore (0.49s)

                                                
                                    
x
+
TestImageBuild/serial/BuildWithSpecifiedDockerfile (0.51s)

                                                
                                                
=== RUN   TestImageBuild/serial/BuildWithSpecifiedDockerfile
image_test.go:88: (dbg) Run:  out/minikube-linux-amd64 image build -t aaa:latest -f inner/Dockerfile ./testdata/image-build/test-f -p image-365884
--- PASS: TestImageBuild/serial/BuildWithSpecifiedDockerfile (0.51s)

                                                
                                    
x
+
TestJSONOutput/start/Command (65.12s)

                                                
                                                
=== RUN   TestJSONOutput/start/Command
json_output_test.go:63: (dbg) Run:  out/minikube-linux-amd64 start -p json-output-539399 --output=json --user=testUser --memory=2200 --wait=true --driver=docker  --container-runtime=docker
E0913 18:49:03.998973   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
json_output_test.go:63: (dbg) Done: out/minikube-linux-amd64 start -p json-output-539399 --output=json --user=testUser --memory=2200 --wait=true --driver=docker  --container-runtime=docker: (1m5.114786215s)
--- PASS: TestJSONOutput/start/Command (65.12s)

                                                
                                    
x
+
TestJSONOutput/start/Audit (0s)

                                                
                                                
=== RUN   TestJSONOutput/start/Audit
--- PASS: TestJSONOutput/start/Audit (0.00s)

                                                
                                    
x
+
TestJSONOutput/start/parallel/DistinctCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/start/parallel/DistinctCurrentSteps
=== PAUSE TestJSONOutput/start/parallel/DistinctCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/start/parallel/DistinctCurrentSteps
--- PASS: TestJSONOutput/start/parallel/DistinctCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/start/parallel/IncreasingCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/start/parallel/IncreasingCurrentSteps
=== PAUSE TestJSONOutput/start/parallel/IncreasingCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/start/parallel/IncreasingCurrentSteps
--- PASS: TestJSONOutput/start/parallel/IncreasingCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/pause/Command (0.52s)

                                                
                                                
=== RUN   TestJSONOutput/pause/Command
json_output_test.go:63: (dbg) Run:  out/minikube-linux-amd64 pause -p json-output-539399 --output=json --user=testUser
--- PASS: TestJSONOutput/pause/Command (0.52s)

                                                
                                    
x
+
TestJSONOutput/pause/Audit (0s)

                                                
                                                
=== RUN   TestJSONOutput/pause/Audit
--- PASS: TestJSONOutput/pause/Audit (0.00s)

                                                
                                    
x
+
TestJSONOutput/pause/parallel/DistinctCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/pause/parallel/DistinctCurrentSteps
=== PAUSE TestJSONOutput/pause/parallel/DistinctCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/pause/parallel/DistinctCurrentSteps
--- PASS: TestJSONOutput/pause/parallel/DistinctCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/pause/parallel/IncreasingCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/pause/parallel/IncreasingCurrentSteps
=== PAUSE TestJSONOutput/pause/parallel/IncreasingCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/pause/parallel/IncreasingCurrentSteps
--- PASS: TestJSONOutput/pause/parallel/IncreasingCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/unpause/Command (0.41s)

                                                
                                                
=== RUN   TestJSONOutput/unpause/Command
json_output_test.go:63: (dbg) Run:  out/minikube-linux-amd64 unpause -p json-output-539399 --output=json --user=testUser
--- PASS: TestJSONOutput/unpause/Command (0.41s)

                                                
                                    
x
+
TestJSONOutput/unpause/Audit (0s)

                                                
                                                
=== RUN   TestJSONOutput/unpause/Audit
--- PASS: TestJSONOutput/unpause/Audit (0.00s)

                                                
                                    
x
+
TestJSONOutput/unpause/parallel/DistinctCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/unpause/parallel/DistinctCurrentSteps
=== PAUSE TestJSONOutput/unpause/parallel/DistinctCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/unpause/parallel/DistinctCurrentSteps
--- PASS: TestJSONOutput/unpause/parallel/DistinctCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/unpause/parallel/IncreasingCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/unpause/parallel/IncreasingCurrentSteps
=== PAUSE TestJSONOutput/unpause/parallel/IncreasingCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/unpause/parallel/IncreasingCurrentSteps
--- PASS: TestJSONOutput/unpause/parallel/IncreasingCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/stop/Command (10.78s)

                                                
                                                
=== RUN   TestJSONOutput/stop/Command
json_output_test.go:63: (dbg) Run:  out/minikube-linux-amd64 stop -p json-output-539399 --output=json --user=testUser
json_output_test.go:63: (dbg) Done: out/minikube-linux-amd64 stop -p json-output-539399 --output=json --user=testUser: (10.783102138s)
--- PASS: TestJSONOutput/stop/Command (10.78s)

                                                
                                    
x
+
TestJSONOutput/stop/Audit (0s)

                                                
                                                
=== RUN   TestJSONOutput/stop/Audit
--- PASS: TestJSONOutput/stop/Audit (0.00s)

                                                
                                    
x
+
TestJSONOutput/stop/parallel/DistinctCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/stop/parallel/DistinctCurrentSteps
=== PAUSE TestJSONOutput/stop/parallel/DistinctCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/stop/parallel/DistinctCurrentSteps
--- PASS: TestJSONOutput/stop/parallel/DistinctCurrentSteps (0.00s)

                                                
                                    
x
+
TestJSONOutput/stop/parallel/IncreasingCurrentSteps (0s)

                                                
                                                
=== RUN   TestJSONOutput/stop/parallel/IncreasingCurrentSteps
=== PAUSE TestJSONOutput/stop/parallel/IncreasingCurrentSteps

                                                
                                                

                                                
                                                
=== CONT  TestJSONOutput/stop/parallel/IncreasingCurrentSteps
--- PASS: TestJSONOutput/stop/parallel/IncreasingCurrentSteps (0.00s)

                                                
                                    
x
+
TestErrorJSONOutput (0.19s)

                                                
                                                
=== RUN   TestErrorJSONOutput
json_output_test.go:160: (dbg) Run:  out/minikube-linux-amd64 start -p json-output-error-747750 --memory=2200 --output=json --wait=true --driver=fail
json_output_test.go:160: (dbg) Non-zero exit: out/minikube-linux-amd64 start -p json-output-error-747750 --memory=2200 --output=json --wait=true --driver=fail: exit status 56 (59.84418ms)

                                                
                                                
-- stdout --
	{"specversion":"1.0","id":"3eacf1ab-dde6-4603-abd1-ba36a945f85a","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.step","datacontenttype":"application/json","data":{"currentstep":"0","message":"[json-output-error-747750] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)","name":"Initial Minikube Setup","totalsteps":"19"}}
	{"specversion":"1.0","id":"107c189a-936e-4081-b81a-46e86848587e","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_LOCATION=19636"}}
	{"specversion":"1.0","id":"81bc6894-fda0-4c32-af5d-1a485375ad48","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true"}}
	{"specversion":"1.0","id":"387697e6-3318-4d8a-a782-bdc78dd474dc","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"KUBECONFIG=/home/jenkins/minikube-integration/19636-3724/kubeconfig"}}
	{"specversion":"1.0","id":"3111b742-cc5d-46db-a14b-fb08a6e60f28","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_HOME=/home/jenkins/minikube-integration/19636-3724/.minikube"}}
	{"specversion":"1.0","id":"46bfc802-a379-4122-a909-8bc3b1ff2e4c","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_BIN=out/minikube-linux-amd64"}}
	{"specversion":"1.0","id":"a7419bda-bc38-4f31-9a2d-b7475f22c91a","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_FORCE_SYSTEMD="}}
	{"specversion":"1.0","id":"6c1bc69f-aa89-4279-b59e-7447e9ce8552","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.error","datacontenttype":"application/json","data":{"advice":"","exitcode":"56","issues":"","message":"The driver 'fail' is not supported on linux/amd64","name":"DRV_UNSUPPORTED_OS","url":""}}

                                                
                                                
-- /stdout --
helpers_test.go:175: Cleaning up "json-output-error-747750" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p json-output-error-747750
--- PASS: TestErrorJSONOutput (0.19s)

                                                
                                    
x
+
TestKicCustomNetwork/create_custom_network (22.58s)

                                                
                                                
=== RUN   TestKicCustomNetwork/create_custom_network
kic_custom_network_test.go:57: (dbg) Run:  out/minikube-linux-amd64 start -p docker-network-763933 --network=
kic_custom_network_test.go:57: (dbg) Done: out/minikube-linux-amd64 start -p docker-network-763933 --network=: (20.662340884s)
kic_custom_network_test.go:150: (dbg) Run:  docker network ls --format {{.Name}}
helpers_test.go:175: Cleaning up "docker-network-763933" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p docker-network-763933
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p docker-network-763933: (1.901122732s)
--- PASS: TestKicCustomNetwork/create_custom_network (22.58s)

                                                
                                    
x
+
TestKicCustomNetwork/use_default_bridge_network (24.79s)

                                                
                                                
=== RUN   TestKicCustomNetwork/use_default_bridge_network
kic_custom_network_test.go:57: (dbg) Run:  out/minikube-linux-amd64 start -p docker-network-715319 --network=bridge
kic_custom_network_test.go:57: (dbg) Done: out/minikube-linux-amd64 start -p docker-network-715319 --network=bridge: (22.981935697s)
kic_custom_network_test.go:150: (dbg) Run:  docker network ls --format {{.Name}}
helpers_test.go:175: Cleaning up "docker-network-715319" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p docker-network-715319
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p docker-network-715319: (1.789835544s)
--- PASS: TestKicCustomNetwork/use_default_bridge_network (24.79s)

                                                
                                    
x
+
TestKicExistingNetwork (22.02s)

                                                
                                                
=== RUN   TestKicExistingNetwork
kic_custom_network_test.go:150: (dbg) Run:  docker network ls --format {{.Name}}
kic_custom_network_test.go:93: (dbg) Run:  out/minikube-linux-amd64 start -p existing-network-082065 --network=existing-network
kic_custom_network_test.go:93: (dbg) Done: out/minikube-linux-amd64 start -p existing-network-082065 --network=existing-network: (20.02249967s)
helpers_test.go:175: Cleaning up "existing-network-082065" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p existing-network-082065
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p existing-network-082065: (1.863640834s)
--- PASS: TestKicExistingNetwork (22.02s)

                                                
                                    
x
+
TestKicCustomSubnet (25.34s)

                                                
                                                
=== RUN   TestKicCustomSubnet
kic_custom_network_test.go:112: (dbg) Run:  out/minikube-linux-amd64 start -p custom-subnet-259125 --subnet=192.168.60.0/24
kic_custom_network_test.go:112: (dbg) Done: out/minikube-linux-amd64 start -p custom-subnet-259125 --subnet=192.168.60.0/24: (23.36287943s)
kic_custom_network_test.go:161: (dbg) Run:  docker network inspect custom-subnet-259125 --format "{{(index .IPAM.Config 0).Subnet}}"
helpers_test.go:175: Cleaning up "custom-subnet-259125" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p custom-subnet-259125
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p custom-subnet-259125: (1.95635082s)
--- PASS: TestKicCustomSubnet (25.34s)

                                                
                                    
x
+
TestKicStaticIP (25.46s)

                                                
                                                
=== RUN   TestKicStaticIP
kic_custom_network_test.go:132: (dbg) Run:  out/minikube-linux-amd64 start -p static-ip-183646 --static-ip=192.168.200.200
kic_custom_network_test.go:132: (dbg) Done: out/minikube-linux-amd64 start -p static-ip-183646 --static-ip=192.168.200.200: (23.426206736s)
kic_custom_network_test.go:138: (dbg) Run:  out/minikube-linux-amd64 -p static-ip-183646 ip
helpers_test.go:175: Cleaning up "static-ip-183646" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p static-ip-183646
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p static-ip-183646: (1.913162793s)
--- PASS: TestKicStaticIP (25.46s)

                                                
                                    
x
+
TestMainNoArgs (0.04s)

                                                
                                                
=== RUN   TestMainNoArgs
main_test.go:68: (dbg) Run:  out/minikube-linux-amd64
--- PASS: TestMainNoArgs (0.04s)

                                                
                                    
x
+
TestMinikubeProfile (52.08s)

                                                
                                                
=== RUN   TestMinikubeProfile
minikube_profile_test.go:44: (dbg) Run:  out/minikube-linux-amd64 start -p first-352814 --driver=docker  --container-runtime=docker
E0913 18:52:11.538840   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
minikube_profile_test.go:44: (dbg) Done: out/minikube-linux-amd64 start -p first-352814 --driver=docker  --container-runtime=docker: (23.452360803s)
minikube_profile_test.go:44: (dbg) Run:  out/minikube-linux-amd64 start -p second-366613 --driver=docker  --container-runtime=docker
minikube_profile_test.go:44: (dbg) Done: out/minikube-linux-amd64 start -p second-366613 --driver=docker  --container-runtime=docker: (23.674176974s)
minikube_profile_test.go:51: (dbg) Run:  out/minikube-linux-amd64 profile first-352814
minikube_profile_test.go:55: (dbg) Run:  out/minikube-linux-amd64 profile list -ojson
minikube_profile_test.go:51: (dbg) Run:  out/minikube-linux-amd64 profile second-366613
minikube_profile_test.go:55: (dbg) Run:  out/minikube-linux-amd64 profile list -ojson
helpers_test.go:175: Cleaning up "second-366613" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p second-366613
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p second-366613: (1.980244048s)
helpers_test.go:175: Cleaning up "first-352814" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p first-352814
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p first-352814: (1.996843842s)
--- PASS: TestMinikubeProfile (52.08s)

                                                
                                    
x
+
TestMountStart/serial/StartWithMountFirst (6.09s)

                                                
                                                
=== RUN   TestMountStart/serial/StartWithMountFirst
mount_start_test.go:98: (dbg) Run:  out/minikube-linux-amd64 start -p mount-start-1-080603 --memory=2048 --mount --mount-gid 0 --mount-msize 6543 --mount-port 46464 --mount-uid 0 --no-kubernetes --driver=docker  --container-runtime=docker
mount_start_test.go:98: (dbg) Done: out/minikube-linux-amd64 start -p mount-start-1-080603 --memory=2048 --mount --mount-gid 0 --mount-msize 6543 --mount-port 46464 --mount-uid 0 --no-kubernetes --driver=docker  --container-runtime=docker: (5.088188641s)
--- PASS: TestMountStart/serial/StartWithMountFirst (6.09s)

                                                
                                    
x
+
TestMountStart/serial/VerifyMountFirst (0.23s)

                                                
                                                
=== RUN   TestMountStart/serial/VerifyMountFirst
mount_start_test.go:114: (dbg) Run:  out/minikube-linux-amd64 -p mount-start-1-080603 ssh -- ls /minikube-host
--- PASS: TestMountStart/serial/VerifyMountFirst (0.23s)

                                                
                                    
x
+
TestMountStart/serial/StartWithMountSecond (9.07s)

                                                
                                                
=== RUN   TestMountStart/serial/StartWithMountSecond
mount_start_test.go:98: (dbg) Run:  out/minikube-linux-amd64 start -p mount-start-2-093112 --memory=2048 --mount --mount-gid 0 --mount-msize 6543 --mount-port 46465 --mount-uid 0 --no-kubernetes --driver=docker  --container-runtime=docker
mount_start_test.go:98: (dbg) Done: out/minikube-linux-amd64 start -p mount-start-2-093112 --memory=2048 --mount --mount-gid 0 --mount-msize 6543 --mount-port 46465 --mount-uid 0 --no-kubernetes --driver=docker  --container-runtime=docker: (8.06921457s)
--- PASS: TestMountStart/serial/StartWithMountSecond (9.07s)

                                                
                                    
x
+
TestMountStart/serial/VerifyMountSecond (0.23s)

                                                
                                                
=== RUN   TestMountStart/serial/VerifyMountSecond
mount_start_test.go:114: (dbg) Run:  out/minikube-linux-amd64 -p mount-start-2-093112 ssh -- ls /minikube-host
--- PASS: TestMountStart/serial/VerifyMountSecond (0.23s)

                                                
                                    
x
+
TestMountStart/serial/DeleteFirst (1.43s)

                                                
                                                
=== RUN   TestMountStart/serial/DeleteFirst
pause_test.go:132: (dbg) Run:  out/minikube-linux-amd64 delete -p mount-start-1-080603 --alsologtostderr -v=5
pause_test.go:132: (dbg) Done: out/minikube-linux-amd64 delete -p mount-start-1-080603 --alsologtostderr -v=5: (1.431551515s)
--- PASS: TestMountStart/serial/DeleteFirst (1.43s)

                                                
                                    
x
+
TestMountStart/serial/VerifyMountPostDelete (0.23s)

                                                
                                                
=== RUN   TestMountStart/serial/VerifyMountPostDelete
mount_start_test.go:114: (dbg) Run:  out/minikube-linux-amd64 -p mount-start-2-093112 ssh -- ls /minikube-host
--- PASS: TestMountStart/serial/VerifyMountPostDelete (0.23s)

                                                
                                    
x
+
TestMountStart/serial/Stop (1.16s)

                                                
                                                
=== RUN   TestMountStart/serial/Stop
mount_start_test.go:155: (dbg) Run:  out/minikube-linux-amd64 stop -p mount-start-2-093112
mount_start_test.go:155: (dbg) Done: out/minikube-linux-amd64 stop -p mount-start-2-093112: (1.160136933s)
--- PASS: TestMountStart/serial/Stop (1.16s)

                                                
                                    
x
+
TestMountStart/serial/RestartStopped (7.44s)

                                                
                                                
=== RUN   TestMountStart/serial/RestartStopped
mount_start_test.go:166: (dbg) Run:  out/minikube-linux-amd64 start -p mount-start-2-093112
mount_start_test.go:166: (dbg) Done: out/minikube-linux-amd64 start -p mount-start-2-093112: (6.443158089s)
--- PASS: TestMountStart/serial/RestartStopped (7.44s)

                                                
                                    
x
+
TestMountStart/serial/VerifyMountPostStop (0.23s)

                                                
                                                
=== RUN   TestMountStart/serial/VerifyMountPostStop
mount_start_test.go:114: (dbg) Run:  out/minikube-linux-amd64 -p mount-start-2-093112 ssh -- ls /minikube-host
--- PASS: TestMountStart/serial/VerifyMountPostStop (0.23s)

                                                
                                    
x
+
TestMultiNode/serial/FreshStart2Nodes (68.98s)

                                                
                                                
=== RUN   TestMultiNode/serial/FreshStart2Nodes
multinode_test.go:96: (dbg) Run:  out/minikube-linux-amd64 start -p multinode-715475 --wait=true --memory=2200 --nodes=2 -v=8 --alsologtostderr --driver=docker  --container-runtime=docker
E0913 18:54:03.999407   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
multinode_test.go:96: (dbg) Done: out/minikube-linux-amd64 start -p multinode-715475 --wait=true --memory=2200 --nodes=2 -v=8 --alsologtostderr --driver=docker  --container-runtime=docker: (1m8.550159835s)
multinode_test.go:102: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 status --alsologtostderr
--- PASS: TestMultiNode/serial/FreshStart2Nodes (68.98s)

                                                
                                    
x
+
TestMultiNode/serial/DeployApp2Nodes (38.37s)

                                                
                                                
=== RUN   TestMultiNode/serial/DeployApp2Nodes
multinode_test.go:493: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- apply -f ./testdata/multinodes/multinode-pod-dns-test.yaml
multinode_test.go:498: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- rollout status deployment/busybox
multinode_test.go:498: (dbg) Done: out/minikube-linux-amd64 kubectl -p multinode-715475 -- rollout status deployment/busybox: (2.839367313s)
multinode_test.go:505: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- get pods -o jsonpath='{.items[*].status.podIP}'
multinode_test.go:514: expected 2 Pod IPs but got 1 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.3'\n\n-- /stdout --"
multinode_test.go:505: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- get pods -o jsonpath='{.items[*].status.podIP}'
multinode_test.go:514: expected 2 Pod IPs but got 1 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.3'\n\n-- /stdout --"
multinode_test.go:505: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- get pods -o jsonpath='{.items[*].status.podIP}'
multinode_test.go:514: expected 2 Pod IPs but got 1 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.3'\n\n-- /stdout --"
multinode_test.go:505: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- get pods -o jsonpath='{.items[*].status.podIP}'
multinode_test.go:514: expected 2 Pod IPs but got 1 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.3'\n\n-- /stdout --"
multinode_test.go:505: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- get pods -o jsonpath='{.items[*].status.podIP}'
multinode_test.go:514: expected 2 Pod IPs but got 1 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.3'\n\n-- /stdout --"
multinode_test.go:505: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- get pods -o jsonpath='{.items[*].status.podIP}'
multinode_test.go:514: expected 2 Pod IPs but got 1 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.3'\n\n-- /stdout --"
multinode_test.go:505: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- get pods -o jsonpath='{.items[*].status.podIP}'
multinode_test.go:514: expected 2 Pod IPs but got 1 (may be temporary), output: "\n-- stdout --\n\t'10.244.0.3'\n\n-- /stdout --"
multinode_test.go:505: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- get pods -o jsonpath='{.items[*].status.podIP}'
multinode_test.go:528: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- get pods -o jsonpath='{.items[*].metadata.name}'
multinode_test.go:536: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- exec busybox-7dff88458-679xw -- nslookup kubernetes.io
multinode_test.go:536: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- exec busybox-7dff88458-68mnl -- nslookup kubernetes.io
multinode_test.go:546: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- exec busybox-7dff88458-679xw -- nslookup kubernetes.default
multinode_test.go:546: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- exec busybox-7dff88458-68mnl -- nslookup kubernetes.default
multinode_test.go:554: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- exec busybox-7dff88458-679xw -- nslookup kubernetes.default.svc.cluster.local
multinode_test.go:554: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- exec busybox-7dff88458-68mnl -- nslookup kubernetes.default.svc.cluster.local
--- PASS: TestMultiNode/serial/DeployApp2Nodes (38.37s)

                                                
                                    
x
+
TestMultiNode/serial/PingHostFrom2Pods (0.69s)

                                                
                                                
=== RUN   TestMultiNode/serial/PingHostFrom2Pods
multinode_test.go:564: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- get pods -o jsonpath='{.items[*].metadata.name}'
multinode_test.go:572: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- exec busybox-7dff88458-679xw -- sh -c "nslookup host.minikube.internal | awk 'NR==5' | cut -d' ' -f3"
multinode_test.go:583: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- exec busybox-7dff88458-679xw -- sh -c "ping -c 1 192.168.67.1"
multinode_test.go:572: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- exec busybox-7dff88458-68mnl -- sh -c "nslookup host.minikube.internal | awk 'NR==5' | cut -d' ' -f3"
multinode_test.go:583: (dbg) Run:  out/minikube-linux-amd64 kubectl -p multinode-715475 -- exec busybox-7dff88458-68mnl -- sh -c "ping -c 1 192.168.67.1"
--- PASS: TestMultiNode/serial/PingHostFrom2Pods (0.69s)

                                                
                                    
x
+
TestMultiNode/serial/AddNode (18.43s)

                                                
                                                
=== RUN   TestMultiNode/serial/AddNode
multinode_test.go:121: (dbg) Run:  out/minikube-linux-amd64 node add -p multinode-715475 -v 3 --alsologtostderr
E0913 18:55:27.063474   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
multinode_test.go:121: (dbg) Done: out/minikube-linux-amd64 node add -p multinode-715475 -v 3 --alsologtostderr: (17.8497174s)
multinode_test.go:127: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 status --alsologtostderr
--- PASS: TestMultiNode/serial/AddNode (18.43s)

                                                
                                    
x
+
TestMultiNode/serial/MultiNodeLabels (0.06s)

                                                
                                                
=== RUN   TestMultiNode/serial/MultiNodeLabels
multinode_test.go:221: (dbg) Run:  kubectl --context multinode-715475 get nodes -o "jsonpath=[{range .items[*]}{.metadata.labels},{end}]"
--- PASS: TestMultiNode/serial/MultiNodeLabels (0.06s)

                                                
                                    
x
+
TestMultiNode/serial/ProfileList (0.27s)

                                                
                                                
=== RUN   TestMultiNode/serial/ProfileList
multinode_test.go:143: (dbg) Run:  out/minikube-linux-amd64 profile list --output json
--- PASS: TestMultiNode/serial/ProfileList (0.27s)

                                                
                                    
x
+
TestMultiNode/serial/CopyFile (8.6s)

                                                
                                                
=== RUN   TestMultiNode/serial/CopyFile
multinode_test.go:184: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 status --output json --alsologtostderr
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 cp testdata/cp-test.txt multinode-715475:/home/docker/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 cp multinode-715475:/home/docker/cp-test.txt /tmp/TestMultiNodeserialCopyFile611648241/001/cp-test_multinode-715475.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 cp multinode-715475:/home/docker/cp-test.txt multinode-715475-m02:/home/docker/cp-test_multinode-715475_multinode-715475-m02.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475-m02 "sudo cat /home/docker/cp-test_multinode-715475_multinode-715475-m02.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 cp multinode-715475:/home/docker/cp-test.txt multinode-715475-m03:/home/docker/cp-test_multinode-715475_multinode-715475-m03.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475-m03 "sudo cat /home/docker/cp-test_multinode-715475_multinode-715475-m03.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 cp testdata/cp-test.txt multinode-715475-m02:/home/docker/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475-m02 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 cp multinode-715475-m02:/home/docker/cp-test.txt /tmp/TestMultiNodeserialCopyFile611648241/001/cp-test_multinode-715475-m02.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475-m02 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 cp multinode-715475-m02:/home/docker/cp-test.txt multinode-715475:/home/docker/cp-test_multinode-715475-m02_multinode-715475.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475-m02 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475 "sudo cat /home/docker/cp-test_multinode-715475-m02_multinode-715475.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 cp multinode-715475-m02:/home/docker/cp-test.txt multinode-715475-m03:/home/docker/cp-test_multinode-715475-m02_multinode-715475-m03.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475-m02 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475-m03 "sudo cat /home/docker/cp-test_multinode-715475-m02_multinode-715475-m03.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 cp testdata/cp-test.txt multinode-715475-m03:/home/docker/cp-test.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475-m03 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 cp multinode-715475-m03:/home/docker/cp-test.txt /tmp/TestMultiNodeserialCopyFile611648241/001/cp-test_multinode-715475-m03.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475-m03 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 cp multinode-715475-m03:/home/docker/cp-test.txt multinode-715475:/home/docker/cp-test_multinode-715475-m03_multinode-715475.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475-m03 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475 "sudo cat /home/docker/cp-test_multinode-715475-m03_multinode-715475.txt"
helpers_test.go:556: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 cp multinode-715475-m03:/home/docker/cp-test.txt multinode-715475-m02:/home/docker/cp-test_multinode-715475-m03_multinode-715475-m02.txt
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475-m03 "sudo cat /home/docker/cp-test.txt"
helpers_test.go:534: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 ssh -n multinode-715475-m02 "sudo cat /home/docker/cp-test_multinode-715475-m03_multinode-715475-m02.txt"
--- PASS: TestMultiNode/serial/CopyFile (8.60s)

                                                
                                    
x
+
TestMultiNode/serial/StopNode (2.03s)

                                                
                                                
=== RUN   TestMultiNode/serial/StopNode
multinode_test.go:248: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 node stop m03
multinode_test.go:248: (dbg) Done: out/minikube-linux-amd64 -p multinode-715475 node stop m03: (1.16414317s)
multinode_test.go:254: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 status
multinode_test.go:254: (dbg) Non-zero exit: out/minikube-linux-amd64 -p multinode-715475 status: exit status 7 (428.066453ms)

                                                
                                                
-- stdout --
	multinode-715475
	type: Control Plane
	host: Running
	kubelet: Running
	apiserver: Running
	kubeconfig: Configured
	
	multinode-715475-m02
	type: Worker
	host: Running
	kubelet: Running
	
	multinode-715475-m03
	type: Worker
	host: Stopped
	kubelet: Stopped
	

                                                
                                                
-- /stdout --
multinode_test.go:261: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 status --alsologtostderr
multinode_test.go:261: (dbg) Non-zero exit: out/minikube-linux-amd64 -p multinode-715475 status --alsologtostderr: exit status 7 (435.637202ms)

                                                
                                                
-- stdout --
	multinode-715475
	type: Control Plane
	host: Running
	kubelet: Running
	apiserver: Running
	kubeconfig: Configured
	
	multinode-715475-m02
	type: Worker
	host: Running
	kubelet: Running
	
	multinode-715475-m03
	type: Worker
	host: Stopped
	kubelet: Stopped
	

                                                
                                                
-- /stdout --
** stderr ** 
	I0913 18:55:43.630621  212792 out.go:345] Setting OutFile to fd 1 ...
	I0913 18:55:43.630724  212792 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:55:43.630734  212792 out.go:358] Setting ErrFile to fd 2...
	I0913 18:55:43.630739  212792 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:55:43.630914  212792 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
	I0913 18:55:43.631106  212792 out.go:352] Setting JSON to false
	I0913 18:55:43.631140  212792 mustload.go:65] Loading cluster: multinode-715475
	I0913 18:55:43.631233  212792 notify.go:220] Checking for updates...
	I0913 18:55:43.631630  212792 config.go:182] Loaded profile config "multinode-715475": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
	I0913 18:55:43.631646  212792 status.go:255] checking status of multinode-715475 ...
	I0913 18:55:43.632049  212792 cli_runner.go:164] Run: docker container inspect multinode-715475 --format={{.State.Status}}
	I0913 18:55:43.649044  212792 status.go:330] multinode-715475 host status = "Running" (err=<nil>)
	I0913 18:55:43.649071  212792 host.go:66] Checking if "multinode-715475" exists ...
	I0913 18:55:43.649335  212792 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" multinode-715475
	I0913 18:55:43.665642  212792 host.go:66] Checking if "multinode-715475" exists ...
	I0913 18:55:43.665923  212792 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0913 18:55:43.665977  212792 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" multinode-715475
	I0913 18:55:43.681807  212792 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32909 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/multinode-715475/id_rsa Username:docker}
	I0913 18:55:43.770419  212792 ssh_runner.go:195] Run: systemctl --version
	I0913 18:55:43.774174  212792 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0913 18:55:43.783869  212792 cli_runner.go:164] Run: docker system info --format "{{json .}}"
	I0913 18:55:43.831822  212792 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:3 ContainersRunning:2 ContainersPaused:0 ContainersStopped:1 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:42 OomKillDisable:true NGoroutines:62 SystemTime:2024-09-13 18:55:43.822763385 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
	I0913 18:55:43.832388  212792 kubeconfig.go:125] found "multinode-715475" server: "https://192.168.67.2:8443"
	I0913 18:55:43.832415  212792 api_server.go:166] Checking apiserver status ...
	I0913 18:55:43.832455  212792 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
	I0913 18:55:43.842714  212792 ssh_runner.go:195] Run: sudo egrep ^[0-9]+:freezer: /proc/2341/cgroup
	I0913 18:55:43.850732  212792 api_server.go:182] apiserver freezer: "7:freezer:/docker/45d07c5f255eec2266c4e0b5f8db7270e15961f6b85db184b4ae031f833c4c76/kubepods/burstable/pod15a1d261f85a96f0d06b142e5be65f29/093e98570dc4cd34e7a03e570dfca17334fdbb2eeb2bb65e7408dfed94fabb0c"
	I0913 18:55:43.850787  212792 ssh_runner.go:195] Run: sudo cat /sys/fs/cgroup/freezer/docker/45d07c5f255eec2266c4e0b5f8db7270e15961f6b85db184b4ae031f833c4c76/kubepods/burstable/pod15a1d261f85a96f0d06b142e5be65f29/093e98570dc4cd34e7a03e570dfca17334fdbb2eeb2bb65e7408dfed94fabb0c/freezer.state
	I0913 18:55:43.858025  212792 api_server.go:204] freezer state: "THAWED"
	I0913 18:55:43.858046  212792 api_server.go:253] Checking apiserver healthz at https://192.168.67.2:8443/healthz ...
	I0913 18:55:43.861584  212792 api_server.go:279] https://192.168.67.2:8443/healthz returned 200:
	ok
	I0913 18:55:43.861607  212792 status.go:422] multinode-715475 apiserver status = Running (err=<nil>)
	I0913 18:55:43.861617  212792 status.go:257] multinode-715475 status: &{Name:multinode-715475 Host:Running Kubelet:Running APIServer:Running Kubeconfig:Configured Worker:false TimeToStop: DockerEnv: PodManEnv:}
	I0913 18:55:43.861644  212792 status.go:255] checking status of multinode-715475-m02 ...
	I0913 18:55:43.861898  212792 cli_runner.go:164] Run: docker container inspect multinode-715475-m02 --format={{.State.Status}}
	I0913 18:55:43.877767  212792 status.go:330] multinode-715475-m02 host status = "Running" (err=<nil>)
	I0913 18:55:43.877789  212792 host.go:66] Checking if "multinode-715475-m02" exists ...
	I0913 18:55:43.878074  212792 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" multinode-715475-m02
	I0913 18:55:43.893239  212792 host.go:66] Checking if "multinode-715475-m02" exists ...
	I0913 18:55:43.893543  212792 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
	I0913 18:55:43.893596  212792 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" multinode-715475-m02
	I0913 18:55:43.909037  212792 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32914 SSHKeyPath:/home/jenkins/minikube-integration/19636-3724/.minikube/machines/multinode-715475-m02/id_rsa Username:docker}
	I0913 18:55:43.998318  212792 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
	I0913 18:55:44.008293  212792 status.go:257] multinode-715475-m02 status: &{Name:multinode-715475-m02 Host:Running Kubelet:Running APIServer:Irrelevant Kubeconfig:Irrelevant Worker:true TimeToStop: DockerEnv: PodManEnv:}
	I0913 18:55:44.008324  212792 status.go:255] checking status of multinode-715475-m03 ...
	I0913 18:55:44.008622  212792 cli_runner.go:164] Run: docker container inspect multinode-715475-m03 --format={{.State.Status}}
	I0913 18:55:44.024575  212792 status.go:330] multinode-715475-m03 host status = "Stopped" (err=<nil>)
	I0913 18:55:44.024594  212792 status.go:343] host is not running, skipping remaining checks
	I0913 18:55:44.024605  212792 status.go:257] multinode-715475-m03 status: &{Name:multinode-715475-m03 Host:Stopped Kubelet:Stopped APIServer:Stopped Kubeconfig:Stopped Worker:true TimeToStop: DockerEnv: PodManEnv:}

                                                
                                                
** /stderr **
--- PASS: TestMultiNode/serial/StopNode (2.03s)

                                                
                                    
x
+
TestMultiNode/serial/StartAfterStop (9.53s)

                                                
                                                
=== RUN   TestMultiNode/serial/StartAfterStop
multinode_test.go:282: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 node start m03 -v=7 --alsologtostderr
multinode_test.go:282: (dbg) Done: out/minikube-linux-amd64 -p multinode-715475 node start m03 -v=7 --alsologtostderr: (8.899488003s)
multinode_test.go:290: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 status -v=7 --alsologtostderr
multinode_test.go:306: (dbg) Run:  kubectl get nodes
--- PASS: TestMultiNode/serial/StartAfterStop (9.53s)

                                                
                                    
x
+
TestMultiNode/serial/RestartKeepsNodes (97.9s)

                                                
                                                
=== RUN   TestMultiNode/serial/RestartKeepsNodes
multinode_test.go:314: (dbg) Run:  out/minikube-linux-amd64 node list -p multinode-715475
multinode_test.go:321: (dbg) Run:  out/minikube-linux-amd64 stop -p multinode-715475
multinode_test.go:321: (dbg) Done: out/minikube-linux-amd64 stop -p multinode-715475: (22.459057329s)
multinode_test.go:326: (dbg) Run:  out/minikube-linux-amd64 start -p multinode-715475 --wait=true -v=8 --alsologtostderr
E0913 18:57:11.538460   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
multinode_test.go:326: (dbg) Done: out/minikube-linux-amd64 start -p multinode-715475 --wait=true -v=8 --alsologtostderr: (1m15.355671951s)
multinode_test.go:331: (dbg) Run:  out/minikube-linux-amd64 node list -p multinode-715475
--- PASS: TestMultiNode/serial/RestartKeepsNodes (97.90s)

                                                
                                    
x
+
TestMultiNode/serial/DeleteNode (5.04s)

                                                
                                                
=== RUN   TestMultiNode/serial/DeleteNode
multinode_test.go:416: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 node delete m03
multinode_test.go:416: (dbg) Done: out/minikube-linux-amd64 -p multinode-715475 node delete m03: (4.526821151s)
multinode_test.go:422: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 status --alsologtostderr
multinode_test.go:436: (dbg) Run:  kubectl get nodes
multinode_test.go:444: (dbg) Run:  kubectl get nodes -o "go-template='{{range .items}}{{range .status.conditions}}{{if eq .type "Ready"}} {{.status}}{{"\n"}}{{end}}{{end}}{{end}}'"
--- PASS: TestMultiNode/serial/DeleteNode (5.04s)

                                                
                                    
x
+
TestMultiNode/serial/StopMultiNode (21.3s)

                                                
                                                
=== RUN   TestMultiNode/serial/StopMultiNode
multinode_test.go:345: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 stop
multinode_test.go:345: (dbg) Done: out/minikube-linux-amd64 -p multinode-715475 stop: (21.143589616s)
multinode_test.go:351: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 status
multinode_test.go:351: (dbg) Non-zero exit: out/minikube-linux-amd64 -p multinode-715475 status: exit status 7 (75.26988ms)

                                                
                                                
-- stdout --
	multinode-715475
	type: Control Plane
	host: Stopped
	kubelet: Stopped
	apiserver: Stopped
	kubeconfig: Stopped
	
	multinode-715475-m02
	type: Worker
	host: Stopped
	kubelet: Stopped
	

                                                
                                                
-- /stdout --
multinode_test.go:358: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 status --alsologtostderr
multinode_test.go:358: (dbg) Non-zero exit: out/minikube-linux-amd64 -p multinode-715475 status --alsologtostderr: exit status 7 (76.420843ms)

                                                
                                                
-- stdout --
	multinode-715475
	type: Control Plane
	host: Stopped
	kubelet: Stopped
	apiserver: Stopped
	kubeconfig: Stopped
	
	multinode-715475-m02
	type: Worker
	host: Stopped
	kubelet: Stopped
	

                                                
                                                
-- /stdout --
** stderr ** 
	I0913 18:57:57.751863  228246 out.go:345] Setting OutFile to fd 1 ...
	I0913 18:57:57.752108  228246 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:57:57.752118  228246 out.go:358] Setting ErrFile to fd 2...
	I0913 18:57:57.752124  228246 out.go:392] TERM=,COLORTERM=, which probably does not support color
	I0913 18:57:57.752308  228246 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19636-3724/.minikube/bin
	I0913 18:57:57.752485  228246 out.go:352] Setting JSON to false
	I0913 18:57:57.752518  228246 mustload.go:65] Loading cluster: multinode-715475
	I0913 18:57:57.752596  228246 notify.go:220] Checking for updates...
	I0913 18:57:57.752919  228246 config.go:182] Loaded profile config "multinode-715475": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
	I0913 18:57:57.752934  228246 status.go:255] checking status of multinode-715475 ...
	I0913 18:57:57.753400  228246 cli_runner.go:164] Run: docker container inspect multinode-715475 --format={{.State.Status}}
	I0913 18:57:57.769890  228246 status.go:330] multinode-715475 host status = "Stopped" (err=<nil>)
	I0913 18:57:57.769908  228246 status.go:343] host is not running, skipping remaining checks
	I0913 18:57:57.769913  228246 status.go:257] multinode-715475 status: &{Name:multinode-715475 Host:Stopped Kubelet:Stopped APIServer:Stopped Kubeconfig:Stopped Worker:false TimeToStop: DockerEnv: PodManEnv:}
	I0913 18:57:57.769940  228246 status.go:255] checking status of multinode-715475-m02 ...
	I0913 18:57:57.770169  228246 cli_runner.go:164] Run: docker container inspect multinode-715475-m02 --format={{.State.Status}}
	I0913 18:57:57.785452  228246 status.go:330] multinode-715475-m02 host status = "Stopped" (err=<nil>)
	I0913 18:57:57.785467  228246 status.go:343] host is not running, skipping remaining checks
	I0913 18:57:57.785472  228246 status.go:257] multinode-715475-m02 status: &{Name:multinode-715475-m02 Host:Stopped Kubelet:Stopped APIServer:Stopped Kubeconfig:Stopped Worker:true TimeToStop: DockerEnv: PodManEnv:}

                                                
                                                
** /stderr **
--- PASS: TestMultiNode/serial/StopMultiNode (21.30s)

                                                
                                    
x
+
TestMultiNode/serial/RestartMultiNode (52.67s)

                                                
                                                
=== RUN   TestMultiNode/serial/RestartMultiNode
multinode_test.go:376: (dbg) Run:  out/minikube-linux-amd64 start -p multinode-715475 --wait=true -v=8 --alsologtostderr --driver=docker  --container-runtime=docker
E0913 18:58:34.602434   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
multinode_test.go:376: (dbg) Done: out/minikube-linux-amd64 start -p multinode-715475 --wait=true -v=8 --alsologtostderr --driver=docker  --container-runtime=docker: (52.139814036s)
multinode_test.go:382: (dbg) Run:  out/minikube-linux-amd64 -p multinode-715475 status --alsologtostderr
multinode_test.go:396: (dbg) Run:  kubectl get nodes
multinode_test.go:404: (dbg) Run:  kubectl get nodes -o "go-template='{{range .items}}{{range .status.conditions}}{{if eq .type "Ready"}} {{.status}}{{"\n"}}{{end}}{{end}}{{end}}'"
--- PASS: TestMultiNode/serial/RestartMultiNode (52.67s)

                                                
                                    
x
+
TestMultiNode/serial/ValidateNameConflict (22.82s)

                                                
                                                
=== RUN   TestMultiNode/serial/ValidateNameConflict
multinode_test.go:455: (dbg) Run:  out/minikube-linux-amd64 node list -p multinode-715475
multinode_test.go:464: (dbg) Run:  out/minikube-linux-amd64 start -p multinode-715475-m02 --driver=docker  --container-runtime=docker
multinode_test.go:464: (dbg) Non-zero exit: out/minikube-linux-amd64 start -p multinode-715475-m02 --driver=docker  --container-runtime=docker: exit status 14 (58.549583ms)

                                                
                                                
-- stdout --
	* [multinode-715475-m02] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
	  - MINIKUBE_LOCATION=19636
	  - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	  - KUBECONFIG=/home/jenkins/minikube-integration/19636-3724/kubeconfig
	  - MINIKUBE_HOME=/home/jenkins/minikube-integration/19636-3724/.minikube
	  - MINIKUBE_BIN=out/minikube-linux-amd64
	  - MINIKUBE_FORCE_SYSTEMD=
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	! Profile name 'multinode-715475-m02' is duplicated with machine name 'multinode-715475-m02' in profile 'multinode-715475'
	X Exiting due to MK_USAGE: Profile name should be unique

                                                
                                                
** /stderr **
multinode_test.go:472: (dbg) Run:  out/minikube-linux-amd64 start -p multinode-715475-m03 --driver=docker  --container-runtime=docker
E0913 18:59:03.999131   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
multinode_test.go:472: (dbg) Done: out/minikube-linux-amd64 start -p multinode-715475-m03 --driver=docker  --container-runtime=docker: (20.50019879s)
multinode_test.go:479: (dbg) Run:  out/minikube-linux-amd64 node add -p multinode-715475
multinode_test.go:479: (dbg) Non-zero exit: out/minikube-linux-amd64 node add -p multinode-715475: exit status 80 (243.005848ms)

                                                
                                                
-- stdout --
	* Adding node m03 to cluster multinode-715475 as [worker]
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	X Exiting due to GUEST_NODE_ADD: failed to add node: Node multinode-715475-m03 already exists in multinode-715475-m03 profile
	* 
	╭─────────────────────────────────────────────────────────────────────────────────────────────╮
	│                                                                                             │
	│    * If the above advice does not help, please let us know:                                 │
	│      https://github.com/kubernetes/minikube/issues/new/choose                               │
	│                                                                                             │
	│    * Please run `minikube logs --file=logs.txt` and attach logs.txt to the GitHub issue.    │
	│    * Please also attach the following file to the GitHub issue:                             │
	│    * - /tmp/minikube_node_040ea7097fd6ed71e65be9a474587f81f0ccd21d_0.log                    │
	│                                                                                             │
	╰─────────────────────────────────────────────────────────────────────────────────────────────╯

                                                
                                                
** /stderr **
multinode_test.go:484: (dbg) Run:  out/minikube-linux-amd64 delete -p multinode-715475-m03
multinode_test.go:484: (dbg) Done: out/minikube-linux-amd64 delete -p multinode-715475-m03: (1.974082665s)
--- PASS: TestMultiNode/serial/ValidateNameConflict (22.82s)

                                                
                                    
x
+
TestPreload (85.82s)

                                                
                                                
=== RUN   TestPreload
preload_test.go:44: (dbg) Run:  out/minikube-linux-amd64 start -p test-preload-215445 --memory=2200 --alsologtostderr --wait=true --preload=false --driver=docker  --container-runtime=docker --kubernetes-version=v1.24.4
preload_test.go:44: (dbg) Done: out/minikube-linux-amd64 start -p test-preload-215445 --memory=2200 --alsologtostderr --wait=true --preload=false --driver=docker  --container-runtime=docker --kubernetes-version=v1.24.4: (50.198808173s)
preload_test.go:52: (dbg) Run:  out/minikube-linux-amd64 -p test-preload-215445 image pull gcr.io/k8s-minikube/busybox
preload_test.go:52: (dbg) Done: out/minikube-linux-amd64 -p test-preload-215445 image pull gcr.io/k8s-minikube/busybox: (1.523860466s)
preload_test.go:58: (dbg) Run:  out/minikube-linux-amd64 stop -p test-preload-215445
preload_test.go:58: (dbg) Done: out/minikube-linux-amd64 stop -p test-preload-215445: (10.672522554s)
preload_test.go:66: (dbg) Run:  out/minikube-linux-amd64 start -p test-preload-215445 --memory=2200 --alsologtostderr -v=1 --wait=true --driver=docker  --container-runtime=docker
preload_test.go:66: (dbg) Done: out/minikube-linux-amd64 start -p test-preload-215445 --memory=2200 --alsologtostderr -v=1 --wait=true --driver=docker  --container-runtime=docker: (21.131372476s)
preload_test.go:71: (dbg) Run:  out/minikube-linux-amd64 -p test-preload-215445 image list
helpers_test.go:175: Cleaning up "test-preload-215445" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p test-preload-215445
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p test-preload-215445: (2.098496964s)
--- PASS: TestPreload (85.82s)

                                                
                                    
x
+
TestScheduledStopUnix (96.46s)

                                                
                                                
=== RUN   TestScheduledStopUnix
scheduled_stop_test.go:128: (dbg) Run:  out/minikube-linux-amd64 start -p scheduled-stop-995689 --memory=2048 --driver=docker  --container-runtime=docker
scheduled_stop_test.go:128: (dbg) Done: out/minikube-linux-amd64 start -p scheduled-stop-995689 --memory=2048 --driver=docker  --container-runtime=docker: (23.671903498s)
scheduled_stop_test.go:137: (dbg) Run:  out/minikube-linux-amd64 stop -p scheduled-stop-995689 --schedule 5m
scheduled_stop_test.go:191: (dbg) Run:  out/minikube-linux-amd64 status --format={{.TimeToStop}} -p scheduled-stop-995689 -n scheduled-stop-995689
scheduled_stop_test.go:169: signal error was:  <nil>
scheduled_stop_test.go:137: (dbg) Run:  out/minikube-linux-amd64 stop -p scheduled-stop-995689 --schedule 15s
scheduled_stop_test.go:169: signal error was:  os: process already finished
scheduled_stop_test.go:137: (dbg) Run:  out/minikube-linux-amd64 stop -p scheduled-stop-995689 --cancel-scheduled
scheduled_stop_test.go:176: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Host}} -p scheduled-stop-995689 -n scheduled-stop-995689
scheduled_stop_test.go:205: (dbg) Run:  out/minikube-linux-amd64 status -p scheduled-stop-995689
scheduled_stop_test.go:137: (dbg) Run:  out/minikube-linux-amd64 stop -p scheduled-stop-995689 --schedule 15s
scheduled_stop_test.go:169: signal error was:  os: process already finished
E0913 19:02:11.538719   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
scheduled_stop_test.go:205: (dbg) Run:  out/minikube-linux-amd64 status -p scheduled-stop-995689
scheduled_stop_test.go:205: (dbg) Non-zero exit: out/minikube-linux-amd64 status -p scheduled-stop-995689: exit status 7 (58.325551ms)

                                                
                                                
-- stdout --
	scheduled-stop-995689
	type: Control Plane
	host: Stopped
	kubelet: Stopped
	apiserver: Stopped
	kubeconfig: Stopped
	

                                                
                                                
-- /stdout --
scheduled_stop_test.go:176: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Host}} -p scheduled-stop-995689 -n scheduled-stop-995689
scheduled_stop_test.go:176: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.Host}} -p scheduled-stop-995689 -n scheduled-stop-995689: exit status 7 (59.257962ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
scheduled_stop_test.go:176: status error: exit status 7 (may be ok)
helpers_test.go:175: Cleaning up "scheduled-stop-995689" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p scheduled-stop-995689
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p scheduled-stop-995689: (1.567800942s)
--- PASS: TestScheduledStopUnix (96.46s)

                                                
                                    
x
+
TestSkaffold (93.61s)

                                                
                                                
=== RUN   TestSkaffold
skaffold_test.go:59: (dbg) Run:  /tmp/skaffold.exe1045812693 version
skaffold_test.go:63: skaffold version: v2.13.2
skaffold_test.go:66: (dbg) Run:  out/minikube-linux-amd64 start -p skaffold-408558 --memory=2600 --driver=docker  --container-runtime=docker
skaffold_test.go:66: (dbg) Done: out/minikube-linux-amd64 start -p skaffold-408558 --memory=2600 --driver=docker  --container-runtime=docker: (20.59865444s)
skaffold_test.go:86: copying out/minikube-linux-amd64 to /home/jenkins/workspace/Docker_Linux_integration/out/minikube
skaffold_test.go:105: (dbg) Run:  /tmp/skaffold.exe1045812693 run --minikube-profile skaffold-408558 --kube-context skaffold-408558 --status-check=true --port-forward=false --interactive=false
skaffold_test.go:105: (dbg) Done: /tmp/skaffold.exe1045812693 run --minikube-profile skaffold-408558 --kube-context skaffold-408558 --status-check=true --port-forward=false --interactive=false: (58.473449344s)
skaffold_test.go:111: (dbg) TestSkaffold: waiting 1m0s for pods matching "app=leeroy-app" in namespace "default" ...
helpers_test.go:344: "leeroy-app-7598d75bb8-kkbbf" [0a7eb240-01fd-4c9d-ada9-cd7b6e95b8d9] Running
skaffold_test.go:111: (dbg) TestSkaffold: app=leeroy-app healthy within 6.002826288s
skaffold_test.go:114: (dbg) TestSkaffold: waiting 1m0s for pods matching "app=leeroy-web" in namespace "default" ...
helpers_test.go:344: "leeroy-web-d4586f8c9-v2t46" [0fda0430-609f-4876-8913-d4ccb8f1b5a2] Running
skaffold_test.go:114: (dbg) TestSkaffold: app=leeroy-web healthy within 5.003927922s
helpers_test.go:175: Cleaning up "skaffold-408558" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p skaffold-408558
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p skaffold-408558: (2.710762051s)
--- PASS: TestSkaffold (93.61s)

                                                
                                    
x
+
TestInsufficientStorage (12.51s)

                                                
                                                
=== RUN   TestInsufficientStorage
status_test.go:50: (dbg) Run:  out/minikube-linux-amd64 start -p insufficient-storage-826444 --memory=2048 --output=json --wait=true --driver=docker  --container-runtime=docker
status_test.go:50: (dbg) Non-zero exit: out/minikube-linux-amd64 start -p insufficient-storage-826444 --memory=2048 --output=json --wait=true --driver=docker  --container-runtime=docker: exit status 26 (10.433599467s)

                                                
                                                
-- stdout --
	{"specversion":"1.0","id":"c2fe70b3-6fe1-45d6-a0ad-139b6f90ddbf","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.step","datacontenttype":"application/json","data":{"currentstep":"0","message":"[insufficient-storage-826444] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)","name":"Initial Minikube Setup","totalsteps":"19"}}
	{"specversion":"1.0","id":"3a6d0096-328c-440f-82d0-f97910bcf46f","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_LOCATION=19636"}}
	{"specversion":"1.0","id":"4b4de778-fdeb-4605-ae6f-16fd8b9f4010","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true"}}
	{"specversion":"1.0","id":"c430ab02-2161-4cb3-a9c1-db288cd1d2db","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"KUBECONFIG=/home/jenkins/minikube-integration/19636-3724/kubeconfig"}}
	{"specversion":"1.0","id":"273f936b-880b-40dd-84d6-75eff5f65e61","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_HOME=/home/jenkins/minikube-integration/19636-3724/.minikube"}}
	{"specversion":"1.0","id":"c779e979-beac-47ac-9dad-8a9e139322d4","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_BIN=out/minikube-linux-amd64"}}
	{"specversion":"1.0","id":"f9eb64d0-a23e-4161-a155-5c45d4e97f59","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_FORCE_SYSTEMD="}}
	{"specversion":"1.0","id":"efaec4cc-59ce-4053-ada3-07ab7b324952","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_TEST_STORAGE_CAPACITY=100"}}
	{"specversion":"1.0","id":"f876d4c6-ebfd-49f2-b391-86f2c4de71d4","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"MINIKUBE_TEST_AVAILABLE_STORAGE=19"}}
	{"specversion":"1.0","id":"bba6ed0d-c0d6-49fe-9a39-af57b6b70253","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.step","datacontenttype":"application/json","data":{"currentstep":"1","message":"Using the docker driver based on user configuration","name":"Selecting Driver","totalsteps":"19"}}
	{"specversion":"1.0","id":"300d565c-9534-44fd-b191-f7b9ef17afbc","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.info","datacontenttype":"application/json","data":{"message":"Using Docker driver with root privileges"}}
	{"specversion":"1.0","id":"5eff63fe-b854-4883-923e-6a7c5fd4dae2","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.step","datacontenttype":"application/json","data":{"currentstep":"3","message":"Starting \"insufficient-storage-826444\" primary control-plane node in \"insufficient-storage-826444\" cluster","name":"Starting Node","totalsteps":"19"}}
	{"specversion":"1.0","id":"ba58b0ae-4ceb-4097-83e9-ba727fc993fb","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.step","datacontenttype":"application/json","data":{"currentstep":"5","message":"Pulling base image v0.0.45-1726193793-19634 ...","name":"Pulling Base Image","totalsteps":"19"}}
	{"specversion":"1.0","id":"df739840-4f60-419a-84c4-da897093164f","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.step","datacontenttype":"application/json","data":{"currentstep":"8","message":"Creating docker container (CPUs=2, Memory=2048MB) ...","name":"Creating Container","totalsteps":"19"}}
	{"specversion":"1.0","id":"6e279222-8eda-4c4a-bd39-22bb5f5dc847","source":"https://minikube.sigs.k8s.io/","type":"io.k8s.sigs.minikube.error","datacontenttype":"application/json","data":{"advice":"Try one or more of the following to free up space on the device:\n\t\n\t\t\t1. Run \"docker system prune\" to remove unused Docker data (optionally with \"-a\")\n\t\t\t2. Increase the storage allocated to Docker for Desktop by clicking on:\n\t\t\t\tDocker icon \u003e Preferences \u003e Resources \u003e Disk Image Size\n\t\t\t3. Run \"minikube ssh -- docker system prune\" if using the Docker container runtime","exitcode":"26","issues":"https://github.com/kubernetes/minikube/issues/9024","message":"Docker is out of disk space! (/var is at 100% of capacity). You can pass '--force' to skip this check.","name":"RSRC_DOCKER_STORAGE","url":""}}

                                                
                                                
-- /stdout --
status_test.go:76: (dbg) Run:  out/minikube-linux-amd64 status -p insufficient-storage-826444 --output=json --layout=cluster
status_test.go:76: (dbg) Non-zero exit: out/minikube-linux-amd64 status -p insufficient-storage-826444 --output=json --layout=cluster: exit status 7 (241.001091ms)

                                                
                                                
-- stdout --
	{"Name":"insufficient-storage-826444","StatusCode":507,"StatusName":"InsufficientStorage","StatusDetail":"/var is almost out of disk space","Step":"Creating Container","StepDetail":"Creating docker container (CPUs=2, Memory=2048MB) ...","BinaryVersion":"v1.34.0","Components":{"kubeconfig":{"Name":"kubeconfig","StatusCode":500,"StatusName":"Error"}},"Nodes":[{"Name":"insufficient-storage-826444","StatusCode":507,"StatusName":"InsufficientStorage","Components":{"apiserver":{"Name":"apiserver","StatusCode":405,"StatusName":"Stopped"},"kubelet":{"Name":"kubelet","StatusCode":405,"StatusName":"Stopped"}}}]}

                                                
                                                
-- /stdout --
** stderr ** 
	E0913 19:04:03.540484  268160 status.go:417] kubeconfig endpoint: get endpoint: "insufficient-storage-826444" does not appear in /home/jenkins/minikube-integration/19636-3724/kubeconfig

                                                
                                                
** /stderr **
status_test.go:76: (dbg) Run:  out/minikube-linux-amd64 status -p insufficient-storage-826444 --output=json --layout=cluster
status_test.go:76: (dbg) Non-zero exit: out/minikube-linux-amd64 status -p insufficient-storage-826444 --output=json --layout=cluster: exit status 7 (246.20785ms)

                                                
                                                
-- stdout --
	{"Name":"insufficient-storage-826444","StatusCode":507,"StatusName":"InsufficientStorage","StatusDetail":"/var is almost out of disk space","BinaryVersion":"v1.34.0","Components":{"kubeconfig":{"Name":"kubeconfig","StatusCode":500,"StatusName":"Error"}},"Nodes":[{"Name":"insufficient-storage-826444","StatusCode":507,"StatusName":"InsufficientStorage","Components":{"apiserver":{"Name":"apiserver","StatusCode":405,"StatusName":"Stopped"},"kubelet":{"Name":"kubelet","StatusCode":405,"StatusName":"Stopped"}}}]}

                                                
                                                
-- /stdout --
** stderr ** 
	E0913 19:04:03.787371  268260 status.go:417] kubeconfig endpoint: get endpoint: "insufficient-storage-826444" does not appear in /home/jenkins/minikube-integration/19636-3724/kubeconfig
	E0913 19:04:03.796555  268260 status.go:560] unable to read event log: stat: stat /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/insufficient-storage-826444/events.json: no such file or directory

                                                
                                                
** /stderr **
helpers_test.go:175: Cleaning up "insufficient-storage-826444" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p insufficient-storage-826444
E0913 19:04:03.999203   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p insufficient-storage-826444: (1.589602669s)
--- PASS: TestInsufficientStorage (12.51s)

                                                
                                    
x
+
TestRunningBinaryUpgrade (67.12s)

                                                
                                                
=== RUN   TestRunningBinaryUpgrade
=== PAUSE TestRunningBinaryUpgrade

                                                
                                                

                                                
                                                
=== CONT  TestRunningBinaryUpgrade
version_upgrade_test.go:120: (dbg) Run:  /tmp/minikube-v1.26.0.353899750 start -p running-upgrade-739706 --memory=2200 --vm-driver=docker  --container-runtime=docker
version_upgrade_test.go:120: (dbg) Done: /tmp/minikube-v1.26.0.353899750 start -p running-upgrade-739706 --memory=2200 --vm-driver=docker  --container-runtime=docker: (31.339662309s)
version_upgrade_test.go:130: (dbg) Run:  out/minikube-linux-amd64 start -p running-upgrade-739706 --memory=2200 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker
version_upgrade_test.go:130: (dbg) Done: out/minikube-linux-amd64 start -p running-upgrade-739706 --memory=2200 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker: (32.789702481s)
helpers_test.go:175: Cleaning up "running-upgrade-739706" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p running-upgrade-739706
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p running-upgrade-739706: (2.251351523s)
--- PASS: TestRunningBinaryUpgrade (67.12s)

                                                
                                    
x
+
TestKubernetesUpgrade (332.95s)

                                                
                                                
=== RUN   TestKubernetesUpgrade
=== PAUSE TestKubernetesUpgrade

                                                
                                                

                                                
                                                
=== CONT  TestKubernetesUpgrade
version_upgrade_test.go:222: (dbg) Run:  out/minikube-linux-amd64 start -p kubernetes-upgrade-725508 --memory=2200 --kubernetes-version=v1.20.0 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker
version_upgrade_test.go:222: (dbg) Done: out/minikube-linux-amd64 start -p kubernetes-upgrade-725508 --memory=2200 --kubernetes-version=v1.20.0 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker: (31.886288936s)
version_upgrade_test.go:227: (dbg) Run:  out/minikube-linux-amd64 stop -p kubernetes-upgrade-725508
version_upgrade_test.go:227: (dbg) Done: out/minikube-linux-amd64 stop -p kubernetes-upgrade-725508: (3.232699522s)
version_upgrade_test.go:232: (dbg) Run:  out/minikube-linux-amd64 -p kubernetes-upgrade-725508 status --format={{.Host}}
version_upgrade_test.go:232: (dbg) Non-zero exit: out/minikube-linux-amd64 -p kubernetes-upgrade-725508 status --format={{.Host}}: exit status 7 (80.695016ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
version_upgrade_test.go:234: status error: exit status 7 (may be ok)
version_upgrade_test.go:243: (dbg) Run:  out/minikube-linux-amd64 start -p kubernetes-upgrade-725508 --memory=2200 --kubernetes-version=v1.31.1 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker
E0913 19:07:11.538256   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
version_upgrade_test.go:243: (dbg) Done: out/minikube-linux-amd64 start -p kubernetes-upgrade-725508 --memory=2200 --kubernetes-version=v1.31.1 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker: (4m29.278087029s)
version_upgrade_test.go:248: (dbg) Run:  kubectl --context kubernetes-upgrade-725508 version --output=json
version_upgrade_test.go:267: Attempting to downgrade Kubernetes (should fail)
version_upgrade_test.go:269: (dbg) Run:  out/minikube-linux-amd64 start -p kubernetes-upgrade-725508 --memory=2200 --kubernetes-version=v1.20.0 --driver=docker  --container-runtime=docker
version_upgrade_test.go:269: (dbg) Non-zero exit: out/minikube-linux-amd64 start -p kubernetes-upgrade-725508 --memory=2200 --kubernetes-version=v1.20.0 --driver=docker  --container-runtime=docker: exit status 106 (88.751621ms)

                                                
                                                
-- stdout --
	* [kubernetes-upgrade-725508] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
	  - MINIKUBE_LOCATION=19636
	  - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	  - KUBECONFIG=/home/jenkins/minikube-integration/19636-3724/kubeconfig
	  - MINIKUBE_HOME=/home/jenkins/minikube-integration/19636-3724/.minikube
	  - MINIKUBE_BIN=out/minikube-linux-amd64
	  - MINIKUBE_FORCE_SYSTEMD=
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	X Exiting due to K8S_DOWNGRADE_UNSUPPORTED: Unable to safely downgrade existing Kubernetes v1.31.1 cluster to v1.20.0
	* Suggestion: 
	
	    1) Recreate the cluster with Kubernetes 1.20.0, by running:
	    
	    minikube delete -p kubernetes-upgrade-725508
	    minikube start -p kubernetes-upgrade-725508 --kubernetes-version=v1.20.0
	    
	    2) Create a second cluster with Kubernetes 1.20.0, by running:
	    
	    minikube start -p kubernetes-upgrade-7255082 --kubernetes-version=v1.20.0
	    
	    3) Use the existing cluster at version Kubernetes 1.31.1, by running:
	    
	    minikube start -p kubernetes-upgrade-725508 --kubernetes-version=v1.31.1
	    

                                                
                                                
** /stderr **
version_upgrade_test.go:273: Attempting restart after unsuccessful downgrade
version_upgrade_test.go:275: (dbg) Run:  out/minikube-linux-amd64 start -p kubernetes-upgrade-725508 --memory=2200 --kubernetes-version=v1.31.1 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker
version_upgrade_test.go:275: (dbg) Done: out/minikube-linux-amd64 start -p kubernetes-upgrade-725508 --memory=2200 --kubernetes-version=v1.31.1 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker: (22.864394039s)
helpers_test.go:175: Cleaning up "kubernetes-upgrade-725508" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p kubernetes-upgrade-725508
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p kubernetes-upgrade-725508: (5.451729263s)
--- PASS: TestKubernetesUpgrade (332.95s)

                                                
                                    
x
+
TestMissingContainerUpgrade (93.48s)

                                                
                                                
=== RUN   TestMissingContainerUpgrade
=== PAUSE TestMissingContainerUpgrade

                                                
                                                

                                                
                                                
=== CONT  TestMissingContainerUpgrade
version_upgrade_test.go:309: (dbg) Run:  /tmp/minikube-v1.26.0.2965501112 start -p missing-upgrade-526212 --memory=2200 --driver=docker  --container-runtime=docker
version_upgrade_test.go:309: (dbg) Done: /tmp/minikube-v1.26.0.2965501112 start -p missing-upgrade-526212 --memory=2200 --driver=docker  --container-runtime=docker: (28.252277229s)
version_upgrade_test.go:318: (dbg) Run:  docker stop missing-upgrade-526212
version_upgrade_test.go:318: (dbg) Done: docker stop missing-upgrade-526212: (10.709488484s)
version_upgrade_test.go:323: (dbg) Run:  docker rm missing-upgrade-526212
version_upgrade_test.go:329: (dbg) Run:  out/minikube-linux-amd64 start -p missing-upgrade-526212 --memory=2200 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker
version_upgrade_test.go:329: (dbg) Done: out/minikube-linux-amd64 start -p missing-upgrade-526212 --memory=2200 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker: (51.923263082s)
helpers_test.go:175: Cleaning up "missing-upgrade-526212" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p missing-upgrade-526212
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p missing-upgrade-526212: (2.020925096s)
--- PASS: TestMissingContainerUpgrade (93.48s)

                                                
                                    
x
+
TestStoppedBinaryUpgrade/Setup (0.61s)

                                                
                                                
=== RUN   TestStoppedBinaryUpgrade/Setup
--- PASS: TestStoppedBinaryUpgrade/Setup (0.61s)

                                                
                                    
x
+
TestNoKubernetes/serial/StartNoK8sWithVersion (0.07s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/StartNoK8sWithVersion
no_kubernetes_test.go:83: (dbg) Run:  out/minikube-linux-amd64 start -p NoKubernetes-380550 --no-kubernetes --kubernetes-version=1.20 --driver=docker  --container-runtime=docker
no_kubernetes_test.go:83: (dbg) Non-zero exit: out/minikube-linux-amd64 start -p NoKubernetes-380550 --no-kubernetes --kubernetes-version=1.20 --driver=docker  --container-runtime=docker: exit status 14 (69.254338ms)

                                                
                                                
-- stdout --
	* [NoKubernetes-380550] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
	  - MINIKUBE_LOCATION=19636
	  - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
	  - KUBECONFIG=/home/jenkins/minikube-integration/19636-3724/kubeconfig
	  - MINIKUBE_HOME=/home/jenkins/minikube-integration/19636-3724/.minikube
	  - MINIKUBE_BIN=out/minikube-linux-amd64
	  - MINIKUBE_FORCE_SYSTEMD=
	
	

                                                
                                                
-- /stdout --
** stderr ** 
	X Exiting due to MK_USAGE: cannot specify --kubernetes-version with --no-kubernetes,
	to unset a global config run:
	
	$ minikube config unset kubernetes-version

                                                
                                                
** /stderr **
--- PASS: TestNoKubernetes/serial/StartNoK8sWithVersion (0.07s)

                                                
                                    
x
+
TestNoKubernetes/serial/StartWithK8s (38.82s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/StartWithK8s
no_kubernetes_test.go:95: (dbg) Run:  out/minikube-linux-amd64 start -p NoKubernetes-380550 --driver=docker  --container-runtime=docker
no_kubernetes_test.go:95: (dbg) Done: out/minikube-linux-amd64 start -p NoKubernetes-380550 --driver=docker  --container-runtime=docker: (38.525014002s)
no_kubernetes_test.go:200: (dbg) Run:  out/minikube-linux-amd64 -p NoKubernetes-380550 status -o json
--- PASS: TestNoKubernetes/serial/StartWithK8s (38.82s)

                                                
                                    
x
+
TestStoppedBinaryUpgrade/Upgrade (109.87s)

                                                
                                                
=== RUN   TestStoppedBinaryUpgrade/Upgrade
version_upgrade_test.go:183: (dbg) Run:  /tmp/minikube-v1.26.0.4062975392 start -p stopped-upgrade-389943 --memory=2200 --vm-driver=docker  --container-runtime=docker
version_upgrade_test.go:183: (dbg) Done: /tmp/minikube-v1.26.0.4062975392 start -p stopped-upgrade-389943 --memory=2200 --vm-driver=docker  --container-runtime=docker: (1m11.872198407s)
version_upgrade_test.go:192: (dbg) Run:  /tmp/minikube-v1.26.0.4062975392 -p stopped-upgrade-389943 stop
version_upgrade_test.go:192: (dbg) Done: /tmp/minikube-v1.26.0.4062975392 -p stopped-upgrade-389943 stop: (10.850663499s)
version_upgrade_test.go:198: (dbg) Run:  out/minikube-linux-amd64 start -p stopped-upgrade-389943 --memory=2200 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker
version_upgrade_test.go:198: (dbg) Done: out/minikube-linux-amd64 start -p stopped-upgrade-389943 --memory=2200 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker: (27.142846037s)
--- PASS: TestStoppedBinaryUpgrade/Upgrade (109.87s)

                                                
                                    
x
+
TestNoKubernetes/serial/StartWithStopK8s (9.87s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/StartWithStopK8s
no_kubernetes_test.go:112: (dbg) Run:  out/minikube-linux-amd64 start -p NoKubernetes-380550 --no-kubernetes --driver=docker  --container-runtime=docker
no_kubernetes_test.go:112: (dbg) Done: out/minikube-linux-amd64 start -p NoKubernetes-380550 --no-kubernetes --driver=docker  --container-runtime=docker: (7.944410127s)
no_kubernetes_test.go:200: (dbg) Run:  out/minikube-linux-amd64 -p NoKubernetes-380550 status -o json
no_kubernetes_test.go:200: (dbg) Non-zero exit: out/minikube-linux-amd64 -p NoKubernetes-380550 status -o json: exit status 2 (257.165095ms)

                                                
                                                
-- stdout --
	{"Name":"NoKubernetes-380550","Host":"Running","Kubelet":"Stopped","APIServer":"Stopped","Kubeconfig":"Configured","Worker":false}

                                                
                                                
-- /stdout --
no_kubernetes_test.go:124: (dbg) Run:  out/minikube-linux-amd64 delete -p NoKubernetes-380550
no_kubernetes_test.go:124: (dbg) Done: out/minikube-linux-amd64 delete -p NoKubernetes-380550: (1.664324091s)
--- PASS: TestNoKubernetes/serial/StartWithStopK8s (9.87s)

                                                
                                    
x
+
TestNoKubernetes/serial/Start (10.58s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/Start
no_kubernetes_test.go:136: (dbg) Run:  out/minikube-linux-amd64 start -p NoKubernetes-380550 --no-kubernetes --driver=docker  --container-runtime=docker
no_kubernetes_test.go:136: (dbg) Done: out/minikube-linux-amd64 start -p NoKubernetes-380550 --no-kubernetes --driver=docker  --container-runtime=docker: (10.575238854s)
--- PASS: TestNoKubernetes/serial/Start (10.58s)

                                                
                                    
x
+
TestNoKubernetes/serial/VerifyK8sNotRunning (0.27s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/VerifyK8sNotRunning
no_kubernetes_test.go:147: (dbg) Run:  out/minikube-linux-amd64 ssh -p NoKubernetes-380550 "sudo systemctl is-active --quiet service kubelet"
no_kubernetes_test.go:147: (dbg) Non-zero exit: out/minikube-linux-amd64 ssh -p NoKubernetes-380550 "sudo systemctl is-active --quiet service kubelet": exit status 1 (268.488029ms)

                                                
                                                
** stderr ** 
	ssh: Process exited with status 3

                                                
                                                
** /stderr **
--- PASS: TestNoKubernetes/serial/VerifyK8sNotRunning (0.27s)

                                                
                                    
x
+
TestNoKubernetes/serial/ProfileList (1.27s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/ProfileList
no_kubernetes_test.go:169: (dbg) Run:  out/minikube-linux-amd64 profile list
no_kubernetes_test.go:179: (dbg) Run:  out/minikube-linux-amd64 profile list --output=json
--- PASS: TestNoKubernetes/serial/ProfileList (1.27s)

                                                
                                    
x
+
TestNoKubernetes/serial/Stop (1.19s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/Stop
no_kubernetes_test.go:158: (dbg) Run:  out/minikube-linux-amd64 stop -p NoKubernetes-380550
no_kubernetes_test.go:158: (dbg) Done: out/minikube-linux-amd64 stop -p NoKubernetes-380550: (1.192053456s)
--- PASS: TestNoKubernetes/serial/Stop (1.19s)

                                                
                                    
x
+
TestNoKubernetes/serial/StartNoArgs (7.29s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/StartNoArgs
no_kubernetes_test.go:191: (dbg) Run:  out/minikube-linux-amd64 start -p NoKubernetes-380550 --driver=docker  --container-runtime=docker
no_kubernetes_test.go:191: (dbg) Done: out/minikube-linux-amd64 start -p NoKubernetes-380550 --driver=docker  --container-runtime=docker: (7.291149609s)
--- PASS: TestNoKubernetes/serial/StartNoArgs (7.29s)

                                                
                                    
x
+
TestNoKubernetes/serial/VerifyK8sNotRunningSecond (0.26s)

                                                
                                                
=== RUN   TestNoKubernetes/serial/VerifyK8sNotRunningSecond
no_kubernetes_test.go:147: (dbg) Run:  out/minikube-linux-amd64 ssh -p NoKubernetes-380550 "sudo systemctl is-active --quiet service kubelet"
no_kubernetes_test.go:147: (dbg) Non-zero exit: out/minikube-linux-amd64 ssh -p NoKubernetes-380550 "sudo systemctl is-active --quiet service kubelet": exit status 1 (261.380051ms)

                                                
                                                
** stderr ** 
	ssh: Process exited with status 3

                                                
                                                
** /stderr **
--- PASS: TestNoKubernetes/serial/VerifyK8sNotRunningSecond (0.26s)

                                                
                                    
x
+
TestStoppedBinaryUpgrade/MinikubeLogs (1.13s)

                                                
                                                
=== RUN   TestStoppedBinaryUpgrade/MinikubeLogs
version_upgrade_test.go:206: (dbg) Run:  out/minikube-linux-amd64 logs -p stopped-upgrade-389943
version_upgrade_test.go:206: (dbg) Done: out/minikube-linux-amd64 logs -p stopped-upgrade-389943: (1.134910512s)
--- PASS: TestStoppedBinaryUpgrade/MinikubeLogs (1.13s)

                                                
                                    
x
+
TestPause/serial/Start (65.57s)

                                                
                                                
=== RUN   TestPause/serial/Start
pause_test.go:80: (dbg) Run:  out/minikube-linux-amd64 start -p pause-841778 --memory=2048 --install-addons=false --wait=all --driver=docker  --container-runtime=docker
pause_test.go:80: (dbg) Done: out/minikube-linux-amd64 start -p pause-841778 --memory=2048 --install-addons=false --wait=all --driver=docker  --container-runtime=docker: (1m5.56628023s)
--- PASS: TestPause/serial/Start (65.57s)

                                                
                                    
x
+
TestPause/serial/SecondStartNoReconfiguration (34.86s)

                                                
                                                
=== RUN   TestPause/serial/SecondStartNoReconfiguration
pause_test.go:92: (dbg) Run:  out/minikube-linux-amd64 start -p pause-841778 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker
pause_test.go:92: (dbg) Done: out/minikube-linux-amd64 start -p pause-841778 --alsologtostderr -v=1 --driver=docker  --container-runtime=docker: (34.846101357s)
--- PASS: TestPause/serial/SecondStartNoReconfiguration (34.86s)

                                                
                                    
x
+
TestNetworkPlugins/group/auto/Start (64.17s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/auto/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-amd64 start -p auto-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --driver=docker  --container-runtime=docker
net_test.go:112: (dbg) Done: out/minikube-linux-amd64 start -p auto-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --driver=docker  --container-runtime=docker: (1m4.171426787s)
--- PASS: TestNetworkPlugins/group/auto/Start (64.17s)

                                                
                                    
x
+
TestPause/serial/Pause (0.53s)

                                                
                                                
=== RUN   TestPause/serial/Pause
pause_test.go:110: (dbg) Run:  out/minikube-linux-amd64 pause -p pause-841778 --alsologtostderr -v=5
--- PASS: TestPause/serial/Pause (0.53s)

                                                
                                    
x
+
TestPause/serial/VerifyStatus (0.28s)

                                                
                                                
=== RUN   TestPause/serial/VerifyStatus
status_test.go:76: (dbg) Run:  out/minikube-linux-amd64 status -p pause-841778 --output=json --layout=cluster
status_test.go:76: (dbg) Non-zero exit: out/minikube-linux-amd64 status -p pause-841778 --output=json --layout=cluster: exit status 2 (280.883766ms)

                                                
                                                
-- stdout --
	{"Name":"pause-841778","StatusCode":418,"StatusName":"Paused","Step":"Done","StepDetail":"* Paused 12 containers in: kube-system, kubernetes-dashboard, storage-gluster, istio-operator","BinaryVersion":"v1.34.0","Components":{"kubeconfig":{"Name":"kubeconfig","StatusCode":200,"StatusName":"OK"}},"Nodes":[{"Name":"pause-841778","StatusCode":200,"StatusName":"OK","Components":{"apiserver":{"Name":"apiserver","StatusCode":418,"StatusName":"Paused"},"kubelet":{"Name":"kubelet","StatusCode":405,"StatusName":"Stopped"}}}]}

                                                
                                                
-- /stdout --
--- PASS: TestPause/serial/VerifyStatus (0.28s)

                                                
                                    
x
+
TestPause/serial/Unpause (0.43s)

                                                
                                                
=== RUN   TestPause/serial/Unpause
pause_test.go:121: (dbg) Run:  out/minikube-linux-amd64 unpause -p pause-841778 --alsologtostderr -v=5
--- PASS: TestPause/serial/Unpause (0.43s)

                                                
                                    
x
+
TestPause/serial/PauseAgain (0.6s)

                                                
                                                
=== RUN   TestPause/serial/PauseAgain
pause_test.go:110: (dbg) Run:  out/minikube-linux-amd64 pause -p pause-841778 --alsologtostderr -v=5
--- PASS: TestPause/serial/PauseAgain (0.60s)

                                                
                                    
x
+
TestPause/serial/DeletePaused (2.01s)

                                                
                                                
=== RUN   TestPause/serial/DeletePaused
pause_test.go:132: (dbg) Run:  out/minikube-linux-amd64 delete -p pause-841778 --alsologtostderr -v=5
pause_test.go:132: (dbg) Done: out/minikube-linux-amd64 delete -p pause-841778 --alsologtostderr -v=5: (2.012687075s)
--- PASS: TestPause/serial/DeletePaused (2.01s)

                                                
                                    
x
+
TestPause/serial/VerifyDeletedResources (14.76s)

                                                
                                                
=== RUN   TestPause/serial/VerifyDeletedResources
pause_test.go:142: (dbg) Run:  out/minikube-linux-amd64 profile list --output json
pause_test.go:142: (dbg) Done: out/minikube-linux-amd64 profile list --output json: (14.714956047s)
pause_test.go:168: (dbg) Run:  docker ps -a
pause_test.go:173: (dbg) Run:  docker volume inspect pause-841778
pause_test.go:173: (dbg) Non-zero exit: docker volume inspect pause-841778: exit status 1 (15.139117ms)

                                                
                                                
-- stdout --
	[]

                                                
                                                
-- /stdout --
** stderr ** 
	Error response from daemon: get pause-841778: no such volume

                                                
                                                
** /stderr **
pause_test.go:178: (dbg) Run:  docker network ls
--- PASS: TestPause/serial/VerifyDeletedResources (14.76s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/Start (40.86s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-amd64 start -p kindnet-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=kindnet --driver=docker  --container-runtime=docker
net_test.go:112: (dbg) Done: out/minikube-linux-amd64 start -p kindnet-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=kindnet --driver=docker  --container-runtime=docker: (40.857219974s)
--- PASS: TestNetworkPlugins/group/kindnet/Start (40.86s)

                                                
                                    
x
+
TestNetworkPlugins/group/auto/KubeletFlags (0.31s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/auto/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-amd64 ssh -p auto-997126 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/auto/KubeletFlags (0.31s)

                                                
                                    
x
+
TestNetworkPlugins/group/auto/NetCatPod (9.23s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/auto/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context auto-997126 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/auto/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-6fc964789b-q5pp4" [f5905ee3-23e2-4b25-96d8-8dce67150f15] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
E0913 19:08:39.147712   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:08:39.154067   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:08:39.165340   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:08:39.186686   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:08:39.228060   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:08:39.309449   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:08:39.471815   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:08:39.793511   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:08:40.434994   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
helpers_test.go:344: "netcat-6fc964789b-q5pp4" [f5905ee3-23e2-4b25-96d8-8dce67150f15] Running
E0913 19:08:41.717183   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:08:44.278579   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
net_test.go:163: (dbg) TestNetworkPlugins/group/auto/NetCatPod: app=netcat healthy within 9.004557606s
--- PASS: TestNetworkPlugins/group/auto/NetCatPod (9.23s)

                                                
                                    
x
+
TestNetworkPlugins/group/auto/DNS (0.18s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/auto/DNS
net_test.go:175: (dbg) Run:  kubectl --context auto-997126 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/auto/DNS (0.18s)

                                                
                                    
x
+
TestNetworkPlugins/group/auto/Localhost (0.23s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/auto/Localhost
net_test.go:194: (dbg) Run:  kubectl --context auto-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/auto/Localhost (0.23s)

                                                
                                    
x
+
TestNetworkPlugins/group/auto/HairPin (0.14s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/auto/HairPin
net_test.go:264: (dbg) Run:  kubectl --context auto-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/auto/HairPin (0.14s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/Start (55.96s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-amd64 start -p calico-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=calico --driver=docker  --container-runtime=docker
net_test.go:112: (dbg) Done: out/minikube-linux-amd64 start -p calico-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=calico --driver=docker  --container-runtime=docker: (55.958087959s)
--- PASS: TestNetworkPlugins/group/calico/Start (55.96s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/ControllerPod (6.01s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/ControllerPod
net_test.go:120: (dbg) TestNetworkPlugins/group/kindnet/ControllerPod: waiting 10m0s for pods matching "app=kindnet" in namespace "kube-system" ...
helpers_test.go:344: "kindnet-dzf52" [c4275be4-efda-466b-ac01-58178b44e2f9] Running
E0913 19:08:59.642897   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
net_test.go:120: (dbg) TestNetworkPlugins/group/kindnet/ControllerPod: app=kindnet healthy within 6.005286262s
--- PASS: TestNetworkPlugins/group/kindnet/ControllerPod (6.01s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/KubeletFlags (0.29s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-amd64 ssh -p kindnet-997126 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/kindnet/KubeletFlags (0.29s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/NetCatPod (9.23s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context kindnet-997126 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/kindnet/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-6fc964789b-4pvhg" [d5b7a9f9-6c68-43f0-8e6d-4bb6f5e788de] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
helpers_test.go:344: "netcat-6fc964789b-4pvhg" [d5b7a9f9-6c68-43f0-8e6d-4bb6f5e788de] Running
net_test.go:163: (dbg) TestNetworkPlugins/group/kindnet/NetCatPod: app=netcat healthy within 9.00442299s
--- PASS: TestNetworkPlugins/group/kindnet/NetCatPod (9.23s)

                                                
                                    
x
+
TestNetworkPlugins/group/custom-flannel/Start (47.17s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/custom-flannel/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-amd64 start -p custom-flannel-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=testdata/kube-flannel.yaml --driver=docker  --container-runtime=docker
net_test.go:112: (dbg) Done: out/minikube-linux-amd64 start -p custom-flannel-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=testdata/kube-flannel.yaml --driver=docker  --container-runtime=docker: (47.167518946s)
--- PASS: TestNetworkPlugins/group/custom-flannel/Start (47.17s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/DNS (0.14s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/DNS
net_test.go:175: (dbg) Run:  kubectl --context kindnet-997126 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/kindnet/DNS (0.14s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/Localhost (0.13s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/Localhost
net_test.go:194: (dbg) Run:  kubectl --context kindnet-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/kindnet/Localhost (0.13s)

                                                
                                    
x
+
TestNetworkPlugins/group/kindnet/HairPin (0.12s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kindnet/HairPin
net_test.go:264: (dbg) Run:  kubectl --context kindnet-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/kindnet/HairPin (0.12s)

                                                
                                    
x
+
TestNetworkPlugins/group/false/Start (66.04s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/false/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-amd64 start -p false-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=false --driver=docker  --container-runtime=docker
net_test.go:112: (dbg) Done: out/minikube-linux-amd64 start -p false-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=false --driver=docker  --container-runtime=docker: (1m6.039822254s)
--- PASS: TestNetworkPlugins/group/false/Start (66.04s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/ControllerPod (6.01s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/ControllerPod
net_test.go:120: (dbg) TestNetworkPlugins/group/calico/ControllerPod: waiting 10m0s for pods matching "k8s-app=calico-node" in namespace "kube-system" ...
helpers_test.go:344: "calico-node-8blrp" [2334a0fb-12a2-4c07-8137-6a375e2db493] Running
net_test.go:120: (dbg) TestNetworkPlugins/group/calico/ControllerPod: k8s-app=calico-node healthy within 6.004787623s
--- PASS: TestNetworkPlugins/group/calico/ControllerPod (6.01s)

                                                
                                    
x
+
TestNetworkPlugins/group/custom-flannel/KubeletFlags (0.26s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/custom-flannel/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-amd64 ssh -p custom-flannel-997126 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/custom-flannel/KubeletFlags (0.26s)

                                                
                                    
x
+
TestNetworkPlugins/group/custom-flannel/NetCatPod (9.19s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/custom-flannel/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context custom-flannel-997126 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/custom-flannel/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-6fc964789b-s9d6p" [5e516178-8eb3-42d8-8f4d-a9d71fa35fc8] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
helpers_test.go:344: "netcat-6fc964789b-s9d6p" [5e516178-8eb3-42d8-8f4d-a9d71fa35fc8] Running
net_test.go:163: (dbg) TestNetworkPlugins/group/custom-flannel/NetCatPod: app=netcat healthy within 9.004020782s
--- PASS: TestNetworkPlugins/group/custom-flannel/NetCatPod (9.19s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/KubeletFlags (0.3s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-amd64 ssh -p calico-997126 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/calico/KubeletFlags (0.30s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/NetCatPod (10.2s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context calico-997126 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/calico/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-6fc964789b-rc7xd" [d9a19147-0457-4838-9690-48ff564862c1] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
helpers_test.go:344: "netcat-6fc964789b-rc7xd" [d9a19147-0457-4838-9690-48ff564862c1] Running
E0913 19:10:01.086958   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
net_test.go:163: (dbg) TestNetworkPlugins/group/calico/NetCatPod: app=netcat healthy within 10.003233433s
--- PASS: TestNetworkPlugins/group/calico/NetCatPod (10.20s)

                                                
                                    
x
+
TestNetworkPlugins/group/custom-flannel/DNS (0.14s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/custom-flannel/DNS
net_test.go:175: (dbg) Run:  kubectl --context custom-flannel-997126 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/custom-flannel/DNS (0.14s)

                                                
                                    
x
+
TestNetworkPlugins/group/custom-flannel/Localhost (0.11s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/custom-flannel/Localhost
net_test.go:194: (dbg) Run:  kubectl --context custom-flannel-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/custom-flannel/Localhost (0.11s)

                                                
                                    
x
+
TestNetworkPlugins/group/custom-flannel/HairPin (0.12s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/custom-flannel/HairPin
net_test.go:264: (dbg) Run:  kubectl --context custom-flannel-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/custom-flannel/HairPin (0.12s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/DNS (0.12s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/DNS
net_test.go:175: (dbg) Run:  kubectl --context calico-997126 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/calico/DNS (0.12s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/Localhost (0.11s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/Localhost
net_test.go:194: (dbg) Run:  kubectl --context calico-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/calico/Localhost (0.11s)

                                                
                                    
x
+
TestNetworkPlugins/group/calico/HairPin (0.1s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/calico/HairPin
net_test.go:264: (dbg) Run:  kubectl --context calico-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/calico/HairPin (0.10s)

                                                
                                    
x
+
TestNetworkPlugins/group/enable-default-cni/Start (40.86s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/enable-default-cni/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-amd64 start -p enable-default-cni-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --enable-default-cni=true --driver=docker  --container-runtime=docker
net_test.go:112: (dbg) Done: out/minikube-linux-amd64 start -p enable-default-cni-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --enable-default-cni=true --driver=docker  --container-runtime=docker: (40.863387045s)
--- PASS: TestNetworkPlugins/group/enable-default-cni/Start (40.86s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/Start (42.15s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-amd64 start -p flannel-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=flannel --driver=docker  --container-runtime=docker
net_test.go:112: (dbg) Done: out/minikube-linux-amd64 start -p flannel-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=flannel --driver=docker  --container-runtime=docker: (42.150472738s)
--- PASS: TestNetworkPlugins/group/flannel/Start (42.15s)

                                                
                                    
x
+
TestNetworkPlugins/group/false/KubeletFlags (0.26s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/false/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-amd64 ssh -p false-997126 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/false/KubeletFlags (0.26s)

                                                
                                    
x
+
TestNetworkPlugins/group/false/NetCatPod (10.2s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/false/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context false-997126 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/false/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-6fc964789b-dx28j" [c5d7859c-32df-467b-a1ac-c589f0defa82] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
helpers_test.go:344: "netcat-6fc964789b-dx28j" [c5d7859c-32df-467b-a1ac-c589f0defa82] Running
net_test.go:163: (dbg) TestNetworkPlugins/group/false/NetCatPod: app=netcat healthy within 10.003950238s
--- PASS: TestNetworkPlugins/group/false/NetCatPod (10.20s)

                                                
                                    
x
+
TestNetworkPlugins/group/false/DNS (0.17s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/false/DNS
net_test.go:175: (dbg) Run:  kubectl --context false-997126 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/false/DNS (0.17s)

                                                
                                    
x
+
TestNetworkPlugins/group/false/Localhost (0.11s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/false/Localhost
net_test.go:194: (dbg) Run:  kubectl --context false-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/false/Localhost (0.11s)

                                                
                                    
x
+
TestNetworkPlugins/group/false/HairPin (0.11s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/false/HairPin
net_test.go:264: (dbg) Run:  kubectl --context false-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/false/HairPin (0.11s)

                                                
                                    
x
+
TestNetworkPlugins/group/enable-default-cni/KubeletFlags (0.31s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/enable-default-cni/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-amd64 ssh -p enable-default-cni-997126 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/enable-default-cni/KubeletFlags (0.31s)

                                                
                                    
x
+
TestNetworkPlugins/group/enable-default-cni/NetCatPod (10.27s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/enable-default-cni/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context enable-default-cni-997126 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/enable-default-cni/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-6fc964789b-qk4l2" [57340776-d46e-4f40-b5db-28d196bb3921] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
helpers_test.go:344: "netcat-6fc964789b-qk4l2" [57340776-d46e-4f40-b5db-28d196bb3921] Running
net_test.go:163: (dbg) TestNetworkPlugins/group/enable-default-cni/NetCatPod: app=netcat healthy within 10.004011132s
--- PASS: TestNetworkPlugins/group/enable-default-cni/NetCatPod (10.27s)

                                                
                                    
x
+
TestNetworkPlugins/group/bridge/Start (66.67s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/bridge/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-amd64 start -p bridge-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=bridge --driver=docker  --container-runtime=docker
net_test.go:112: (dbg) Done: out/minikube-linux-amd64 start -p bridge-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --cni=bridge --driver=docker  --container-runtime=docker: (1m6.674321777s)
--- PASS: TestNetworkPlugins/group/bridge/Start (66.67s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/ControllerPod (6.01s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/ControllerPod
net_test.go:120: (dbg) TestNetworkPlugins/group/flannel/ControllerPod: waiting 10m0s for pods matching "app=flannel" in namespace "kube-flannel" ...
helpers_test.go:344: "kube-flannel-ds-j9xf5" [f76df0dc-129b-4010-842f-59693a25d343] Running
net_test.go:120: (dbg) TestNetworkPlugins/group/flannel/ControllerPod: app=flannel healthy within 6.003651841s
--- PASS: TestNetworkPlugins/group/flannel/ControllerPod (6.01s)

                                                
                                    
x
+
TestNetworkPlugins/group/enable-default-cni/DNS (0.13s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/enable-default-cni/DNS
net_test.go:175: (dbg) Run:  kubectl --context enable-default-cni-997126 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/enable-default-cni/DNS (0.13s)

                                                
                                    
x
+
TestNetworkPlugins/group/enable-default-cni/Localhost (0.1s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/enable-default-cni/Localhost
net_test.go:194: (dbg) Run:  kubectl --context enable-default-cni-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/enable-default-cni/Localhost (0.10s)

                                                
                                    
x
+
TestNetworkPlugins/group/enable-default-cni/HairPin (0.11s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/enable-default-cni/HairPin
net_test.go:264: (dbg) Run:  kubectl --context enable-default-cni-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/enable-default-cni/HairPin (0.11s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/KubeletFlags (0.24s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-amd64 ssh -p flannel-997126 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/flannel/KubeletFlags (0.24s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/NetCatPod (9.17s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context flannel-997126 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/flannel/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-6fc964789b-l6mv5" [a60466cf-fe89-475b-8abc-9db71eb96dfb] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
helpers_test.go:344: "netcat-6fc964789b-l6mv5" [a60466cf-fe89-475b-8abc-9db71eb96dfb] Running
E0913 19:11:23.009241   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
net_test.go:163: (dbg) TestNetworkPlugins/group/flannel/NetCatPod: app=netcat healthy within 9.004672444s
--- PASS: TestNetworkPlugins/group/flannel/NetCatPod (9.17s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/DNS (0.21s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/DNS
net_test.go:175: (dbg) Run:  kubectl --context flannel-997126 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/flannel/DNS (0.21s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/Localhost (0.18s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/Localhost
net_test.go:194: (dbg) Run:  kubectl --context flannel-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/flannel/Localhost (0.18s)

                                                
                                    
x
+
TestNetworkPlugins/group/flannel/HairPin (0.15s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/flannel/HairPin
net_test.go:264: (dbg) Run:  kubectl --context flannel-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/flannel/HairPin (0.15s)

                                                
                                    
x
+
TestNetworkPlugins/group/kubenet/Start (65.63s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kubenet/Start
net_test.go:112: (dbg) Run:  out/minikube-linux-amd64 start -p kubenet-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --network-plugin=kubenet --driver=docker  --container-runtime=docker
net_test.go:112: (dbg) Done: out/minikube-linux-amd64 start -p kubenet-997126 --memory=3072 --alsologtostderr --wait=true --wait-timeout=15m --network-plugin=kubenet --driver=docker  --container-runtime=docker: (1m5.629904404s)
--- PASS: TestNetworkPlugins/group/kubenet/Start (65.63s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/FirstStart (151.1s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/FirstStart
start_stop_delete_test.go:186: (dbg) Run:  out/minikube-linux-amd64 start -p old-k8s-version-963966 --memory=2200 --alsologtostderr --wait=true --kvm-network=default --kvm-qemu-uri=qemu:///system --disable-driver-mounts --keep-context=false --driver=docker  --container-runtime=docker --kubernetes-version=v1.20.0
start_stop_delete_test.go:186: (dbg) Done: out/minikube-linux-amd64 start -p old-k8s-version-963966 --memory=2200 --alsologtostderr --wait=true --kvm-network=default --kvm-qemu-uri=qemu:///system --disable-driver-mounts --keep-context=false --driver=docker  --container-runtime=docker --kubernetes-version=v1.20.0: (2m31.096078908s)
--- PASS: TestStartStop/group/old-k8s-version/serial/FirstStart (151.10s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/FirstStart (70.09s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/FirstStart
start_stop_delete_test.go:186: (dbg) Run:  out/minikube-linux-amd64 start -p no-preload-230953 --memory=2200 --alsologtostderr --wait=true --preload=false --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1
E0913 19:12:07.065633   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:12:11.538171   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:186: (dbg) Done: out/minikube-linux-amd64 start -p no-preload-230953 --memory=2200 --alsologtostderr --wait=true --preload=false --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1: (1m10.091681168s)
--- PASS: TestStartStop/group/no-preload/serial/FirstStart (70.09s)

                                                
                                    
x
+
TestNetworkPlugins/group/bridge/KubeletFlags (0.29s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/bridge/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-amd64 ssh -p bridge-997126 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/bridge/KubeletFlags (0.29s)

                                                
                                    
x
+
TestNetworkPlugins/group/bridge/NetCatPod (11.21s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/bridge/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context bridge-997126 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/bridge/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-6fc964789b-zmr87" [980f7426-13b2-44b7-9503-be800a41d17f] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
helpers_test.go:344: "netcat-6fc964789b-zmr87" [980f7426-13b2-44b7-9503-be800a41d17f] Running
net_test.go:163: (dbg) TestNetworkPlugins/group/bridge/NetCatPod: app=netcat healthy within 11.003678346s
--- PASS: TestNetworkPlugins/group/bridge/NetCatPod (11.21s)

                                                
                                    
x
+
TestNetworkPlugins/group/bridge/DNS (0.13s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/bridge/DNS
net_test.go:175: (dbg) Run:  kubectl --context bridge-997126 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/bridge/DNS (0.13s)

                                                
                                    
x
+
TestNetworkPlugins/group/bridge/Localhost (0.1s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/bridge/Localhost
net_test.go:194: (dbg) Run:  kubectl --context bridge-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/bridge/Localhost (0.10s)

                                                
                                    
x
+
TestNetworkPlugins/group/bridge/HairPin (0.11s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/bridge/HairPin
net_test.go:264: (dbg) Run:  kubectl --context bridge-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/bridge/HairPin (0.11s)

                                                
                                    
x
+
TestNetworkPlugins/group/kubenet/KubeletFlags (0.27s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kubenet/KubeletFlags
net_test.go:133: (dbg) Run:  out/minikube-linux-amd64 ssh -p kubenet-997126 "pgrep -a kubelet"
--- PASS: TestNetworkPlugins/group/kubenet/KubeletFlags (0.27s)

                                                
                                    
x
+
TestNetworkPlugins/group/kubenet/NetCatPod (10.22s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kubenet/NetCatPod
net_test.go:149: (dbg) Run:  kubectl --context kubenet-997126 replace --force -f testdata/netcat-deployment.yaml
net_test.go:163: (dbg) TestNetworkPlugins/group/kubenet/NetCatPod: waiting 15m0s for pods matching "app=netcat" in namespace "default" ...
helpers_test.go:344: "netcat-6fc964789b-t8jg6" [d8979906-5dbd-4c5c-8ef8-16d266dd4c5c] Pending / Ready:ContainersNotReady (containers with unready status: [dnsutils]) / ContainersReady:ContainersNotReady (containers with unready status: [dnsutils])
helpers_test.go:344: "netcat-6fc964789b-t8jg6" [d8979906-5dbd-4c5c-8ef8-16d266dd4c5c] Running
net_test.go:163: (dbg) TestNetworkPlugins/group/kubenet/NetCatPod: app=netcat healthy within 10.002923499s
--- PASS: TestNetworkPlugins/group/kubenet/NetCatPod (10.22s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/FirstStart (68.73s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/FirstStart
start_stop_delete_test.go:186: (dbg) Run:  out/minikube-linux-amd64 start -p embed-certs-556241 --memory=2200 --alsologtostderr --wait=true --embed-certs --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1
start_stop_delete_test.go:186: (dbg) Done: out/minikube-linux-amd64 start -p embed-certs-556241 --memory=2200 --alsologtostderr --wait=true --embed-certs --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1: (1m8.725568419s)
--- PASS: TestStartStop/group/embed-certs/serial/FirstStart (68.73s)

                                                
                                    
x
+
TestNetworkPlugins/group/kubenet/DNS (0.13s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kubenet/DNS
net_test.go:175: (dbg) Run:  kubectl --context kubenet-997126 exec deployment/netcat -- nslookup kubernetes.default
--- PASS: TestNetworkPlugins/group/kubenet/DNS (0.13s)

                                                
                                    
x
+
TestNetworkPlugins/group/kubenet/Localhost (0.1s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kubenet/Localhost
net_test.go:194: (dbg) Run:  kubectl --context kubenet-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z localhost 8080"
--- PASS: TestNetworkPlugins/group/kubenet/Localhost (0.10s)

                                                
                                    
x
+
TestNetworkPlugins/group/kubenet/HairPin (0.1s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/kubenet/HairPin
net_test.go:264: (dbg) Run:  kubectl --context kubenet-997126 exec deployment/netcat -- /bin/sh -c "nc -w 5 -i 5 -z netcat 8080"
--- PASS: TestNetworkPlugins/group/kubenet/HairPin (0.10s)
E0913 19:17:54.207209   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/FirstStart (64.28s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/FirstStart
start_stop_delete_test.go:186: (dbg) Run:  out/minikube-linux-amd64 start -p default-k8s-diff-port-315900 --memory=2200 --alsologtostderr --wait=true --apiserver-port=8444 --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1
start_stop_delete_test.go:186: (dbg) Done: out/minikube-linux-amd64 start -p default-k8s-diff-port-315900 --memory=2200 --alsologtostderr --wait=true --apiserver-port=8444 --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1: (1m4.275361872s)
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/FirstStart (64.28s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/DeployApp (9.3s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/DeployApp
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context no-preload-230953 create -f testdata/busybox.yaml
start_stop_delete_test.go:196: (dbg) TestStartStop/group/no-preload/serial/DeployApp: waiting 8m0s for pods matching "integration-test=busybox" in namespace "default" ...
helpers_test.go:344: "busybox" [792981ea-5da9-4ba2-8bd5-eb1d27847a3d] Pending
helpers_test.go:344: "busybox" [792981ea-5da9-4ba2-8bd5-eb1d27847a3d] Pending / Ready:ContainersNotReady (containers with unready status: [busybox]) / ContainersReady:ContainersNotReady (containers with unready status: [busybox])
helpers_test.go:344: "busybox" [792981ea-5da9-4ba2-8bd5-eb1d27847a3d] Running
start_stop_delete_test.go:196: (dbg) TestStartStop/group/no-preload/serial/DeployApp: integration-test=busybox healthy within 9.004536482s
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context no-preload-230953 exec busybox -- /bin/sh -c "ulimit -n"
--- PASS: TestStartStop/group/no-preload/serial/DeployApp (9.30s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/EnableAddonWhileActive (0.8s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/EnableAddonWhileActive
start_stop_delete_test.go:205: (dbg) Run:  out/minikube-linux-amd64 addons enable metrics-server -p no-preload-230953 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain
start_stop_delete_test.go:215: (dbg) Run:  kubectl --context no-preload-230953 describe deploy/metrics-server -n kube-system
--- PASS: TestStartStop/group/no-preload/serial/EnableAddonWhileActive (0.80s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/Stop (10.72s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/Stop
start_stop_delete_test.go:228: (dbg) Run:  out/minikube-linux-amd64 stop -p no-preload-230953 --alsologtostderr -v=3
start_stop_delete_test.go:228: (dbg) Done: out/minikube-linux-amd64 stop -p no-preload-230953 --alsologtostderr -v=3: (10.71893149s)
--- PASS: TestStartStop/group/no-preload/serial/Stop (10.72s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/EnableAddonAfterStop (0.23s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/EnableAddonAfterStop
start_stop_delete_test.go:239: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Host}} -p no-preload-230953 -n no-preload-230953
start_stop_delete_test.go:239: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.Host}} -p no-preload-230953 -n no-preload-230953: exit status 7 (132.522861ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:239: status error: exit status 7 (may be ok)
start_stop_delete_test.go:246: (dbg) Run:  out/minikube-linux-amd64 addons enable dashboard -p no-preload-230953 --images=MetricsScraper=registry.k8s.io/echoserver:1.4
--- PASS: TestStartStop/group/no-preload/serial/EnableAddonAfterStop (0.23s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/SecondStart (263.04s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/SecondStart
start_stop_delete_test.go:256: (dbg) Run:  out/minikube-linux-amd64 start -p no-preload-230953 --memory=2200 --alsologtostderr --wait=true --preload=false --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1
E0913 19:13:37.320150   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:37.326491   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:37.337821   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:37.359134   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:37.400634   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:37.482069   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:37.643737   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:37.965196   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:38.607108   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:39.147763   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:39.888645   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:42.450475   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:47.572210   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:256: (dbg) Done: out/minikube-linux-amd64 start -p no-preload-230953 --memory=2200 --alsologtostderr --wait=true --preload=false --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1: (4m22.755215547s)
start_stop_delete_test.go:262: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Host}} -p no-preload-230953 -n no-preload-230953
--- PASS: TestStartStop/group/no-preload/serial/SecondStart (263.04s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/DeployApp (9.25s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/DeployApp
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context embed-certs-556241 create -f testdata/busybox.yaml
start_stop_delete_test.go:196: (dbg) TestStartStop/group/embed-certs/serial/DeployApp: waiting 8m0s for pods matching "integration-test=busybox" in namespace "default" ...
helpers_test.go:344: "busybox" [a15d6e01-51a0-4885-832f-2bd74e28a9ee] Pending / Ready:ContainersNotReady (containers with unready status: [busybox]) / ContainersReady:ContainersNotReady (containers with unready status: [busybox])
E0913 19:13:54.401256   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:54.407667   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:54.419055   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:54.440362   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:54.481735   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:54.563177   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:54.724588   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:55.046390   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:55.688194   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
helpers_test.go:344: "busybox" [a15d6e01-51a0-4885-832f-2bd74e28a9ee] Running
E0913 19:13:56.970263   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:57.814343   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:13:59.532132   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:196: (dbg) TestStartStop/group/embed-certs/serial/DeployApp: integration-test=busybox healthy within 9.004115488s
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context embed-certs-556241 exec busybox -- /bin/sh -c "ulimit -n"
--- PASS: TestStartStop/group/embed-certs/serial/DeployApp (9.25s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/EnableAddonWhileActive (0.82s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/EnableAddonWhileActive
start_stop_delete_test.go:205: (dbg) Run:  out/minikube-linux-amd64 addons enable metrics-server -p embed-certs-556241 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain
start_stop_delete_test.go:215: (dbg) Run:  kubectl --context embed-certs-556241 describe deploy/metrics-server -n kube-system
--- PASS: TestStartStop/group/embed-certs/serial/EnableAddonWhileActive (0.82s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/Stop (10.77s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/Stop
start_stop_delete_test.go:228: (dbg) Run:  out/minikube-linux-amd64 stop -p embed-certs-556241 --alsologtostderr -v=3
E0913 19:14:03.998551   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:04.653630   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:06.851513   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:228: (dbg) Done: out/minikube-linux-amd64 stop -p embed-certs-556241 --alsologtostderr -v=3: (10.77433087s)
--- PASS: TestStartStop/group/embed-certs/serial/Stop (10.77s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/DeployApp (9.24s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/DeployApp
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context default-k8s-diff-port-315900 create -f testdata/busybox.yaml
start_stop_delete_test.go:196: (dbg) TestStartStop/group/default-k8s-diff-port/serial/DeployApp: waiting 8m0s for pods matching "integration-test=busybox" in namespace "default" ...
helpers_test.go:344: "busybox" [992bd382-bd07-43fc-b209-fbcec086fc75] Pending / Ready:ContainersNotReady (containers with unready status: [busybox]) / ContainersReady:ContainersNotReady (containers with unready status: [busybox])
helpers_test.go:344: "busybox" [992bd382-bd07-43fc-b209-fbcec086fc75] Running
E0913 19:14:14.895806   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:196: (dbg) TestStartStop/group/default-k8s-diff-port/serial/DeployApp: integration-test=busybox healthy within 9.004520871s
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context default-k8s-diff-port-315900 exec busybox -- /bin/sh -c "ulimit -n"
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/DeployApp (9.24s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/EnableAddonAfterStop (0.15s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/EnableAddonAfterStop
start_stop_delete_test.go:239: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Host}} -p embed-certs-556241 -n embed-certs-556241
start_stop_delete_test.go:239: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.Host}} -p embed-certs-556241 -n embed-certs-556241: exit status 7 (61.540663ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:239: status error: exit status 7 (may be ok)
start_stop_delete_test.go:246: (dbg) Run:  out/minikube-linux-amd64 addons enable dashboard -p embed-certs-556241 --images=MetricsScraper=registry.k8s.io/echoserver:1.4
--- PASS: TestStartStop/group/embed-certs/serial/EnableAddonAfterStop (0.15s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/SecondStart (262.1s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/SecondStart
start_stop_delete_test.go:256: (dbg) Run:  out/minikube-linux-amd64 start -p embed-certs-556241 --memory=2200 --alsologtostderr --wait=true --embed-certs --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1
start_stop_delete_test.go:256: (dbg) Done: out/minikube-linux-amd64 start -p embed-certs-556241 --memory=2200 --alsologtostderr --wait=true --embed-certs --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1: (4m21.811325898s)
start_stop_delete_test.go:262: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Host}} -p embed-certs-556241 -n embed-certs-556241
--- PASS: TestStartStop/group/embed-certs/serial/SecondStart (262.10s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/DeployApp (8.36s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/DeployApp
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context old-k8s-version-963966 create -f testdata/busybox.yaml
start_stop_delete_test.go:196: (dbg) TestStartStop/group/old-k8s-version/serial/DeployApp: waiting 8m0s for pods matching "integration-test=busybox" in namespace "default" ...
helpers_test.go:344: "busybox" [2979872e-9528-4fb2-ae15-c2f598863908] Pending / Ready:ContainersNotReady (containers with unready status: [busybox]) / ContainersReady:ContainersNotReady (containers with unready status: [busybox])
E0913 19:14:18.295999   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
helpers_test.go:344: "busybox" [2979872e-9528-4fb2-ae15-c2f598863908] Running
start_stop_delete_test.go:196: (dbg) TestStartStop/group/old-k8s-version/serial/DeployApp: integration-test=busybox healthy within 8.004101278s
start_stop_delete_test.go:196: (dbg) Run:  kubectl --context old-k8s-version-963966 exec busybox -- /bin/sh -c "ulimit -n"
--- PASS: TestStartStop/group/old-k8s-version/serial/DeployApp (8.36s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/EnableAddonWhileActive (0.93s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/EnableAddonWhileActive
start_stop_delete_test.go:205: (dbg) Run:  out/minikube-linux-amd64 addons enable metrics-server -p default-k8s-diff-port-315900 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain
start_stop_delete_test.go:215: (dbg) Run:  kubectl --context default-k8s-diff-port-315900 describe deploy/metrics-server -n kube-system
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/EnableAddonWhileActive (0.93s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/Stop (10.91s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/Stop
start_stop_delete_test.go:228: (dbg) Run:  out/minikube-linux-amd64 stop -p default-k8s-diff-port-315900 --alsologtostderr -v=3
start_stop_delete_test.go:228: (dbg) Done: out/minikube-linux-amd64 stop -p default-k8s-diff-port-315900 --alsologtostderr -v=3: (10.912647553s)
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/Stop (10.91s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/EnableAddonWhileActive (0.8s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/EnableAddonWhileActive
start_stop_delete_test.go:205: (dbg) Run:  out/minikube-linux-amd64 addons enable metrics-server -p old-k8s-version-963966 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain
start_stop_delete_test.go:215: (dbg) Run:  kubectl --context old-k8s-version-963966 describe deploy/metrics-server -n kube-system
--- PASS: TestStartStop/group/old-k8s-version/serial/EnableAddonWhileActive (0.80s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/Stop (10.8s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/Stop
start_stop_delete_test.go:228: (dbg) Run:  out/minikube-linux-amd64 stop -p old-k8s-version-963966 --alsologtostderr -v=3
start_stop_delete_test.go:228: (dbg) Done: out/minikube-linux-amd64 stop -p old-k8s-version-963966 --alsologtostderr -v=3: (10.803483061s)
--- PASS: TestStartStop/group/old-k8s-version/serial/Stop (10.80s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/EnableAddonAfterStop (0.22s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/EnableAddonAfterStop
start_stop_delete_test.go:239: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Host}} -p default-k8s-diff-port-315900 -n default-k8s-diff-port-315900
start_stop_delete_test.go:239: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.Host}} -p default-k8s-diff-port-315900 -n default-k8s-diff-port-315900: exit status 7 (110.402796ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:239: status error: exit status 7 (may be ok)
start_stop_delete_test.go:246: (dbg) Run:  out/minikube-linux-amd64 addons enable dashboard -p default-k8s-diff-port-315900 --images=MetricsScraper=registry.k8s.io/echoserver:1.4
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/EnableAddonAfterStop (0.22s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/SecondStart (263.3s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/SecondStart
start_stop_delete_test.go:256: (dbg) Run:  out/minikube-linux-amd64 start -p default-k8s-diff-port-315900 --memory=2200 --alsologtostderr --wait=true --apiserver-port=8444 --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1
E0913 19:14:35.377850   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:256: (dbg) Done: out/minikube-linux-amd64 start -p default-k8s-diff-port-315900 --memory=2200 --alsologtostderr --wait=true --apiserver-port=8444 --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1: (4m23.029034661s)
start_stop_delete_test.go:262: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Host}} -p default-k8s-diff-port-315900 -n default-k8s-diff-port-315900
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/SecondStart (263.30s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/EnableAddonAfterStop (0.16s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/EnableAddonAfterStop
start_stop_delete_test.go:239: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Host}} -p old-k8s-version-963966 -n old-k8s-version-963966
start_stop_delete_test.go:239: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.Host}} -p old-k8s-version-963966 -n old-k8s-version-963966: exit status 7 (60.103369ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:239: status error: exit status 7 (may be ok)
start_stop_delete_test.go:246: (dbg) Run:  out/minikube-linux-amd64 addons enable dashboard -p old-k8s-version-963966 --images=MetricsScraper=registry.k8s.io/echoserver:1.4
--- PASS: TestStartStop/group/old-k8s-version/serial/EnableAddonAfterStop (0.16s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/SecondStart (119.41s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/SecondStart
start_stop_delete_test.go:256: (dbg) Run:  out/minikube-linux-amd64 start -p old-k8s-version-963966 --memory=2200 --alsologtostderr --wait=true --kvm-network=default --kvm-qemu-uri=qemu:///system --disable-driver-mounts --keep-context=false --driver=docker  --container-runtime=docker --kubernetes-version=v1.20.0
E0913 19:14:48.822967   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:48.829352   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:48.840691   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:48.862134   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:48.903963   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:48.985784   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:49.147331   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:49.468871   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:50.111069   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:51.392990   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:53.324145   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:53.330498   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:53.341826   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:53.363138   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:53.404544   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:53.485971   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:53.647378   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:53.954991   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:53.969342   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:54.610948   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:55.892633   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:58.454145   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:59.077267   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:14:59.257882   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:03.575873   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:09.318583   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:13.817454   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:14.604583   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:16.339214   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:29.800904   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:34.298788   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:36.458377   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:36.464736   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:36.476061   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:36.497410   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:36.538813   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:36.620232   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:36.781767   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:37.103452   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:37.745608   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:39.027661   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:41.589632   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:46.711981   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:15:56.953525   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:03.115110   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:03.121457   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:03.132792   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:03.154161   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:03.196407   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:03.277705   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:03.439850   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:03.761627   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:04.403253   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:05.685456   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:08.247605   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:08.492629   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:08.498914   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:08.510233   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:08.531543   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:08.572894   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:08.654288   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:08.815786   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:09.137743   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:09.779963   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:10.762910   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:11.062157   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:13.369088   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:13.624448   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:15.260854   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:17.435520   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:18.746701   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:21.179713   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:23.610460   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:16:28.988151   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:256: (dbg) Done: out/minikube-linux-amd64 start -p old-k8s-version-963966 --memory=2200 --alsologtostderr --wait=true --kvm-network=default --kvm-qemu-uri=qemu:///system --disable-driver-mounts --keep-context=false --driver=docker  --container-runtime=docker --kubernetes-version=v1.20.0: (1m59.131028553s)
start_stop_delete_test.go:262: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Host}} -p old-k8s-version-963966 -n old-k8s-version-963966
--- PASS: TestStartStop/group/old-k8s-version/serial/SecondStart (119.41s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/UserAppExistsAfterStop (6.01s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/UserAppExistsAfterStop
start_stop_delete_test.go:274: (dbg) TestStartStop/group/old-k8s-version/serial/UserAppExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-cd95d586-n26kp" [92b4ed94-037f-47fd-b1d2-4ba629a1e47c] Running
E0913 19:16:38.261263   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kindnet-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:274: (dbg) TestStartStop/group/old-k8s-version/serial/UserAppExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 6.003562711s
--- PASS: TestStartStop/group/old-k8s-version/serial/UserAppExistsAfterStop (6.01s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/AddonExistsAfterStop (5.07s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/AddonExistsAfterStop
start_stop_delete_test.go:287: (dbg) TestStartStop/group/old-k8s-version/serial/AddonExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-cd95d586-n26kp" [92b4ed94-037f-47fd-b1d2-4ba629a1e47c] Running
E0913 19:16:44.091808   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:287: (dbg) TestStartStop/group/old-k8s-version/serial/AddonExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 5.00300518s
start_stop_delete_test.go:291: (dbg) Run:  kubectl --context old-k8s-version-963966 describe deploy/dashboard-metrics-scraper -n kubernetes-dashboard
--- PASS: TestStartStop/group/old-k8s-version/serial/AddonExistsAfterStop (5.07s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/VerifyKubernetesImages (0.19s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/VerifyKubernetesImages
start_stop_delete_test.go:304: (dbg) Run:  out/minikube-linux-amd64 -p old-k8s-version-963966 image list --format=json
start_stop_delete_test.go:304: Found non-minikube image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
--- PASS: TestStartStop/group/old-k8s-version/serial/VerifyKubernetesImages (0.19s)

                                                
                                    
x
+
TestStartStop/group/old-k8s-version/serial/Pause (2.3s)

                                                
                                                
=== RUN   TestStartStop/group/old-k8s-version/serial/Pause
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 pause -p old-k8s-version-963966 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.APIServer}} -p old-k8s-version-963966 -n old-k8s-version-963966
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.APIServer}} -p old-k8s-version-963966 -n old-k8s-version-963966: exit status 2 (270.696121ms)

                                                
                                                
-- stdout --
	Paused

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Kubelet}} -p old-k8s-version-963966 -n old-k8s-version-963966
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.Kubelet}} -p old-k8s-version-963966 -n old-k8s-version-963966: exit status 2 (274.775578ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 unpause -p old-k8s-version-963966 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.APIServer}} -p old-k8s-version-963966 -n old-k8s-version-963966
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Kubelet}} -p old-k8s-version-963966 -n old-k8s-version-963966
--- PASS: TestStartStop/group/old-k8s-version/serial/Pause (2.30s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/FirstStart (29.21s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/FirstStart
start_stop_delete_test.go:186: (dbg) Run:  out/minikube-linux-amd64 start -p newest-cni-719781 --memory=2200 --alsologtostderr --wait=apiserver,system_pods,default_sa --feature-gates ServerSideApply=true --network-plugin=cni --extra-config=kubeadm.pod-network-cidr=10.42.0.0/16 --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1
E0913 19:16:58.397256   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/false-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:11.538989   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/functional-463208/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:13.228958   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:13.235381   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:13.246795   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:13.268219   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:13.309742   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:13.391370   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:13.552882   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:13.874597   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:14.516677   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:15.798414   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:18.360610   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:186: (dbg) Done: out/minikube-linux-amd64 start -p newest-cni-719781 --memory=2200 --alsologtostderr --wait=apiserver,system_pods,default_sa --feature-gates ServerSideApply=true --network-plugin=cni --extra-config=kubeadm.pod-network-cidr=10.42.0.0/16 --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1: (29.209352643s)
--- PASS: TestStartStop/group/newest-cni/serial/FirstStart (29.21s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/DeployApp (0s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/DeployApp
--- PASS: TestStartStop/group/newest-cni/serial/DeployApp (0.00s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/EnableAddonWhileActive (1.08s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/EnableAddonWhileActive
start_stop_delete_test.go:205: (dbg) Run:  out/minikube-linux-amd64 addons enable metrics-server -p newest-cni-719781 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain
start_stop_delete_test.go:205: (dbg) Done: out/minikube-linux-amd64 addons enable metrics-server -p newest-cni-719781 --images=MetricsServer=registry.k8s.io/echoserver:1.4 --registries=MetricsServer=fake.domain: (1.081139727s)
start_stop_delete_test.go:211: WARNING: cni mode requires additional setup before pods can schedule :(
--- PASS: TestStartStop/group/newest-cni/serial/EnableAddonWhileActive (1.08s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/Stop (10.78s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/Stop
start_stop_delete_test.go:228: (dbg) Run:  out/minikube-linux-amd64 stop -p newest-cni-719781 --alsologtostderr -v=3
E0913 19:17:23.482904   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:25.053361   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:30.431854   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:228: (dbg) Done: out/minikube-linux-amd64 stop -p newest-cni-719781 --alsologtostderr -v=3: (10.776428362s)
--- PASS: TestStartStop/group/newest-cni/serial/Stop (10.78s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/EnableAddonAfterStop (0.16s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/EnableAddonAfterStop
start_stop_delete_test.go:239: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Host}} -p newest-cni-719781 -n newest-cni-719781
start_stop_delete_test.go:239: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.Host}} -p newest-cni-719781 -n newest-cni-719781: exit status 7 (62.171648ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:239: status error: exit status 7 (may be ok)
start_stop_delete_test.go:246: (dbg) Run:  out/minikube-linux-amd64 addons enable dashboard -p newest-cni-719781 --images=MetricsScraper=registry.k8s.io/echoserver:1.4
E0913 19:17:32.684813   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/calico-997126/client.crt: no such file or directory" logger="UnhandledError"
--- PASS: TestStartStop/group/newest-cni/serial/EnableAddonAfterStop (0.16s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/SecondStart (14.63s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/SecondStart
start_stop_delete_test.go:256: (dbg) Run:  out/minikube-linux-amd64 start -p newest-cni-719781 --memory=2200 --alsologtostderr --wait=apiserver,system_pods,default_sa --feature-gates ServerSideApply=true --network-plugin=cni --extra-config=kubeadm.pod-network-cidr=10.42.0.0/16 --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1
E0913 19:17:33.725207   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:37.182910   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/custom-flannel-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:38.167821   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:38.174297   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:38.186576   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:38.208298   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:38.249781   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:38.331601   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:38.494010   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:38.816065   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:39.457937   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:40.740138   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:17:43.301636   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:256: (dbg) Done: out/minikube-linux-amd64 start -p newest-cni-719781 --memory=2200 --alsologtostderr --wait=apiserver,system_pods,default_sa --feature-gates ServerSideApply=true --network-plugin=cni --extra-config=kubeadm.pod-network-cidr=10.42.0.0/16 --driver=docker  --container-runtime=docker --kubernetes-version=v1.31.1: (14.258598302s)
start_stop_delete_test.go:262: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Host}} -p newest-cni-719781 -n newest-cni-719781
--- PASS: TestStartStop/group/newest-cni/serial/SecondStart (14.63s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/UserAppExistsAfterStop (0s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/UserAppExistsAfterStop
start_stop_delete_test.go:273: WARNING: cni mode requires additional setup before pods can schedule :(
--- PASS: TestStartStop/group/newest-cni/serial/UserAppExistsAfterStop (0.00s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/AddonExistsAfterStop (0s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/AddonExistsAfterStop
start_stop_delete_test.go:284: WARNING: cni mode requires additional setup before pods can schedule :(
--- PASS: TestStartStop/group/newest-cni/serial/AddonExistsAfterStop (0.00s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/VerifyKubernetesImages (0.21s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/VerifyKubernetesImages
start_stop_delete_test.go:304: (dbg) Run:  out/minikube-linux-amd64 -p newest-cni-719781 image list --format=json
--- PASS: TestStartStop/group/newest-cni/serial/VerifyKubernetesImages (0.21s)

                                                
                                    
x
+
TestStartStop/group/newest-cni/serial/Pause (2.48s)

                                                
                                                
=== RUN   TestStartStop/group/newest-cni/serial/Pause
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 pause -p newest-cni-719781 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.APIServer}} -p newest-cni-719781 -n newest-cni-719781
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.APIServer}} -p newest-cni-719781 -n newest-cni-719781: exit status 2 (296.302881ms)

                                                
                                                
-- stdout --
	Paused

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Kubelet}} -p newest-cni-719781 -n newest-cni-719781
E0913 19:17:48.423919   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.Kubelet}} -p newest-cni-719781 -n newest-cni-719781: exit status 2 (290.876584ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 unpause -p newest-cni-719781 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.APIServer}} -p newest-cni-719781 -n newest-cni-719781
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Kubelet}} -p newest-cni-719781 -n newest-cni-719781
--- PASS: TestStartStop/group/newest-cni/serial/Pause (2.48s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/UserAppExistsAfterStop (6.01s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/UserAppExistsAfterStop
start_stop_delete_test.go:274: (dbg) TestStartStop/group/no-preload/serial/UserAppExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-695b96c756-jfn2b" [3fdc55af-c621-43ab-a8df-4fdaa07ba4a6] Running
E0913 19:17:58.666114   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:274: (dbg) TestStartStop/group/no-preload/serial/UserAppExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 6.004496794s
--- PASS: TestStartStop/group/no-preload/serial/UserAppExistsAfterStop (6.01s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/AddonExistsAfterStop (5.07s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/AddonExistsAfterStop
start_stop_delete_test.go:287: (dbg) TestStartStop/group/no-preload/serial/AddonExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-695b96c756-jfn2b" [3fdc55af-c621-43ab-a8df-4fdaa07ba4a6] Running
start_stop_delete_test.go:287: (dbg) TestStartStop/group/no-preload/serial/AddonExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 5.003639525s
start_stop_delete_test.go:291: (dbg) Run:  kubectl --context no-preload-230953 describe deploy/dashboard-metrics-scraper -n kubernetes-dashboard
--- PASS: TestStartStop/group/no-preload/serial/AddonExistsAfterStop (5.07s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/VerifyKubernetesImages (0.19s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/VerifyKubernetesImages
start_stop_delete_test.go:304: (dbg) Run:  out/minikube-linux-amd64 -p no-preload-230953 image list --format=json
start_stop_delete_test.go:304: Found non-minikube image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
--- PASS: TestStartStop/group/no-preload/serial/VerifyKubernetesImages (0.19s)

                                                
                                    
x
+
TestStartStop/group/no-preload/serial/Pause (2.24s)

                                                
                                                
=== RUN   TestStartStop/group/no-preload/serial/Pause
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 pause -p no-preload-230953 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.APIServer}} -p no-preload-230953 -n no-preload-230953
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.APIServer}} -p no-preload-230953 -n no-preload-230953: exit status 2 (268.939505ms)

                                                
                                                
-- stdout --
	Paused

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Kubelet}} -p no-preload-230953 -n no-preload-230953
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.Kubelet}} -p no-preload-230953 -n no-preload-230953: exit status 2 (265.145351ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 unpause -p no-preload-230953 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.APIServer}} -p no-preload-230953 -n no-preload-230953
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Kubelet}} -p no-preload-230953 -n no-preload-230953
--- PASS: TestStartStop/group/no-preload/serial/Pause (2.24s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/UserAppExistsAfterStop (6s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/UserAppExistsAfterStop
start_stop_delete_test.go:274: (dbg) TestStartStop/group/embed-certs/serial/UserAppExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-695b96c756-qh8gz" [96cd82e2-176d-4d06-9d19-bab9ccf4eaa5] Running
E0913 19:18:35.168952   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/bridge-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:18:37.320383   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:18:39.147591   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/skaffold-408558/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:274: (dbg) TestStartStop/group/embed-certs/serial/UserAppExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 6.00369257s
--- PASS: TestStartStop/group/embed-certs/serial/UserAppExistsAfterStop (6.00s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/AddonExistsAfterStop (5.07s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/AddonExistsAfterStop
start_stop_delete_test.go:287: (dbg) TestStartStop/group/embed-certs/serial/AddonExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-695b96c756-qh8gz" [96cd82e2-176d-4d06-9d19-bab9ccf4eaa5] Running
start_stop_delete_test.go:287: (dbg) TestStartStop/group/embed-certs/serial/AddonExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 5.004136988s
start_stop_delete_test.go:291: (dbg) Run:  kubectl --context embed-certs-556241 describe deploy/dashboard-metrics-scraper -n kubernetes-dashboard
--- PASS: TestStartStop/group/embed-certs/serial/AddonExistsAfterStop (5.07s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/VerifyKubernetesImages (0.19s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/VerifyKubernetesImages
start_stop_delete_test.go:304: (dbg) Run:  out/minikube-linux-amd64 -p embed-certs-556241 image list --format=json
start_stop_delete_test.go:304: Found non-minikube image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
--- PASS: TestStartStop/group/embed-certs/serial/VerifyKubernetesImages (0.19s)

                                                
                                    
x
+
TestStartStop/group/embed-certs/serial/Pause (2.23s)

                                                
                                                
=== RUN   TestStartStop/group/embed-certs/serial/Pause
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 pause -p embed-certs-556241 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.APIServer}} -p embed-certs-556241 -n embed-certs-556241
E0913 19:18:46.975249   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/enable-default-cni-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.APIServer}} -p embed-certs-556241 -n embed-certs-556241: exit status 2 (264.629343ms)

                                                
                                                
-- stdout --
	Paused

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Kubelet}} -p embed-certs-556241 -n embed-certs-556241
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.Kubelet}} -p embed-certs-556241 -n embed-certs-556241: exit status 2 (271.01721ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 unpause -p embed-certs-556241 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.APIServer}} -p embed-certs-556241 -n embed-certs-556241
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Kubelet}} -p embed-certs-556241 -n embed-certs-556241
--- PASS: TestStartStop/group/embed-certs/serial/Pause (2.23s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/UserAppExistsAfterStop (6.01s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/UserAppExistsAfterStop
start_stop_delete_test.go:274: (dbg) TestStartStop/group/default-k8s-diff-port/serial/UserAppExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-695b96c756-nn5sw" [4b130516-270e-4c8e-a73e-6aeaa07dd373] Running
E0913 19:19:00.109522   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/kubenet-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:274: (dbg) TestStartStop/group/default-k8s-diff-port/serial/UserAppExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 6.004328281s
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/UserAppExistsAfterStop (6.01s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/AddonExistsAfterStop (6.07s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/AddonExistsAfterStop
start_stop_delete_test.go:287: (dbg) TestStartStop/group/default-k8s-diff-port/serial/AddonExistsAfterStop: waiting 9m0s for pods matching "k8s-app=kubernetes-dashboard" in namespace "kubernetes-dashboard" ...
helpers_test.go:344: "kubernetes-dashboard-695b96c756-nn5sw" [4b130516-270e-4c8e-a73e-6aeaa07dd373] Running
E0913 19:19:03.999437   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/addons-315216/client.crt: no such file or directory" logger="UnhandledError"
E0913 19:19:05.021263   10475 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/auto-997126/client.crt: no such file or directory" logger="UnhandledError"
start_stop_delete_test.go:287: (dbg) TestStartStop/group/default-k8s-diff-port/serial/AddonExistsAfterStop: k8s-app=kubernetes-dashboard healthy within 6.005016659s
start_stop_delete_test.go:291: (dbg) Run:  kubectl --context default-k8s-diff-port-315900 describe deploy/dashboard-metrics-scraper -n kubernetes-dashboard
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/AddonExistsAfterStop (6.07s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/VerifyKubernetesImages (0.19s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/VerifyKubernetesImages
start_stop_delete_test.go:304: (dbg) Run:  out/minikube-linux-amd64 -p default-k8s-diff-port-315900 image list --format=json
start_stop_delete_test.go:304: Found non-minikube image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/VerifyKubernetesImages (0.19s)

                                                
                                    
x
+
TestStartStop/group/default-k8s-diff-port/serial/Pause (2.21s)

                                                
                                                
=== RUN   TestStartStop/group/default-k8s-diff-port/serial/Pause
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 pause -p default-k8s-diff-port-315900 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.APIServer}} -p default-k8s-diff-port-315900 -n default-k8s-diff-port-315900
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.APIServer}} -p default-k8s-diff-port-315900 -n default-k8s-diff-port-315900: exit status 2 (262.752509ms)

                                                
                                                
-- stdout --
	Paused

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Kubelet}} -p default-k8s-diff-port-315900 -n default-k8s-diff-port-315900
start_stop_delete_test.go:311: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.Kubelet}} -p default-k8s-diff-port-315900 -n default-k8s-diff-port-315900: exit status 2 (264.878168ms)

                                                
                                                
-- stdout --
	Stopped

                                                
                                                
-- /stdout --
start_stop_delete_test.go:311: status error: exit status 2 (may be ok)
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 unpause -p default-k8s-diff-port-315900 --alsologtostderr -v=1
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.APIServer}} -p default-k8s-diff-port-315900 -n default-k8s-diff-port-315900
start_stop_delete_test.go:311: (dbg) Run:  out/minikube-linux-amd64 status --format={{.Kubelet}} -p default-k8s-diff-port-315900 -n default-k8s-diff-port-315900
--- PASS: TestStartStop/group/default-k8s-diff-port/serial/Pause (2.21s)

                                                
                                    

Test skip (20/342)

x
+
TestDownloadOnly/v1.20.0/cached-images (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.20.0/cached-images
aaa_download_only_test.go:129: Preload exists, images won't be cached
--- SKIP: TestDownloadOnly/v1.20.0/cached-images (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.20.0/binaries (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.20.0/binaries
aaa_download_only_test.go:151: Preload exists, binaries are present within.
--- SKIP: TestDownloadOnly/v1.20.0/binaries (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.20.0/kubectl (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.20.0/kubectl
aaa_download_only_test.go:167: Test for darwin and windows
--- SKIP: TestDownloadOnly/v1.20.0/kubectl (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.31.1/cached-images (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.31.1/cached-images
aaa_download_only_test.go:129: Preload exists, images won't be cached
--- SKIP: TestDownloadOnly/v1.31.1/cached-images (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.31.1/binaries (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.31.1/binaries
aaa_download_only_test.go:151: Preload exists, binaries are present within.
--- SKIP: TestDownloadOnly/v1.31.1/binaries (0.00s)

                                                
                                    
x
+
TestDownloadOnly/v1.31.1/kubectl (0s)

                                                
                                                
=== RUN   TestDownloadOnly/v1.31.1/kubectl
aaa_download_only_test.go:167: Test for darwin and windows
--- SKIP: TestDownloadOnly/v1.31.1/kubectl (0.00s)

                                                
                                    
x
+
TestAddons/parallel/Olm (0s)

                                                
                                                
=== RUN   TestAddons/parallel/Olm
=== PAUSE TestAddons/parallel/Olm

                                                
                                                

                                                
                                                
=== CONT  TestAddons/parallel/Olm
addons_test.go:438: Skipping OLM addon test until https://github.com/operator-framework/operator-lifecycle-manager/issues/2534 is resolved
--- SKIP: TestAddons/parallel/Olm (0.00s)

                                                
                                    
x
+
TestDockerEnvContainerd (0s)

                                                
                                                
=== RUN   TestDockerEnvContainerd
docker_test.go:170: running with docker true linux amd64
docker_test.go:172: skipping: TestDockerEnvContainerd can only be run with the containerd runtime on Docker driver
--- SKIP: TestDockerEnvContainerd (0.00s)

                                                
                                    
x
+
TestHyperKitDriverInstallOrUpdate (0s)

                                                
                                                
=== RUN   TestHyperKitDriverInstallOrUpdate
driver_install_or_update_test.go:105: Skip if not darwin.
--- SKIP: TestHyperKitDriverInstallOrUpdate (0.00s)

                                                
                                    
x
+
TestHyperkitDriverSkipUpgrade (0s)

                                                
                                                
=== RUN   TestHyperkitDriverSkipUpgrade
driver_install_or_update_test.go:169: Skip if not darwin.
--- SKIP: TestHyperkitDriverSkipUpgrade (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/PodmanEnv (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/PodmanEnv
=== PAUSE TestFunctional/parallel/PodmanEnv

                                                
                                                

                                                
                                                
=== CONT  TestFunctional/parallel/PodmanEnv
functional_test.go:550: only validate podman env with docker container runtime, currently testing docker
--- SKIP: TestFunctional/parallel/PodmanEnv (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/DNSResolutionByDig (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/DNSResolutionByDig
functional_test_tunnel_test.go:99: DNS forwarding is only supported for Hyperkit on Darwin, skipping test DNS forwarding
--- SKIP: TestFunctional/parallel/TunnelCmd/serial/DNSResolutionByDig (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/DNSResolutionByDscacheutil (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/DNSResolutionByDscacheutil
functional_test_tunnel_test.go:99: DNS forwarding is only supported for Hyperkit on Darwin, skipping test DNS forwarding
--- SKIP: TestFunctional/parallel/TunnelCmd/serial/DNSResolutionByDscacheutil (0.00s)

                                                
                                    
x
+
TestFunctional/parallel/TunnelCmd/serial/AccessThroughDNS (0s)

                                                
                                                
=== RUN   TestFunctional/parallel/TunnelCmd/serial/AccessThroughDNS
functional_test_tunnel_test.go:99: DNS forwarding is only supported for Hyperkit on Darwin, skipping test DNS forwarding
--- SKIP: TestFunctional/parallel/TunnelCmd/serial/AccessThroughDNS (0.00s)

                                                
                                    
x
+
TestGvisorAddon (0s)

                                                
                                                
=== RUN   TestGvisorAddon
gvisor_addon_test.go:34: skipping test because --gvisor=false
--- SKIP: TestGvisorAddon (0.00s)

                                                
                                    
x
+
TestImageBuild/serial/validateImageBuildWithBuildEnv (0s)

                                                
                                                
=== RUN   TestImageBuild/serial/validateImageBuildWithBuildEnv
image_test.go:114: skipping due to https://github.com/kubernetes/minikube/issues/12431
--- SKIP: TestImageBuild/serial/validateImageBuildWithBuildEnv (0.00s)

                                                
                                    
x
+
TestChangeNoneUser (0s)

                                                
                                                
=== RUN   TestChangeNoneUser
none_test.go:38: Test requires none driver and SUDO_USER env to not be empty
--- SKIP: TestChangeNoneUser (0.00s)

                                                
                                    
x
+
TestScheduledStopWindows (0s)

                                                
                                                
=== RUN   TestScheduledStopWindows
scheduled_stop_test.go:42: test only runs on windows
--- SKIP: TestScheduledStopWindows (0.00s)

                                                
                                    
x
+
TestNetworkPlugins/group/cilium (5.38s)

                                                
                                                
=== RUN   TestNetworkPlugins/group/cilium
net_test.go:102: Skipping the test as it's interfering with other tests and is outdated
panic.go:629: 
----------------------- debugLogs start: cilium-997126 [pass: true] --------------------------------
>>> netcat: nslookup kubernetes.default:
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> netcat: nslookup debug kubernetes.default a-records:
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> netcat: dig search kubernetes.default:
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> netcat: dig @10.96.0.10 kubernetes.default.svc.cluster.local udp/53:
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> netcat: dig @10.96.0.10 kubernetes.default.svc.cluster.local tcp/53:
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> netcat: nc 10.96.0.10 udp/53:
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> netcat: nc 10.96.0.10 tcp/53:
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> netcat: /etc/nsswitch.conf:
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> netcat: /etc/hosts:
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> netcat: /etc/resolv.conf:
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> host: /etc/nsswitch.conf:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: /etc/hosts:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: /etc/resolv.conf:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> k8s: nodes, services, endpoints, daemon sets, deployments and pods, :
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> host: crictl pods:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: crictl containers:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> k8s: describe netcat deployment:
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe netcat pod(s):
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> k8s: netcat logs:
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe coredns deployment:
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe coredns pods:
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> k8s: coredns logs:
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe api server pod(s):
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> k8s: api server logs:
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> host: /etc/cni:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: ip a s:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: ip r s:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: iptables-save:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: iptables table nat:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> k8s: describe cilium daemon set:
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> k8s: describe cilium daemon set pod(s):
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> k8s: cilium daemon set container(s) logs (current):
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> k8s: cilium daemon set container(s) logs (previous):
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe cilium deployment:
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> k8s: describe cilium deployment pod(s):
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> k8s: cilium deployment container(s) logs (current):
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> k8s: cilium deployment container(s) logs (previous):
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe kube-proxy daemon set:
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> k8s: describe kube-proxy pod(s):
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> k8s: kube-proxy logs:
error: context "cilium-997126" does not exist

                                                
                                                

                                                
                                                
>>> host: kubelet daemon status:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: kubelet daemon config:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> k8s: kubelet logs:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: /etc/kubernetes/kubelet.conf:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: /var/lib/kubelet/config.yaml:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> k8s: kubectl config:
apiVersion: v1
clusters:
- cluster:
certificate-authority: /home/jenkins/minikube-integration/19636-3724/.minikube/ca.crt
extensions:
- extension:
last-update: Fri, 13 Sep 2024 19:04:42 UTC
provider: minikube.sigs.k8s.io
version: v1.34.0
name: cluster_info
server: https://192.168.94.2:8443
name: NoKubernetes-380550
- cluster:
certificate-authority: /home/jenkins/minikube-integration/19636-3724/.minikube/ca.crt
extensions:
- extension:
last-update: Fri, 13 Sep 2024 19:04:46 UTC
provider: minikube.sigs.k8s.io
version: v1.34.0
name: cluster_info
server: https://192.168.76.2:8443
name: offline-docker-363399
contexts:
- context:
cluster: NoKubernetes-380550
extensions:
- extension:
last-update: Fri, 13 Sep 2024 19:04:42 UTC
provider: minikube.sigs.k8s.io
version: v1.34.0
name: context_info
namespace: default
user: NoKubernetes-380550
name: NoKubernetes-380550
- context:
cluster: offline-docker-363399
extensions:
- extension:
last-update: Fri, 13 Sep 2024 19:04:46 UTC
provider: minikube.sigs.k8s.io
version: v1.34.0
name: context_info
namespace: default
user: offline-docker-363399
name: offline-docker-363399
current-context: offline-docker-363399
kind: Config
preferences: {}
users:
- name: NoKubernetes-380550
user:
client-certificate: /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/NoKubernetes-380550/client.crt
client-key: /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/NoKubernetes-380550/client.key
- name: offline-docker-363399
user:
client-certificate: /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/offline-docker-363399/client.crt
client-key: /home/jenkins/minikube-integration/19636-3724/.minikube/profiles/offline-docker-363399/client.key

                                                
                                                

                                                
                                                
>>> k8s: cms:
Error in configuration: context was not found for specified context: cilium-997126

                                                
                                                

                                                
                                                
>>> host: docker daemon status:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: docker daemon config:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: /etc/docker/daemon.json:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: docker system info:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: cri-docker daemon status:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: cri-docker daemon config:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: /etc/systemd/system/cri-docker.service.d/10-cni.conf:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: /usr/lib/systemd/system/cri-docker.service:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: cri-dockerd version:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: containerd daemon status:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: containerd daemon config:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: /lib/systemd/system/containerd.service:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: /etc/containerd/config.toml:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: containerd config dump:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: crio daemon status:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: crio daemon config:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: /etc/crio:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                

                                                
                                                
>>> host: crio config:
* Profile "cilium-997126" not found. Run "minikube profile list" to view all profiles.
To start a cluster, run: "minikube start -p cilium-997126"

                                                
                                                
----------------------- debugLogs end: cilium-997126 [took: 5.210166205s] --------------------------------
helpers_test.go:175: Cleaning up "cilium-997126" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p cilium-997126
--- SKIP: TestNetworkPlugins/group/cilium (5.38s)

                                                
                                    
x
+
TestStartStop/group/disable-driver-mounts (0.14s)

                                                
                                                
=== RUN   TestStartStop/group/disable-driver-mounts
=== PAUSE TestStartStop/group/disable-driver-mounts

                                                
                                                

                                                
                                                
=== CONT  TestStartStop/group/disable-driver-mounts
start_stop_delete_test.go:103: skipping TestStartStop/group/disable-driver-mounts - only runs on virtualbox
helpers_test.go:175: Cleaning up "disable-driver-mounts-988646" profile ...
helpers_test.go:178: (dbg) Run:  out/minikube-linux-amd64 delete -p disable-driver-mounts-988646
--- SKIP: TestStartStop/group/disable-driver-mounts (0.14s)

                                                
                                    
Copied to clipboard