=== RUN TestFunctional/serial/ComponentHealth
functional_test.go:825: (dbg) Run: kubectl --context functional-866869 get po -l tier=control-plane -n kube-system -o=json
functional_test.go:840: etcd phase: Running
functional_test.go:848: etcd is not Ready: {Phase:Running Conditions:[{Type:PodReadyToStartContainers Status:True} {Type:Initialized Status:True} {Type:Ready Status:False} {Type:ContainersReady Status:False} {Type:PodScheduled Status:True}] Message: Reason: HostIP:192.168.39.233 PodIP:192.168.39.233 StartTime:2025-12-27 20:02:56 +0000 UTC ContainerStatuses:[{Name:etcd State:{Waiting:<nil> Running:0xc001f5bf50 Terminated:<nil>} LastTerminationState:{Waiting:<nil> Running:<nil> Terminated:0xc001d7f180} Ready:false RestartCount:2 Image:registry.k8s.io/etcd:3.6.6-0 ImageID:registry.k8s.io/etcd@sha256:5279f56db4f32772bb41e47ca44c553f5c87a08fdf339d74c23a4cdc3c388d6a ContainerID:cri-o://2d79b661aef28434e47e009490240e32f12ccd12aa87a23f4c8cd21f83bb358b}]}
functional_test.go:840: kube-apiserver phase: Running
functional_test.go:850: kube-apiserver status: Ready
functional_test.go:840: kube-controller-manager phase: Running
functional_test.go:850: kube-controller-manager status: Ready
functional_test.go:840: kube-scheduler phase: Running
functional_test.go:850: kube-scheduler status: Ready
helpers_test.go:223: -----------------------post-mortem--------------------------------
helpers_test.go:224: ======> post-mortem[TestFunctional/serial/ComponentHealth]: network settings <======
helpers_test.go:231: HOST ENV snapshots: PROXY env: HTTP_PROXY="<empty>" HTTPS_PROXY="<empty>" NO_PROXY="<empty>"
helpers_test.go:248: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p functional-866869 -n functional-866869
helpers_test.go:253: <<< TestFunctional/serial/ComponentHealth FAILED: start of post-mortem logs <<<
helpers_test.go:254: ======> post-mortem[TestFunctional/serial/ComponentHealth]: minikube logs <======
helpers_test.go:256: (dbg) Run: out/minikube-linux-amd64 -p functional-866869 logs -n 25
helpers_test.go:256: (dbg) Done: out/minikube-linux-amd64 -p functional-866869 logs -n 25: (1.322210369s)
helpers_test.go:261: TestFunctional/serial/ComponentHealth logs:
-- stdout --
==> Audit <==
┌─────────┬─────────────────────────────────────────────────────────────────────────────────────────────────────────────┬───────────────────┬─────────┬─────────┬─────────────────────┬─────────────────────┐
│ COMMAND │ ARGS │ PROFILE │ USER │ VERSION │ START TIME │ END TIME │
├─────────┼─────────────────────────────────────────────────────────────────────────────────────────────────────────────┼───────────────────┼─────────┼─────────┼─────────────────────┼─────────────────────┤
│ unpause │ nospam-067430 --log_dir /tmp/nospam-067430 unpause │ nospam-067430 │ jenkins │ v1.37.0 │ 27 Dec 25 19:59 UTC │ 27 Dec 25 19:59 UTC │
│ unpause │ nospam-067430 --log_dir /tmp/nospam-067430 unpause │ nospam-067430 │ jenkins │ v1.37.0 │ 27 Dec 25 19:59 UTC │ 27 Dec 25 19:59 UTC │
│ unpause │ nospam-067430 --log_dir /tmp/nospam-067430 unpause │ nospam-067430 │ jenkins │ v1.37.0 │ 27 Dec 25 19:59 UTC │ 27 Dec 25 19:59 UTC │
│ stop │ nospam-067430 --log_dir /tmp/nospam-067430 stop │ nospam-067430 │ jenkins │ v1.37.0 │ 27 Dec 25 19:59 UTC │ 27 Dec 25 20:00 UTC │
│ stop │ nospam-067430 --log_dir /tmp/nospam-067430 stop │ nospam-067430 │ jenkins │ v1.37.0 │ 27 Dec 25 20:00 UTC │ 27 Dec 25 20:00 UTC │
│ stop │ nospam-067430 --log_dir /tmp/nospam-067430 stop │ nospam-067430 │ jenkins │ v1.37.0 │ 27 Dec 25 20:00 UTC │ 27 Dec 25 20:00 UTC │
│ delete │ -p nospam-067430 │ nospam-067430 │ jenkins │ v1.37.0 │ 27 Dec 25 20:00 UTC │ 27 Dec 25 20:00 UTC │
│ start │ -p functional-866869 --memory=4096 --apiserver-port=8441 --wait=all --driver=kvm2 --container-runtime=crio │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:00 UTC │ 27 Dec 25 20:01 UTC │
│ start │ -p functional-866869 --alsologtostderr -v=8 │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:01 UTC │ 27 Dec 25 20:02 UTC │
│ cache │ functional-866869 cache add registry.k8s.io/pause:3.1 │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ cache │ functional-866869 cache add registry.k8s.io/pause:3.3 │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ cache │ functional-866869 cache add registry.k8s.io/pause:latest │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ cache │ functional-866869 cache add minikube-local-cache-test:functional-866869 │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ cache │ functional-866869 cache delete minikube-local-cache-test:functional-866869 │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ cache │ delete registry.k8s.io/pause:3.3 │ minikube │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ cache │ list │ minikube │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ ssh │ functional-866869 ssh sudo crictl images │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ ssh │ functional-866869 ssh sudo crictl rmi registry.k8s.io/pause:latest │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ ssh │ functional-866869 ssh sudo crictl inspecti registry.k8s.io/pause:latest │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ │
│ cache │ functional-866869 cache reload │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ ssh │ functional-866869 ssh sudo crictl inspecti registry.k8s.io/pause:latest │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ cache │ delete registry.k8s.io/pause:3.1 │ minikube │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ cache │ delete registry.k8s.io/pause:latest │ minikube │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ kubectl │ functional-866869 kubectl -- --context functional-866869 get pods │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:02 UTC │
│ start │ -p functional-866869 --extra-config=apiserver.enable-admission-plugins=NamespaceAutoProvision --wait=all │ functional-866869 │ jenkins │ v1.37.0 │ 27 Dec 25 20:02 UTC │ 27 Dec 25 20:03 UTC │
└─────────┴─────────────────────────────────────────────────────────────────────────────────────────────────────────────┴───────────────────┴─────────┴─────────┴─────────────────────┴─────────────────────┘
==> Last Start <==
Log file created at: 2025/12/27 20:02:34
Running on machine: ubuntu-20-agent-13
Binary: Built with gc go1.25.5 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I1227 20:02:34.156903 67071 out.go:360] Setting OutFile to fd 1 ...
I1227 20:02:34.156996 67071 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1227 20:02:34.156998 67071 out.go:374] Setting ErrFile to fd 2...
I1227 20:02:34.157001 67071 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1227 20:02:34.157233 67071 root.go:338] Updating PATH: /home/jenkins/minikube-integration/22332-59055/.minikube/bin
I1227 20:02:34.157669 67071 out.go:368] Setting JSON to false
I1227 20:02:34.158534 67071 start.go:133] hostinfo: {"hostname":"ubuntu-20-agent-13","uptime":6304,"bootTime":1766859450,"procs":177,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"22.04","kernelVersion":"6.8.0-1045-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I1227 20:02:34.158634 67071 start.go:143] virtualization: kvm guest
I1227 20:02:34.160408 67071 out.go:179] * [functional-866869] minikube v1.37.0 on Ubuntu 22.04 (kvm/amd64)
I1227 20:02:34.162199 67071 out.go:179] - MINIKUBE_LOCATION=22332
I1227 20:02:34.162223 67071 notify.go:221] Checking for updates...
I1227 20:02:34.164209 67071 out.go:179] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I1227 20:02:34.165326 67071 out.go:179] - KUBECONFIG=/home/jenkins/minikube-integration/22332-59055/kubeconfig
I1227 20:02:34.166493 67071 out.go:179] - MINIKUBE_HOME=/home/jenkins/minikube-integration/22332-59055/.minikube
I1227 20:02:34.167464 67071 out.go:179] - MINIKUBE_BIN=out/minikube-linux-amd64
I1227 20:02:34.168682 67071 out.go:179] - MINIKUBE_FORCE_SYSTEMD=
I1227 20:02:34.170291 67071 config.go:182] Loaded profile config "functional-866869": Driver=kvm2, ContainerRuntime=crio, KubernetesVersion=v1.35.0
I1227 20:02:34.170394 67071 driver.go:422] Setting default libvirt URI to qemu:///system
I1227 20:02:34.203306 67071 out.go:179] * Using the kvm2 driver based on existing profile
I1227 20:02:34.204654 67071 start.go:309] selected driver: kvm2
I1227 20:02:34.204664 67071 start.go:928] validating driver "kvm2" against &{Name:functional-866869 KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube-builds/iso/22332/minikube-v1.37.0-1766811082-22332-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1766570851-22316@sha256:7975a7a1117280f99ad7696c9c80bdca993064fe9e309e9984685e0ce989758a Memory:4096 CPUs:2 DiskSize:20000 Driver:kvm2 HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8441 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{Kuber
netesVersion:v1.35.0 ClusterName:functional-866869 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:crio CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.39.233 Port:8441 KubernetesVersion:v1.35.0 ContainerRuntime:crio ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 M
ountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s Rosetta:false}
I1227 20:02:34.204787 67071 start.go:939] status for kvm2: {Installed:true Healthy:true Running:true NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I1227 20:02:34.205664 67071 start_flags.go:1019] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I1227 20:02:34.205689 67071 cni.go:84] Creating CNI manager for ""
I1227 20:02:34.205763 67071 cni.go:146] "kvm2" driver + "crio" runtime found, recommending bridge
I1227 20:02:34.205810 67071 start.go:353] cluster config:
{Name:functional-866869 KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube-builds/iso/22332/minikube-v1.37.0-1766811082-22332-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1766570851-22316@sha256:7975a7a1117280f99ad7696c9c80bdca993064fe9e309e9984685e0ce989758a Memory:4096 CPUs:2 DiskSize:20000 Driver:kvm2 HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8441 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.35.0 ClusterName:functional-866869 Namespace:default APIServer
HAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:crio CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[{Component:apiserver Key:enable-admission-plugins Value:NamespaceAutoProvision}] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.39.233 Port:8441 KubernetesVersion:v1.35.0 ContainerRuntime:crio ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144
MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s Rosetta:false}
I1227 20:02:34.205891 67071 iso.go:125] acquiring lock: {Name:mka43d70ce37123bef7d956775bb3b0726c5ddc8 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I1227 20:02:34.207500 67071 out.go:179] * Starting "functional-866869" primary control-plane node in "functional-866869" cluster
I1227 20:02:34.208594 67071 preload.go:188] Checking if preload exists for k8s version v1.35.0 and runtime crio
I1227 20:02:34.208623 67071 preload.go:203] Found local preload: /home/jenkins/minikube-integration/22332-59055/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.35.0-cri-o-overlay-amd64.tar.lz4
I1227 20:02:34.208630 67071 cache.go:65] Caching tarball of preloaded images
I1227 20:02:34.208792 67071 preload.go:251] Found /home/jenkins/minikube-integration/22332-59055/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.35.0-cri-o-overlay-amd64.tar.lz4 in cache, skipping download
I1227 20:02:34.208800 67071 cache.go:68] Finished verifying existence of preloaded tar for v1.35.0 on crio
I1227 20:02:34.208885 67071 profile.go:143] Saving config to /home/jenkins/minikube-integration/22332-59055/.minikube/profiles/functional-866869/config.json ...
I1227 20:02:34.209083 67071 start.go:360] acquireMachinesLock for functional-866869: {Name:mka9931fb06a62e71d190bf45bd86894fc3ea87e Clock:{} Delay:500ms Timeout:13m0s Cancel:<nil>}
I1227 20:02:34.209122 67071 start.go:364] duration metric: took 28.137µs to acquireMachinesLock for "functional-866869"
I1227 20:02:34.209137 67071 start.go:96] Skipping create...Using existing machine configuration
I1227 20:02:34.209141 67071 fix.go:54] fixHost starting:
I1227 20:02:34.210853 67071 fix.go:112] recreateIfNeeded on functional-866869: state=Running err=<nil>
W1227 20:02:34.210876 67071 fix.go:138] unexpected machine state, will restart: <nil>
I1227 20:02:34.212436 67071 out.go:252] * Updating the running kvm2 "functional-866869" VM ...
I1227 20:02:34.212455 67071 machine.go:94] provisionDockerMachine start ...
I1227 20:02:34.215124 67071 main.go:144] libmachine: domain functional-866869 has defined MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.215589 67071 main.go:144] libmachine: found host DHCP lease matching {name: "", mac: "52:54:00:61:ec:a4", ip: ""} in network mk-functional-866869: {Iface:virbr1 ExpiryTime:2025-12-27 21:00:42 +0000 UTC Type:0 Mac:52:54:00:61:ec:a4 Iaid: IPaddr:192.168.39.233 Prefix:24 Hostname:functional-866869 Clientid:01:52:54:00:61:ec:a4}
I1227 20:02:34.215608 67071 main.go:144] libmachine: domain functional-866869 has defined IP address 192.168.39.233 and MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.215810 67071 main.go:144] libmachine: Using SSH client type: native
I1227 20:02:34.216047 67071 main.go:144] libmachine: &{{{<nil> 0 [] [] []} docker [0x84e300] 0x850fa0 <nil> [] 0s} 192.168.39.233 22 <nil> <nil>}
I1227 20:02:34.216052 67071 main.go:144] libmachine: About to run SSH command:
hostname
I1227 20:02:34.327023 67071 main.go:144] libmachine: SSH cmd err, output: <nil>: functional-866869
I1227 20:02:34.327043 67071 buildroot.go:166] provisioning hostname "functional-866869"
I1227 20:02:34.329774 67071 main.go:144] libmachine: domain functional-866869 has defined MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.330145 67071 main.go:144] libmachine: found host DHCP lease matching {name: "", mac: "52:54:00:61:ec:a4", ip: ""} in network mk-functional-866869: {Iface:virbr1 ExpiryTime:2025-12-27 21:00:42 +0000 UTC Type:0 Mac:52:54:00:61:ec:a4 Iaid: IPaddr:192.168.39.233 Prefix:24 Hostname:functional-866869 Clientid:01:52:54:00:61:ec:a4}
I1227 20:02:34.330167 67071 main.go:144] libmachine: domain functional-866869 has defined IP address 192.168.39.233 and MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.330399 67071 main.go:144] libmachine: Using SSH client type: native
I1227 20:02:34.330600 67071 main.go:144] libmachine: &{{{<nil> 0 [] [] []} docker [0x84e300] 0x850fa0 <nil> [] 0s} 192.168.39.233 22 <nil> <nil>}
I1227 20:02:34.330606 67071 main.go:144] libmachine: About to run SSH command:
sudo hostname functional-866869 && echo "functional-866869" | sudo tee /etc/hostname
I1227 20:02:34.459572 67071 main.go:144] libmachine: SSH cmd err, output: <nil>: functional-866869
I1227 20:02:34.462412 67071 main.go:144] libmachine: domain functional-866869 has defined MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.462839 67071 main.go:144] libmachine: found host DHCP lease matching {name: "", mac: "52:54:00:61:ec:a4", ip: ""} in network mk-functional-866869: {Iface:virbr1 ExpiryTime:2025-12-27 21:00:42 +0000 UTC Type:0 Mac:52:54:00:61:ec:a4 Iaid: IPaddr:192.168.39.233 Prefix:24 Hostname:functional-866869 Clientid:01:52:54:00:61:ec:a4}
I1227 20:02:34.462867 67071 main.go:144] libmachine: domain functional-866869 has defined IP address 192.168.39.233 and MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.463045 67071 main.go:144] libmachine: Using SSH client type: native
I1227 20:02:34.463265 67071 main.go:144] libmachine: &{{{<nil> 0 [] [] []} docker [0x84e300] 0x850fa0 <nil> [] 0s} 192.168.39.233 22 <nil> <nil>}
I1227 20:02:34.463275 67071 main.go:144] libmachine: About to run SSH command:
if ! grep -xq '.*\sfunctional-866869' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 functional-866869/g' /etc/hosts;
else
echo '127.0.1.1 functional-866869' | sudo tee -a /etc/hosts;
fi
fi
I1227 20:02:34.577329 67071 main.go:144] libmachine: SSH cmd err, output: <nil>:
I1227 20:02:34.577362 67071 buildroot.go:172] set auth options {CertDir:/home/jenkins/minikube-integration/22332-59055/.minikube CaCertPath:/home/jenkins/minikube-integration/22332-59055/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/22332-59055/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/22332-59055/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/22332-59055/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/22332-59055/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/22332-59055/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/22332-59055/.minikube}
I1227 20:02:34.577387 67071 buildroot.go:174] setting up certificates
I1227 20:02:34.577402 67071 provision.go:84] configureAuth start
I1227 20:02:34.580888 67071 main.go:144] libmachine: domain functional-866869 has defined MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.581359 67071 main.go:144] libmachine: found host DHCP lease matching {name: "", mac: "52:54:00:61:ec:a4", ip: ""} in network mk-functional-866869: {Iface:virbr1 ExpiryTime:2025-12-27 21:00:42 +0000 UTC Type:0 Mac:52:54:00:61:ec:a4 Iaid: IPaddr:192.168.39.233 Prefix:24 Hostname:functional-866869 Clientid:01:52:54:00:61:ec:a4}
I1227 20:02:34.581377 67071 main.go:144] libmachine: domain functional-866869 has defined IP address 192.168.39.233 and MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.583968 67071 main.go:144] libmachine: domain functional-866869 has defined MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.584357 67071 main.go:144] libmachine: found host DHCP lease matching {name: "", mac: "52:54:00:61:ec:a4", ip: ""} in network mk-functional-866869: {Iface:virbr1 ExpiryTime:2025-12-27 21:00:42 +0000 UTC Type:0 Mac:52:54:00:61:ec:a4 Iaid: IPaddr:192.168.39.233 Prefix:24 Hostname:functional-866869 Clientid:01:52:54:00:61:ec:a4}
I1227 20:02:34.584372 67071 main.go:144] libmachine: domain functional-866869 has defined IP address 192.168.39.233 and MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.584516 67071 provision.go:143] copyHostCerts
I1227 20:02:34.584573 67071 exec_runner.go:144] found /home/jenkins/minikube-integration/22332-59055/.minikube/ca.pem, removing ...
I1227 20:02:34.584593 67071 exec_runner.go:203] rm: /home/jenkins/minikube-integration/22332-59055/.minikube/ca.pem
I1227 20:02:34.584682 67071 exec_runner.go:151] cp: /home/jenkins/minikube-integration/22332-59055/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/22332-59055/.minikube/ca.pem (1078 bytes)
I1227 20:02:34.584824 67071 exec_runner.go:144] found /home/jenkins/minikube-integration/22332-59055/.minikube/cert.pem, removing ...
I1227 20:02:34.584830 67071 exec_runner.go:203] rm: /home/jenkins/minikube-integration/22332-59055/.minikube/cert.pem
I1227 20:02:34.584860 67071 exec_runner.go:151] cp: /home/jenkins/minikube-integration/22332-59055/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/22332-59055/.minikube/cert.pem (1123 bytes)
I1227 20:02:34.584921 67071 exec_runner.go:144] found /home/jenkins/minikube-integration/22332-59055/.minikube/key.pem, removing ...
I1227 20:02:34.584924 67071 exec_runner.go:203] rm: /home/jenkins/minikube-integration/22332-59055/.minikube/key.pem
I1227 20:02:34.584946 67071 exec_runner.go:151] cp: /home/jenkins/minikube-integration/22332-59055/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/22332-59055/.minikube/key.pem (1679 bytes)
I1227 20:02:34.585003 67071 provision.go:117] generating server cert: /home/jenkins/minikube-integration/22332-59055/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/22332-59055/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/22332-59055/.minikube/certs/ca-key.pem org=jenkins.functional-866869 san=[127.0.0.1 192.168.39.233 functional-866869 localhost minikube]
I1227 20:02:34.669118 67071 provision.go:177] copyRemoteCerts
I1227 20:02:34.669181 67071 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I1227 20:02:34.672044 67071 main.go:144] libmachine: domain functional-866869 has defined MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.672449 67071 main.go:144] libmachine: found host DHCP lease matching {name: "", mac: "52:54:00:61:ec:a4", ip: ""} in network mk-functional-866869: {Iface:virbr1 ExpiryTime:2025-12-27 21:00:42 +0000 UTC Type:0 Mac:52:54:00:61:ec:a4 Iaid: IPaddr:192.168.39.233 Prefix:24 Hostname:functional-866869 Clientid:01:52:54:00:61:ec:a4}
I1227 20:02:34.672467 67071 main.go:144] libmachine: domain functional-866869 has defined IP address 192.168.39.233 and MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.672636 67071 sshutil.go:53] new ssh client: &{IP:192.168.39.233 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/22332-59055/.minikube/machines/functional-866869/id_rsa Username:docker}
I1227 20:02:34.761346 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I1227 20:02:34.795607 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/machines/server.pem --> /etc/docker/server.pem (1220 bytes)
I1227 20:02:34.828674 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
I1227 20:02:34.861087 67071 provision.go:87] duration metric: took 283.6692ms to configureAuth
I1227 20:02:34.861114 67071 buildroot.go:189] setting minikube options for container-runtime
I1227 20:02:34.861341 67071 config.go:182] Loaded profile config "functional-866869": Driver=kvm2, ContainerRuntime=crio, KubernetesVersion=v1.35.0
I1227 20:02:34.864648 67071 main.go:144] libmachine: domain functional-866869 has defined MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.865159 67071 main.go:144] libmachine: found host DHCP lease matching {name: "", mac: "52:54:00:61:ec:a4", ip: ""} in network mk-functional-866869: {Iface:virbr1 ExpiryTime:2025-12-27 21:00:42 +0000 UTC Type:0 Mac:52:54:00:61:ec:a4 Iaid: IPaddr:192.168.39.233 Prefix:24 Hostname:functional-866869 Clientid:01:52:54:00:61:ec:a4}
I1227 20:02:34.865185 67071 main.go:144] libmachine: domain functional-866869 has defined IP address 192.168.39.233 and MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:34.865425 67071 main.go:144] libmachine: Using SSH client type: native
I1227 20:02:34.865646 67071 main.go:144] libmachine: &{{{<nil> 0 [] [] []} docker [0x84e300] 0x850fa0 <nil> [] 0s} 192.168.39.233 22 <nil> <nil>}
I1227 20:02:34.865655 67071 main.go:144] libmachine: About to run SSH command:
sudo mkdir -p /etc/sysconfig && printf %s "
CRIO_MINIKUBE_OPTIONS='--insecure-registry 10.96.0.0/12 '
" | sudo tee /etc/sysconfig/crio.minikube && sudo systemctl restart crio
I1227 20:02:35.446621 67071 main.go:144] libmachine: SSH cmd err, output: <nil>:
CRIO_MINIKUBE_OPTIONS='--insecure-registry 10.96.0.0/12 '
I1227 20:02:35.446644 67071 machine.go:97] duration metric: took 1.23418137s to provisionDockerMachine
I1227 20:02:35.446659 67071 start.go:293] postStartSetup for "functional-866869" (driver="kvm2")
I1227 20:02:35.446671 67071 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I1227 20:02:35.446753 67071 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I1227 20:02:35.450071 67071 main.go:144] libmachine: domain functional-866869 has defined MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:35.450573 67071 main.go:144] libmachine: found host DHCP lease matching {name: "", mac: "52:54:00:61:ec:a4", ip: ""} in network mk-functional-866869: {Iface:virbr1 ExpiryTime:2025-12-27 21:00:42 +0000 UTC Type:0 Mac:52:54:00:61:ec:a4 Iaid: IPaddr:192.168.39.233 Prefix:24 Hostname:functional-866869 Clientid:01:52:54:00:61:ec:a4}
I1227 20:02:35.450619 67071 main.go:144] libmachine: domain functional-866869 has defined IP address 192.168.39.233 and MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:35.450805 67071 sshutil.go:53] new ssh client: &{IP:192.168.39.233 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/22332-59055/.minikube/machines/functional-866869/id_rsa Username:docker}
I1227 20:02:35.541183 67071 ssh_runner.go:195] Run: cat /etc/os-release
I1227 20:02:35.546706 67071 info.go:137] Remote host: Buildroot 2025.02
I1227 20:02:35.546760 67071 filesync.go:126] Scanning /home/jenkins/minikube-integration/22332-59055/.minikube/addons for local assets ...
I1227 20:02:35.546849 67071 filesync.go:126] Scanning /home/jenkins/minikube-integration/22332-59055/.minikube/files for local assets ...
I1227 20:02:35.546922 67071 filesync.go:149] local asset: /home/jenkins/minikube-integration/22332-59055/.minikube/files/etc/ssl/certs/629372.pem -> 629372.pem in /etc/ssl/certs
I1227 20:02:35.546988 67071 filesync.go:149] local asset: /home/jenkins/minikube-integration/22332-59055/.minikube/files/etc/test/nested/copy/62937/hosts -> hosts in /etc/test/nested/copy/62937
I1227 20:02:35.547032 67071 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs /etc/test/nested/copy/62937
I1227 20:02:35.559556 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/files/etc/ssl/certs/629372.pem --> /etc/ssl/certs/629372.pem (1708 bytes)
I1227 20:02:35.596383 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/files/etc/test/nested/copy/62937/hosts --> /etc/test/nested/copy/62937/hosts (40 bytes)
I1227 20:02:35.630158 67071 start.go:296] duration metric: took 183.478668ms for postStartSetup
I1227 20:02:35.630236 67071 fix.go:56] duration metric: took 1.421059499s for fixHost
I1227 20:02:35.633574 67071 main.go:144] libmachine: domain functional-866869 has defined MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:35.633995 67071 main.go:144] libmachine: found host DHCP lease matching {name: "", mac: "52:54:00:61:ec:a4", ip: ""} in network mk-functional-866869: {Iface:virbr1 ExpiryTime:2025-12-27 21:00:42 +0000 UTC Type:0 Mac:52:54:00:61:ec:a4 Iaid: IPaddr:192.168.39.233 Prefix:24 Hostname:functional-866869 Clientid:01:52:54:00:61:ec:a4}
I1227 20:02:35.634016 67071 main.go:144] libmachine: domain functional-866869 has defined IP address 192.168.39.233 and MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:35.634200 67071 main.go:144] libmachine: Using SSH client type: native
I1227 20:02:35.634492 67071 main.go:144] libmachine: &{{{<nil> 0 [] [] []} docker [0x84e300] 0x850fa0 <nil> [] 0s} 192.168.39.233 22 <nil> <nil>}
I1227 20:02:35.634500 67071 main.go:144] libmachine: About to run SSH command:
date +%s.%N
I1227 20:02:35.749619 67071 main.go:144] libmachine: SSH cmd err, output: <nil>: 1766865755.744614389
I1227 20:02:35.749634 67071 fix.go:216] guest clock: 1766865755.744614389
I1227 20:02:35.749640 67071 fix.go:229] Guest: 2025-12-27 20:02:35.744614389 +0000 UTC Remote: 2025-12-27 20:02:35.630241746 +0000 UTC m=+1.522455046 (delta=114.372643ms)
I1227 20:02:35.749656 67071 fix.go:200] guest clock delta is within tolerance: 114.372643ms
I1227 20:02:35.749660 67071 start.go:83] releasing machines lock for "functional-866869", held for 1.540533102s
I1227 20:02:35.752689 67071 main.go:144] libmachine: domain functional-866869 has defined MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:35.753046 67071 main.go:144] libmachine: found host DHCP lease matching {name: "", mac: "52:54:00:61:ec:a4", ip: ""} in network mk-functional-866869: {Iface:virbr1 ExpiryTime:2025-12-27 21:00:42 +0000 UTC Type:0 Mac:52:54:00:61:ec:a4 Iaid: IPaddr:192.168.39.233 Prefix:24 Hostname:functional-866869 Clientid:01:52:54:00:61:ec:a4}
I1227 20:02:35.753064 67071 main.go:144] libmachine: domain functional-866869 has defined IP address 192.168.39.233 and MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:35.753696 67071 ssh_runner.go:195] Run: cat /version.json
I1227 20:02:35.753803 67071 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I1227 20:02:35.756645 67071 main.go:144] libmachine: domain functional-866869 has defined MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:35.757045 67071 main.go:144] libmachine: found host DHCP lease matching {name: "", mac: "52:54:00:61:ec:a4", ip: ""} in network mk-functional-866869: {Iface:virbr1 ExpiryTime:2025-12-27 21:00:42 +0000 UTC Type:0 Mac:52:54:00:61:ec:a4 Iaid: IPaddr:192.168.39.233 Prefix:24 Hostname:functional-866869 Clientid:01:52:54:00:61:ec:a4}
I1227 20:02:35.757061 67071 main.go:144] libmachine: domain functional-866869 has defined IP address 192.168.39.233 and MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:35.757128 67071 main.go:144] libmachine: domain functional-866869 has defined MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:35.757250 67071 sshutil.go:53] new ssh client: &{IP:192.168.39.233 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/22332-59055/.minikube/machines/functional-866869/id_rsa Username:docker}
I1227 20:02:35.757677 67071 main.go:144] libmachine: found host DHCP lease matching {name: "", mac: "52:54:00:61:ec:a4", ip: ""} in network mk-functional-866869: {Iface:virbr1 ExpiryTime:2025-12-27 21:00:42 +0000 UTC Type:0 Mac:52:54:00:61:ec:a4 Iaid: IPaddr:192.168.39.233 Prefix:24 Hostname:functional-866869 Clientid:01:52:54:00:61:ec:a4}
I1227 20:02:35.757701 67071 main.go:144] libmachine: domain functional-866869 has defined IP address 192.168.39.233 and MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:35.757878 67071 sshutil.go:53] new ssh client: &{IP:192.168.39.233 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/22332-59055/.minikube/machines/functional-866869/id_rsa Username:docker}
I1227 20:02:35.839243 67071 ssh_runner.go:195] Run: systemctl --version
I1227 20:02:35.863879 67071 ssh_runner.go:195] Run: sudo sh -c "podman version >/dev/null"
I1227 20:02:36.011937 67071 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
W1227 20:02:36.019383 67071 cni.go:209] loopback cni configuration skipped: "/etc/cni/net.d/*loopback.conf*" not found
I1227 20:02:36.019458 67071 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I1227 20:02:36.031847 67071 cni.go:259] no active bridge cni configs found in "/etc/cni/net.d" - nothing to disable
I1227 20:02:36.031866 67071 start.go:496] detecting cgroup driver to use...
I1227 20:02:36.031891 67071 start.go:519] Kubernetes 1.35.0+ detected, using "systemd" cgroup driver
I1227 20:02:36.031968 67071 ssh_runner.go:195] Run: sudo systemctl stop -f containerd
I1227 20:02:36.053057 67071 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I1227 20:02:36.071206 67071 docker.go:218] disabling cri-docker service (if available) ...
I1227 20:02:36.071256 67071 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
I1227 20:02:36.094078 67071 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
I1227 20:02:36.111432 67071 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
I1227 20:02:36.336363 67071 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
I1227 20:02:36.550167 67071 docker.go:234] disabling docker service ...
I1227 20:02:36.550237 67071 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
I1227 20:02:36.581308 67071 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
I1227 20:02:36.598501 67071 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
I1227 20:02:36.822473 67071 ssh_runner.go:195] Run: sudo systemctl mask docker.service
I1227 20:02:37.023655 67071 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
I1227 20:02:37.040845 67071 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/crio/crio.sock
" | sudo tee /etc/crictl.yaml"
I1227 20:02:37.065547 67071 crio.go:59] configure cri-o to use "registry.k8s.io/pause:3.10.1" pause image...
I1227 20:02:37.065627 67071 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|^.*pause_image = .*$|pause_image = "registry.k8s.io/pause:3.10.1"|' /etc/crio/crio.conf.d/02-crio.conf"
I1227 20:02:37.079160 67071 crio.go:70] configuring cri-o to use "systemd" as cgroup driver...
I1227 20:02:37.079233 67071 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|^.*cgroup_manager = .*$|cgroup_manager = "systemd"|' /etc/crio/crio.conf.d/02-crio.conf"
I1227 20:02:37.093547 67071 ssh_runner.go:195] Run: sh -c "sudo sed -i '/conmon_cgroup = .*/d' /etc/crio/crio.conf.d/02-crio.conf"
I1227 20:02:37.107189 67071 ssh_runner.go:195] Run: sh -c "sudo sed -i '/cgroup_manager = .*/a conmon_cgroup = "pod"' /etc/crio/crio.conf.d/02-crio.conf"
I1227 20:02:37.121098 67071 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I1227 20:02:37.134917 67071 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *"net.ipv4.ip_unprivileged_port_start=.*"/d' /etc/crio/crio.conf.d/02-crio.conf"
I1227 20:02:37.147946 67071 ssh_runner.go:195] Run: sh -c "sudo grep -q "^ *default_sysctls" /etc/crio/crio.conf.d/02-crio.conf || sudo sed -i '/conmon_cgroup = .*/a default_sysctls = \[\n\]' /etc/crio/crio.conf.d/02-crio.conf"
I1227 20:02:37.161940 67071 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^default_sysctls *= *\[|&\n "net.ipv4.ip_unprivileged_port_start=0",|' /etc/crio/crio.conf.d/02-crio.conf"
I1227 20:02:37.175520 67071 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I1227 20:02:37.187633 67071 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I1227 20:02:37.201891 67071 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1227 20:02:37.415049 67071 ssh_runner.go:195] Run: sudo systemctl restart crio
I1227 20:02:37.874955 67071 start.go:553] Will wait 60s for socket path /var/run/crio/crio.sock
I1227 20:02:37.875033 67071 ssh_runner.go:195] Run: stat /var/run/crio/crio.sock
I1227 20:02:37.881399 67071 start.go:574] Will wait 60s for crictl version
I1227 20:02:37.881473 67071 ssh_runner.go:195] Run: which crictl
I1227 20:02:37.886192 67071 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I1227 20:02:37.919852 67071 start.go:590] Version: 0.1.0
RuntimeName: cri-o
RuntimeVersion: 1.35.0
RuntimeApiVersion: v1
I1227 20:02:37.919966 67071 ssh_runner.go:195] Run: crio --version
I1227 20:02:37.954373 67071 ssh_runner.go:195] Run: crio --version
I1227 20:02:37.990227 67071 out.go:179] * Preparing Kubernetes v1.35.0 on CRI-O 1.35.0 ...
I1227 20:02:37.994457 67071 main.go:144] libmachine: domain functional-866869 has defined MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:37.994977 67071 main.go:144] libmachine: found host DHCP lease matching {name: "", mac: "52:54:00:61:ec:a4", ip: ""} in network mk-functional-866869: {Iface:virbr1 ExpiryTime:2025-12-27 21:00:42 +0000 UTC Type:0 Mac:52:54:00:61:ec:a4 Iaid: IPaddr:192.168.39.233 Prefix:24 Hostname:functional-866869 Clientid:01:52:54:00:61:ec:a4}
I1227 20:02:37.995001 67071 main.go:144] libmachine: domain functional-866869 has defined IP address 192.168.39.233 and MAC address 52:54:00:61:ec:a4 in network mk-functional-866869
I1227 20:02:37.995188 67071 ssh_runner.go:195] Run: grep 192.168.39.1 host.minikube.internal$ /etc/hosts
I1227 20:02:38.001807 67071 out.go:179] - apiserver.enable-admission-plugins=NamespaceAutoProvision
I1227 20:02:38.003260 67071 kubeadm.go:884] updating cluster {Name:functional-866869 KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube-builds/iso/22332/minikube-v1.37.0-1766811082-22332-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1766570851-22316@sha256:7975a7a1117280f99ad7696c9c80bdca993064fe9e309e9984685e0ce989758a Memory:4096 CPUs:2 DiskSize:20000 Driver:kvm2 HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8441 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1
.35.0 ClusterName:functional-866869 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:crio CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[{Component:apiserver Key:enable-admission-plugins Value:NamespaceAutoProvision}] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.39.233 Port:8441 KubernetesVersion:v1.35.0 ContainerRuntime:crio ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Moun
t9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s Rosetta:false} ...
I1227 20:02:38.003413 67071 preload.go:188] Checking if preload exists for k8s version v1.35.0 and runtime crio
I1227 20:02:38.003470 67071 ssh_runner.go:195] Run: sudo crictl images --output json
I1227 20:02:38.046840 67071 crio.go:561] all images are preloaded for cri-o runtime.
I1227 20:02:38.046858 67071 crio.go:433] Images already preloaded, skipping extraction
I1227 20:02:38.046911 67071 ssh_runner.go:195] Run: sudo crictl images --output json
I1227 20:02:38.080007 67071 crio.go:561] all images are preloaded for cri-o runtime.
I1227 20:02:38.080025 67071 cache_images.go:86] Images are preloaded, skipping loading
I1227 20:02:38.080035 67071 kubeadm.go:935] updating node { 192.168.39.233 8441 v1.35.0 crio true true} ...
I1227 20:02:38.080158 67071 kubeadm.go:947] kubelet [Unit]
Wants=crio.service
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.35.0/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=functional-866869 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.39.233
[Install]
config:
{KubernetesVersion:v1.35.0 ClusterName:functional-866869 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:crio CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[{Component:apiserver Key:enable-admission-plugins Value:NamespaceAutoProvision}] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I1227 20:02:38.080237 67071 ssh_runner.go:195] Run: crio config
I1227 20:02:38.137458 67071 extraconfig.go:125] Overwriting default enable-admission-plugins=NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota with user provided enable-admission-plugins=NamespaceAutoProvision for component apiserver
I1227 20:02:38.137498 67071 cni.go:84] Creating CNI manager for ""
I1227 20:02:38.137509 67071 cni.go:146] "kvm2" driver + "crio" runtime found, recommending bridge
I1227 20:02:38.137518 67071 kubeadm.go:85] Using pod CIDR: 10.244.0.0/16
I1227 20:02:38.137540 67071 kubeadm.go:197] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.39.233 APIServerPort:8441 KubernetesVersion:v1.35.0 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:functional-866869 NodeName:functional-866869 DNSDomain:cluster.local CRISocket:/var/run/crio/crio.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceAutoProvision] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.39.233"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.39.233 CgroupDriver:systemd ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOp
ts:map[containerRuntimeEndpoint:unix:///var/run/crio/crio.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I1227 20:02:38.137675 67071 kubeadm.go:203] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta4
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.39.233
bindPort: 8441
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/crio/crio.sock
name: "functional-866869"
kubeletExtraArgs:
- name: "node-ip"
value: "192.168.39.233"
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta4
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.39.233"]
extraArgs:
- name: "enable-admission-plugins"
value: "NamespaceAutoProvision"
controllerManager:
extraArgs:
- name: "allocate-node-cidrs"
value: "true"
- name: "leader-elect"
value: "false"
scheduler:
extraArgs:
- name: "leader-elect"
value: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8441
etcd:
local:
dataDir: /var/lib/minikube/etcd
kubernetesVersion: v1.35.0
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: systemd
containerRuntimeEndpoint: unix:///var/run/crio/crio.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I1227 20:02:38.137752 67071 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.35.0
I1227 20:02:38.151970 67071 binaries.go:51] Found k8s binaries, skipping transfer
I1227 20:02:38.152054 67071 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I1227 20:02:38.165274 67071 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (317 bytes)
I1227 20:02:38.187251 67071 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I1227 20:02:38.209856 67071 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2069 bytes)
I1227 20:02:38.233216 67071 ssh_runner.go:195] Run: grep 192.168.39.233 control-plane.minikube.internal$ /etc/hosts
I1227 20:02:38.238172 67071 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1227 20:02:38.450831 67071 ssh_runner.go:195] Run: sudo systemctl start kubelet
I1227 20:02:38.469748 67071 certs.go:69] Setting up /home/jenkins/minikube-integration/22332-59055/.minikube/profiles/functional-866869 for IP: 192.168.39.233
I1227 20:02:38.469761 67071 certs.go:195] generating shared ca certs ...
I1227 20:02:38.469777 67071 certs.go:227] acquiring lock for ca certs: {Name:mkaababc7dc2fa0b2cccf395a6ff1958c07efd0c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1227 20:02:38.469994 67071 certs.go:236] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/22332-59055/.minikube/ca.key
I1227 20:02:38.470050 67071 certs.go:236] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/22332-59055/.minikube/proxy-client-ca.key
I1227 20:02:38.470056 67071 certs.go:257] generating profile certs ...
I1227 20:02:38.470147 67071 certs.go:360] skipping valid signed profile cert regeneration for "minikube-user": /home/jenkins/minikube-integration/22332-59055/.minikube/profiles/functional-866869/client.key
I1227 20:02:38.470188 67071 certs.go:360] skipping valid signed profile cert regeneration for "minikube": /home/jenkins/minikube-integration/22332-59055/.minikube/profiles/functional-866869/apiserver.key.585845fc
I1227 20:02:38.470222 67071 certs.go:360] skipping valid signed profile cert regeneration for "aggregator": /home/jenkins/minikube-integration/22332-59055/.minikube/profiles/functional-866869/proxy-client.key
I1227 20:02:38.470327 67071 certs.go:484] found cert: /home/jenkins/minikube-integration/22332-59055/.minikube/certs/62937.pem (1338 bytes)
W1227 20:02:38.470358 67071 certs.go:480] ignoring /home/jenkins/minikube-integration/22332-59055/.minikube/certs/62937_empty.pem, impossibly tiny 0 bytes
I1227 20:02:38.470363 67071 certs.go:484] found cert: /home/jenkins/minikube-integration/22332-59055/.minikube/certs/ca-key.pem (1675 bytes)
I1227 20:02:38.470384 67071 certs.go:484] found cert: /home/jenkins/minikube-integration/22332-59055/.minikube/certs/ca.pem (1078 bytes)
I1227 20:02:38.470406 67071 certs.go:484] found cert: /home/jenkins/minikube-integration/22332-59055/.minikube/certs/cert.pem (1123 bytes)
I1227 20:02:38.470424 67071 certs.go:484] found cert: /home/jenkins/minikube-integration/22332-59055/.minikube/certs/key.pem (1679 bytes)
I1227 20:02:38.470462 67071 certs.go:484] found cert: /home/jenkins/minikube-integration/22332-59055/.minikube/files/etc/ssl/certs/629372.pem (1708 bytes)
I1227 20:02:38.471158 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I1227 20:02:38.507361 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I1227 20:02:38.541342 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I1227 20:02:38.572863 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I1227 20:02:38.604830 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/profiles/functional-866869/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1424 bytes)
I1227 20:02:38.640648 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/profiles/functional-866869/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I1227 20:02:38.673006 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/profiles/functional-866869/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I1227 20:02:38.706379 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/profiles/functional-866869/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I1227 20:02:38.740498 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/certs/62937.pem --> /usr/share/ca-certificates/62937.pem (1338 bytes)
I1227 20:02:38.773430 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/files/etc/ssl/certs/629372.pem --> /usr/share/ca-certificates/629372.pem (1708 bytes)
I1227 20:02:38.805174 67071 ssh_runner.go:362] scp /home/jenkins/minikube-integration/22332-59055/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I1227 20:02:38.837663 67071 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (722 bytes)
I1227 20:02:38.859754 67071 ssh_runner.go:195] Run: openssl version
I1227 20:02:38.866682 67071 ssh_runner.go:195] Run: sudo test -s /usr/share/ca-certificates/minikubeCA.pem
I1227 20:02:38.879324 67071 ssh_runner.go:195] Run: sudo ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem
I1227 20:02:38.895053 67071 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I1227 20:02:38.902318 67071 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Dec 27 19:55 /usr/share/ca-certificates/minikubeCA.pem
I1227 20:02:38.902373 67071 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I1227 20:02:38.910129 67071 ssh_runner.go:195] Run: sudo test -L /etc/ssl/certs/b5213941.0
I1227 20:02:38.922631 67071 ssh_runner.go:195] Run: sudo test -s /usr/share/ca-certificates/62937.pem
I1227 20:02:38.935684 67071 ssh_runner.go:195] Run: sudo ln -fs /usr/share/ca-certificates/62937.pem /etc/ssl/certs/62937.pem
I1227 20:02:38.948343 67071 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/62937.pem
I1227 20:02:38.954792 67071 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 Dec 27 20:00 /usr/share/ca-certificates/62937.pem
I1227 20:02:38.954847 67071 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/62937.pem
I1227 20:02:38.962851 67071 ssh_runner.go:195] Run: sudo test -L /etc/ssl/certs/51391683.0
I1227 20:02:38.975775 67071 ssh_runner.go:195] Run: sudo test -s /usr/share/ca-certificates/629372.pem
I1227 20:02:38.989491 67071 ssh_runner.go:195] Run: sudo ln -fs /usr/share/ca-certificates/629372.pem /etc/ssl/certs/629372.pem
I1227 20:02:39.002163 67071 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/629372.pem
I1227 20:02:39.008349 67071 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 Dec 27 20:00 /usr/share/ca-certificates/629372.pem
I1227 20:02:39.008409 67071 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/629372.pem
I1227 20:02:39.016617 67071 ssh_runner.go:195] Run: sudo test -L /etc/ssl/certs/3ec20f2e.0
I1227 20:02:39.030577 67071 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I1227 20:02:39.036937 67071 ssh_runner.go:195] Run: openssl x509 -noout -in /var/lib/minikube/certs/apiserver-etcd-client.crt -checkend 86400
I1227 20:02:39.044581 67071 ssh_runner.go:195] Run: openssl x509 -noout -in /var/lib/minikube/certs/apiserver-kubelet-client.crt -checkend 86400
I1227 20:02:39.052507 67071 ssh_runner.go:195] Run: openssl x509 -noout -in /var/lib/minikube/certs/etcd/server.crt -checkend 86400
I1227 20:02:39.060063 67071 ssh_runner.go:195] Run: openssl x509 -noout -in /var/lib/minikube/certs/etcd/healthcheck-client.crt -checkend 86400
I1227 20:02:39.068040 67071 ssh_runner.go:195] Run: openssl x509 -noout -in /var/lib/minikube/certs/etcd/peer.crt -checkend 86400
I1227 20:02:39.075544 67071 ssh_runner.go:195] Run: openssl x509 -noout -in /var/lib/minikube/certs/front-proxy-client.crt -checkend 86400
I1227 20:02:39.082869 67071 kubeadm.go:401] StartCluster: {Name:functional-866869 KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube-builds/iso/22332/minikube-v1.37.0-1766811082-22332-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1766570851-22316@sha256:7975a7a1117280f99ad7696c9c80bdca993064fe9e309e9984685e0ce989758a Memory:4096 CPUs:2 DiskSize:20000 Driver:kvm2 HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8441 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.35
.0 ClusterName:functional-866869 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:crio CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[{Component:apiserver Key:enable-admission-plugins Value:NamespaceAutoProvision}] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.39.233 Port:8441 KubernetesVersion:v1.35.0 ContainerRuntime:crio ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9P
Version:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s Rosetta:false}
I1227 20:02:39.082951 67071 cri.go:61] listing CRI containers in root : {State:paused Name: Namespaces:[kube-system]}
I1227 20:02:39.083017 67071 ssh_runner.go:195] Run: sudo -s eval "crictl --timeout=10s ps -a --quiet --label io.kubernetes.pod.namespace=kube-system"
I1227 20:02:39.122195 67071 cri.go:96] found id: "ce191c599168870ba1b21a5aea73f026100dca03a8f5de06ee9f8c4625002569"
I1227 20:02:39.122212 67071 cri.go:96] found id: "fdcacbca0e88cc7b2078f29fed3d52a0240d76b4af367957526f13d46c6ae327"
I1227 20:02:39.122216 67071 cri.go:96] found id: "bd9d719e21a64fbb304f5c5880c24405c3114395ca851c56caa1fb1eed13d3de"
I1227 20:02:39.122220 67071 cri.go:96] found id: "0c0fc056da8e6ab0f45b663c742ccd950720771653b60978bdf6d2b91ac6a56f"
I1227 20:02:39.122224 67071 cri.go:96] found id: "a3b78daea5b561743fdcfaf4ff523bf3eb65aa9cf7650cdfad11ae53c6c39109"
I1227 20:02:39.122227 67071 cri.go:96] found id: "7cc6ce7f1e32d80636f28c3de137b596be9ea7d125b48649e8a8a1d596b1832d"
I1227 20:02:39.122230 67071 cri.go:96] found id: "f5ee32319e9274cd30cc9cb7ccc8dc3593153810b117a710339c041209f9142a"
I1227 20:02:39.122233 67071 cri.go:96] found id: "184fea3e33981ccb8efdf67323a0d5fd4f6cde9d2963ed1cd9db428879de196f"
I1227 20:02:39.122236 67071 cri.go:96] found id: "2c3773082525b68caef292485a709f468a4ca6174b9b03037fa82769109a7ff5"
I1227 20:02:39.122266 67071 cri.go:96] found id: "74f9eadffeb3f22331e3806fc23ff902a743d9444d1543849836174980ed7096"
I1227 20:02:39.122270 67071 cri.go:96] found id: "716c479bfd66fd1d4795b7fb25f8db84901323555444ecc4711defff384fae8c"
I1227 20:02:39.122273 67071 cri.go:96] found id: "95b6278a1ffacc0a433008eb9c0f0d032b4b98534901dc588e99faa2c7d8114f"
I1227 20:02:39.122276 67071 cri.go:96] found id: "14a0a905e7f2b7a2d9cd72441b0565f562ad83c238deeee6c8767769320c93f7"
I1227 20:02:39.122279 67071 cri.go:96] found id: "184a5fc99508f1c6a749a3c91042358fdd4df627d6f33ad81a6b8af6a5a267e7"
I1227 20:02:39.122282 67071 cri.go:96] found id: ""
I1227 20:02:39.122347 67071 ssh_runner.go:195] Run: sudo runc list -f json
-- /stdout --
helpers_test.go:263: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p functional-866869 -n functional-866869
helpers_test.go:270: (dbg) Run: kubectl --context functional-866869 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:294: <<< TestFunctional/serial/ComponentHealth FAILED: end of post-mortem logs <<<
helpers_test.go:295: ---------------------/post-mortem---------------------------------
--- FAIL: TestFunctional/serial/ComponentHealth (1.96s)