=== RUN TestMultiControlPlane/serial/DeployApp
ha_test.go:128: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- apply -f ./testdata/ha/ha-pod-dns-test.yaml
ha_test.go:133: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- rollout status deployment/busybox
ha_test.go:133: (dbg) Done: out/minikube-linux-amd64 kubectl -p ha-238496 -- rollout status deployment/busybox: (5.929027416s)
ha_test.go:140: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:149: expected 3 Pod IPs but got 4 (may be temporary), output: "\n-- stdout --\n\t'10.244.2.2 10.244.2.3 10.244.0.4 10.244.1.2'\n\n-- /stdout --"
ha_test.go:140: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:149: expected 3 Pod IPs but got 4 (may be temporary), output: "\n-- stdout --\n\t'10.244.2.2 10.244.2.3 10.244.0.4 10.244.1.2'\n\n-- /stdout --"
E0729 23:19:24.228331 19411 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/functional-652848/client.crt: no such file or directory
E0729 23:19:24.233585 19411 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/functional-652848/client.crt: no such file or directory
E0729 23:19:24.243915 19411 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/functional-652848/client.crt: no such file or directory
E0729 23:19:24.264232 19411 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/functional-652848/client.crt: no such file or directory
E0729 23:19:24.304549 19411 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/functional-652848/client.crt: no such file or directory
E0729 23:19:24.384877 19411 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/functional-652848/client.crt: no such file or directory
E0729 23:19:24.545307 19411 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/functional-652848/client.crt: no such file or directory
E0729 23:19:24.865594 19411 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/functional-652848/client.crt: no such file or directory
ha_test.go:140: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- get pods -o jsonpath='{.items[*].status.podIP}'
E0729 23:19:25.505774 19411 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/functional-652848/client.crt: no such file or directory
ha_test.go:149: expected 3 Pod IPs but got 4 (may be temporary), output: "\n-- stdout --\n\t'10.244.2.2 10.244.2.3 10.244.0.4 10.244.1.2'\n\n-- /stdout --"
E0729 23:19:26.787108 19411 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/functional-652848/client.crt: no such file or directory
ha_test.go:140: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:149: expected 3 Pod IPs but got 4 (may be temporary), output: "\n-- stdout --\n\t'10.244.2.2 10.244.2.3 10.244.0.4 10.244.1.2'\n\n-- /stdout --"
E0729 23:19:29.347785 19411 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/functional-652848/client.crt: no such file or directory
ha_test.go:140: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:149: expected 3 Pod IPs but got 4 (may be temporary), output: "\n-- stdout --\n\t'10.244.2.2 10.244.2.3 10.244.0.4 10.244.1.2'\n\n-- /stdout --"
E0729 23:19:34.468040 19411 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/functional-652848/client.crt: no such file or directory
ha_test.go:140: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:149: expected 3 Pod IPs but got 4 (may be temporary), output: "\n-- stdout --\n\t'10.244.2.2 10.244.2.3 10.244.0.4 10.244.1.2'\n\n-- /stdout --"
ha_test.go:140: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:149: expected 3 Pod IPs but got 4 (may be temporary), output: "\n-- stdout --\n\t'10.244.2.2 10.244.2.3 10.244.0.4 10.244.1.2'\n\n-- /stdout --"
E0729 23:19:44.708532 19411 cert_rotation.go:168] key failed with : open /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/functional-652848/client.crt: no such file or directory
ha_test.go:140: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- get pods -o jsonpath='{.items[*].status.podIP}'
ha_test.go:163: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- get pods -o jsonpath='{.items[*].metadata.name}'
ha_test.go:171: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-8ql68 -- nslookup kubernetes.io
ha_test.go:171: (dbg) Non-zero exit: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-8ql68 -- nslookup kubernetes.io: exit status 1 (104.452814ms)
** stderr **
error: cannot exec into a container in a completed pod; current phase is Failed
** /stderr **
ha_test.go:173: Pod busybox-fc5497c4f-8ql68 could not resolve 'kubernetes.io': exit status 1
ha_test.go:171: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-d42qb -- nslookup kubernetes.io
ha_test.go:171: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-ftt4w -- nslookup kubernetes.io
ha_test.go:171: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-scl6h -- nslookup kubernetes.io
ha_test.go:181: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-8ql68 -- nslookup kubernetes.default
ha_test.go:181: (dbg) Non-zero exit: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-8ql68 -- nslookup kubernetes.default: exit status 1 (106.003564ms)
** stderr **
Error from server (NotFound): pods "busybox-fc5497c4f-8ql68" not found
** /stderr **
ha_test.go:183: Pod busybox-fc5497c4f-8ql68 could not resolve 'kubernetes.default': exit status 1
ha_test.go:181: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-d42qb -- nslookup kubernetes.default
ha_test.go:181: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-ftt4w -- nslookup kubernetes.default
ha_test.go:181: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-scl6h -- nslookup kubernetes.default
ha_test.go:189: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-8ql68 -- nslookup kubernetes.default.svc.cluster.local
ha_test.go:189: (dbg) Non-zero exit: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-8ql68 -- nslookup kubernetes.default.svc.cluster.local: exit status 1 (101.516243ms)
** stderr **
Error from server (NotFound): pods "busybox-fc5497c4f-8ql68" not found
** /stderr **
ha_test.go:191: Pod busybox-fc5497c4f-8ql68 could not resolve local service (kubernetes.default.svc.cluster.local): exit status 1
ha_test.go:189: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-d42qb -- nslookup kubernetes.default.svc.cluster.local
ha_test.go:189: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-ftt4w -- nslookup kubernetes.default.svc.cluster.local
ha_test.go:189: (dbg) Run: out/minikube-linux-amd64 kubectl -p ha-238496 -- exec busybox-fc5497c4f-scl6h -- nslookup kubernetes.default.svc.cluster.local
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p ha-238496 -n ha-238496
helpers_test.go:244: <<< TestMultiControlPlane/serial/DeployApp FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestMultiControlPlane/serial/DeployApp]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p ha-238496 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-amd64 -p ha-238496 logs -n 25: (1.084438141s)
helpers_test.go:252: TestMultiControlPlane/serial/DeployApp logs:
-- stdout --
==> Audit <==
|---------|--------------------------------------|-------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|--------------------------------------|-------------------|---------|---------|---------------------|---------------------|
| delete | -p functional-652848 | functional-652848 | jenkins | v1.33.1 | 29 Jul 24 23:15 UTC | 29 Jul 24 23:15 UTC |
| start | -p ha-238496 --wait=true | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:15 UTC | 29 Jul 24 23:19 UTC |
| | --memory=2200 --ha | | | | | |
| | -v=7 --alsologtostderr | | | | | |
| | --driver=kvm2 | | | | | |
| kubectl | -p ha-238496 -- apply -f | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | ./testdata/ha/ha-pod-dns-test.yaml | | | | | |
| kubectl | -p ha-238496 -- rollout status | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | deployment/busybox | | | | | |
| kubectl | -p ha-238496 -- get pods -o | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | jsonpath='{.items[*].status.podIP}' | | | | | |
| kubectl | -p ha-238496 -- get pods -o | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | jsonpath='{.items[*].status.podIP}' | | | | | |
| kubectl | -p ha-238496 -- get pods -o | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | jsonpath='{.items[*].status.podIP}' | | | | | |
| kubectl | -p ha-238496 -- get pods -o | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | jsonpath='{.items[*].status.podIP}' | | | | | |
| kubectl | -p ha-238496 -- get pods -o | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | jsonpath='{.items[*].status.podIP}' | | | | | |
| kubectl | -p ha-238496 -- get pods -o | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | jsonpath='{.items[*].status.podIP}' | | | | | |
| kubectl | -p ha-238496 -- get pods -o | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | jsonpath='{.items[*].status.podIP}' | | | | | |
| kubectl | -p ha-238496 -- get pods -o | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | jsonpath='{.items[*].status.podIP}' | | | | | |
| kubectl | -p ha-238496 -- get pods -o | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | jsonpath='{.items[*].metadata.name}' | | | | | |
| kubectl | -p ha-238496 -- exec | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | |
| | busybox-fc5497c4f-8ql68 -- | | | | | |
| | nslookup kubernetes.io | | | | | |
| kubectl | -p ha-238496 -- exec | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | busybox-fc5497c4f-d42qb -- | | | | | |
| | nslookup kubernetes.io | | | | | |
| kubectl | -p ha-238496 -- exec | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | busybox-fc5497c4f-ftt4w -- | | | | | |
| | nslookup kubernetes.io | | | | | |
| kubectl | -p ha-238496 -- exec | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | busybox-fc5497c4f-scl6h -- | | | | | |
| | nslookup kubernetes.io | | | | | |
| kubectl | -p ha-238496 -- exec | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | |
| | busybox-fc5497c4f-8ql68 -- | | | | | |
| | nslookup kubernetes.default | | | | | |
| kubectl | -p ha-238496 -- exec | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | busybox-fc5497c4f-d42qb -- | | | | | |
| | nslookup kubernetes.default | | | | | |
| kubectl | -p ha-238496 -- exec | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | busybox-fc5497c4f-ftt4w -- | | | | | |
| | nslookup kubernetes.default | | | | | |
| kubectl | -p ha-238496 -- exec | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | busybox-fc5497c4f-scl6h -- | | | | | |
| | nslookup kubernetes.default | | | | | |
| kubectl | -p ha-238496 -- exec | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | |
| | busybox-fc5497c4f-8ql68 -- nslookup | | | | | |
| | kubernetes.default.svc.cluster.local | | | | | |
| kubectl | -p ha-238496 -- exec | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | busybox-fc5497c4f-d42qb -- nslookup | | | | | |
| | kubernetes.default.svc.cluster.local | | | | | |
| kubectl | -p ha-238496 -- exec | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | busybox-fc5497c4f-ftt4w -- nslookup | | | | | |
| | kubernetes.default.svc.cluster.local | | | | | |
| kubectl | -p ha-238496 -- exec | ha-238496 | jenkins | v1.33.1 | 29 Jul 24 23:19 UTC | 29 Jul 24 23:19 UTC |
| | busybox-fc5497c4f-scl6h -- nslookup | | | | | |
| | kubernetes.default.svc.cluster.local | | | | | |
|---------|--------------------------------------|-------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/07/29 23:15:16
Running on machine: ubuntu-20-agent
Binary: Built with gc go1.22.5 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0729 23:15:16.010336 29396 out.go:291] Setting OutFile to fd 1 ...
I0729 23:15:16.010608 29396 out.go:338] TERM=,COLORTERM=, which probably does not support color
I0729 23:15:16.010618 29396 out.go:304] Setting ErrFile to fd 2...
I0729 23:15:16.010622 29396 out.go:338] TERM=,COLORTERM=, which probably does not support color
I0729 23:15:16.010886 29396 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19347-12221/.minikube/bin
I0729 23:15:16.011469 29396 out.go:298] Setting JSON to false
I0729 23:15:16.012305 29396 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent","uptime":3412,"bootTime":1722291504,"procs":177,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1065-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0729 23:15:16.012359 29396 start.go:139] virtualization: kvm guest
I0729 23:15:16.014349 29396 out.go:177] * [ha-238496] minikube v1.33.1 on Ubuntu 20.04 (kvm/amd64)
I0729 23:15:16.015655 29396 out.go:177] - MINIKUBE_LOCATION=19347
I0729 23:15:16.015702 29396 notify.go:220] Checking for updates...
I0729 23:15:16.018055 29396 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0729 23:15:16.019280 29396 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19347-12221/kubeconfig
I0729 23:15:16.020588 29396 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19347-12221/.minikube
I0729 23:15:16.021752 29396 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0729 23:15:16.022911 29396 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0729 23:15:16.024125 29396 driver.go:392] Setting default libvirt URI to qemu:///system
I0729 23:15:16.058288 29396 out.go:177] * Using the kvm2 driver based on user configuration
I0729 23:15:16.059458 29396 start.go:297] selected driver: kvm2
I0729 23:15:16.059470 29396 start.go:901] validating driver "kvm2" against <nil>
I0729 23:15:16.059483 29396 start.go:912] status for kvm2: {Installed:true Healthy:true Running:true NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0729 23:15:16.060204 29396 install.go:52] acquiring lock: {Name:mk900956b073697a4aa6c80a27c6bb0742a99a53 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0729 23:15:16.060324 29396 install.go:117] Validating docker-machine-driver-kvm2, PATH=/home/jenkins/minikube-integration/19347-12221/.minikube/bin:/home/jenkins/workspace/KVM_Linux_integration/out/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games:/usr/local/go/bin:/home/jenkins/go/bin:/usr/local/bin/:/usr/local/go/bin/:/home/jenkins/go/bin
I0729 23:15:16.075224 29396 install.go:137] /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2 version is 1.33.1
I0729 23:15:16.075274 29396 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0729 23:15:16.075538 29396 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0729 23:15:16.075566 29396 cni.go:84] Creating CNI manager for ""
I0729 23:15:16.075575 29396 cni.go:136] multinode detected (0 nodes found), recommending kindnet
I0729 23:15:16.075587 29396 start_flags.go:319] Found "CNI" CNI - setting NetworkPlugin=cni
I0729 23:15:16.075665 29396 start.go:340] cluster config:
{Name:ha-238496 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721902582-19326@sha256:540fb5dc7f38be17ff5276a38dfe6c8a4b1d9ba1c27c62244e6eebd7e37696e7 Memory:2200 CPUs:2 DiskSize:20000 Driver:kvm2 HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.30.3 ClusterName:ha-238496 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CR
ISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:true ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPI
D:0 GPUs: AutoPauseInterval:1m0s}
I0729 23:15:16.075824 29396 iso.go:125] acquiring lock: {Name:mke1b110143262a7fb7eb5e1cbaa1784fa37fd0a Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0729 23:15:16.078247 29396 out.go:177] * Starting "ha-238496" primary control-plane node in "ha-238496" cluster
I0729 23:15:16.079320 29396 preload.go:131] Checking if preload exists for k8s version v1.30.3 and runtime docker
I0729 23:15:16.079359 29396 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19347-12221/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.30.3-docker-overlay2-amd64.tar.lz4
I0729 23:15:16.079369 29396 cache.go:56] Caching tarball of preloaded images
I0729 23:15:16.079460 29396 preload.go:172] Found /home/jenkins/minikube-integration/19347-12221/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.30.3-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0729 23:15:16.079474 29396 cache.go:59] Finished verifying existence of preloaded tar for v1.30.3 on docker
I0729 23:15:16.079750 29396 profile.go:143] Saving config to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/config.json ...
I0729 23:15:16.079767 29396 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/config.json: {Name:mk83765509bbe48dfceafa2fa0be21d32b315310 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:15:16.079890 29396 start.go:360] acquireMachinesLock for ha-238496: {Name:mk79fbc287386032c39e512567e9786663e657a9 Clock:{} Delay:500ms Timeout:13m0s Cancel:<nil>}
I0729 23:15:16.079917 29396 start.go:364] duration metric: took 14.94µs to acquireMachinesLock for "ha-238496"
I0729 23:15:16.079932 29396 start.go:93] Provisioning new machine with config: &{Name:ha-238496 KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube-builds/iso/19319/minikube-v1.33.1-1721690939-19319-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721902582-19326@sha256:540fb5dc7f38be17ff5276a38dfe6c8a4b1d9ba1c27c62244e6eebd7e37696e7 Memory:2200 CPUs:2 DiskSize:20000 Driver:kvm2 HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{Kubernete
sVersion:v1.30.3 ClusterName:ha-238496 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:true ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountT
ype:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}
I0729 23:15:16.079984 29396 start.go:125] createHost starting for "" (driver="kvm2")
I0729 23:15:16.081578 29396 out.go:204] * Creating kvm2 VM (CPUs=2, Memory=2200MB, Disk=20000MB) ...
I0729 23:15:16.081687 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:15:16.081718 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:15:16.095932 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:42111
I0729 23:15:16.096323 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:15:16.096896 29396 main.go:141] libmachine: Using API Version 1
I0729 23:15:16.096923 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:15:16.097216 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:15:16.097413 29396 main.go:141] libmachine: (ha-238496) Calling .GetMachineName
I0729 23:15:16.097547 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:15:16.097777 29396 start.go:159] libmachine.API.Create for "ha-238496" (driver="kvm2")
I0729 23:15:16.097801 29396 client.go:168] LocalClient.Create starting
I0729 23:15:16.097825 29396 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem
I0729 23:15:16.097853 29396 main.go:141] libmachine: Decoding PEM data...
I0729 23:15:16.097873 29396 main.go:141] libmachine: Parsing certificate...
I0729 23:15:16.097926 29396 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem
I0729 23:15:16.097944 29396 main.go:141] libmachine: Decoding PEM data...
I0729 23:15:16.097952 29396 main.go:141] libmachine: Parsing certificate...
I0729 23:15:16.097971 29396 main.go:141] libmachine: Running pre-create checks...
I0729 23:15:16.097981 29396 main.go:141] libmachine: (ha-238496) Calling .PreCreateCheck
I0729 23:15:16.098305 29396 main.go:141] libmachine: (ha-238496) Calling .GetConfigRaw
I0729 23:15:16.098614 29396 main.go:141] libmachine: Creating machine...
I0729 23:15:16.098638 29396 main.go:141] libmachine: (ha-238496) Calling .Create
I0729 23:15:16.098767 29396 main.go:141] libmachine: (ha-238496) Creating KVM machine...
I0729 23:15:16.099974 29396 main.go:141] libmachine: (ha-238496) DBG | found existing default KVM network
I0729 23:15:16.100617 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:16.100493 29419 network.go:206] using free private subnet 192.168.39.0/24: &{IP:192.168.39.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.39.0/24 Gateway:192.168.39.1 ClientMin:192.168.39.2 ClientMax:192.168.39.254 Broadcast:192.168.39.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc00012d990}
I0729 23:15:16.100634 29396 main.go:141] libmachine: (ha-238496) DBG | created network xml:
I0729 23:15:16.100647 29396 main.go:141] libmachine: (ha-238496) DBG | <network>
I0729 23:15:16.100656 29396 main.go:141] libmachine: (ha-238496) DBG | <name>mk-ha-238496</name>
I0729 23:15:16.100664 29396 main.go:141] libmachine: (ha-238496) DBG | <dns enable='no'/>
I0729 23:15:16.100674 29396 main.go:141] libmachine: (ha-238496) DBG |
I0729 23:15:16.100689 29396 main.go:141] libmachine: (ha-238496) DBG | <ip address='192.168.39.1' netmask='255.255.255.0'>
I0729 23:15:16.100697 29396 main.go:141] libmachine: (ha-238496) DBG | <dhcp>
I0729 23:15:16.100704 29396 main.go:141] libmachine: (ha-238496) DBG | <range start='192.168.39.2' end='192.168.39.253'/>
I0729 23:15:16.100716 29396 main.go:141] libmachine: (ha-238496) DBG | </dhcp>
I0729 23:15:16.100724 29396 main.go:141] libmachine: (ha-238496) DBG | </ip>
I0729 23:15:16.100732 29396 main.go:141] libmachine: (ha-238496) DBG |
I0729 23:15:16.100740 29396 main.go:141] libmachine: (ha-238496) DBG | </network>
I0729 23:15:16.100750 29396 main.go:141] libmachine: (ha-238496) DBG |
I0729 23:15:16.105671 29396 main.go:141] libmachine: (ha-238496) DBG | trying to create private KVM network mk-ha-238496 192.168.39.0/24...
I0729 23:15:16.170490 29396 main.go:141] libmachine: (ha-238496) Setting up store path in /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496 ...
I0729 23:15:16.170526 29396 main.go:141] libmachine: (ha-238496) Building disk image from file:///home/jenkins/minikube-integration/19347-12221/.minikube/cache/iso/amd64/minikube-v1.33.1-1721690939-19319-amd64.iso
I0729 23:15:16.170537 29396 main.go:141] libmachine: (ha-238496) DBG | private KVM network mk-ha-238496 192.168.39.0/24 created
I0729 23:15:16.170555 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:16.170413 29419 common.go:145] Making disk image using store path: /home/jenkins/minikube-integration/19347-12221/.minikube
I0729 23:15:16.170571 29396 main.go:141] libmachine: (ha-238496) Downloading /home/jenkins/minikube-integration/19347-12221/.minikube/cache/boot2docker.iso from file:///home/jenkins/minikube-integration/19347-12221/.minikube/cache/iso/amd64/minikube-v1.33.1-1721690939-19319-amd64.iso...
I0729 23:15:16.419701 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:16.419556 29419 common.go:152] Creating ssh key: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/id_rsa...
I0729 23:15:16.565114 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:16.564962 29419 common.go:158] Creating raw disk image: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/ha-238496.rawdisk...
I0729 23:15:16.565144 29396 main.go:141] libmachine: (ha-238496) DBG | Writing magic tar header
I0729 23:15:16.565157 29396 main.go:141] libmachine: (ha-238496) DBG | Writing SSH key tar header
I0729 23:15:16.565167 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:16.565099 29419 common.go:172] Fixing permissions on /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496 ...
I0729 23:15:16.565212 29396 main.go:141] libmachine: (ha-238496) DBG | Checking permissions on dir: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496
I0729 23:15:16.565252 29396 main.go:141] libmachine: (ha-238496) Setting executable bit set on /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496 (perms=drwx------)
I0729 23:15:16.565263 29396 main.go:141] libmachine: (ha-238496) DBG | Checking permissions on dir: /home/jenkins/minikube-integration/19347-12221/.minikube/machines
I0729 23:15:16.565281 29396 main.go:141] libmachine: (ha-238496) DBG | Checking permissions on dir: /home/jenkins/minikube-integration/19347-12221/.minikube
I0729 23:15:16.565293 29396 main.go:141] libmachine: (ha-238496) Setting executable bit set on /home/jenkins/minikube-integration/19347-12221/.minikube/machines (perms=drwxr-xr-x)
I0729 23:15:16.565305 29396 main.go:141] libmachine: (ha-238496) DBG | Checking permissions on dir: /home/jenkins/minikube-integration/19347-12221
I0729 23:15:16.565313 29396 main.go:141] libmachine: (ha-238496) DBG | Checking permissions on dir: /home/jenkins/minikube-integration
I0729 23:15:16.565319 29396 main.go:141] libmachine: (ha-238496) DBG | Checking permissions on dir: /home/jenkins
I0729 23:15:16.565329 29396 main.go:141] libmachine: (ha-238496) Setting executable bit set on /home/jenkins/minikube-integration/19347-12221/.minikube (perms=drwxr-xr-x)
I0729 23:15:16.565334 29396 main.go:141] libmachine: (ha-238496) DBG | Checking permissions on dir: /home
I0729 23:15:16.565341 29396 main.go:141] libmachine: (ha-238496) Setting executable bit set on /home/jenkins/minikube-integration/19347-12221 (perms=drwxrwxr-x)
I0729 23:15:16.565352 29396 main.go:141] libmachine: (ha-238496) Setting executable bit set on /home/jenkins/minikube-integration (perms=drwxrwxr-x)
I0729 23:15:16.565364 29396 main.go:141] libmachine: (ha-238496) Setting executable bit set on /home/jenkins (perms=drwxr-xr-x)
I0729 23:15:16.565377 29396 main.go:141] libmachine: (ha-238496) Creating domain...
I0729 23:15:16.565390 29396 main.go:141] libmachine: (ha-238496) DBG | Skipping /home - not owner
I0729 23:15:16.566416 29396 main.go:141] libmachine: (ha-238496) define libvirt domain using xml:
I0729 23:15:16.566442 29396 main.go:141] libmachine: (ha-238496) <domain type='kvm'>
I0729 23:15:16.566450 29396 main.go:141] libmachine: (ha-238496) <name>ha-238496</name>
I0729 23:15:16.566457 29396 main.go:141] libmachine: (ha-238496) <memory unit='MiB'>2200</memory>
I0729 23:15:16.566462 29396 main.go:141] libmachine: (ha-238496) <vcpu>2</vcpu>
I0729 23:15:16.566467 29396 main.go:141] libmachine: (ha-238496) <features>
I0729 23:15:16.566475 29396 main.go:141] libmachine: (ha-238496) <acpi/>
I0729 23:15:16.566485 29396 main.go:141] libmachine: (ha-238496) <apic/>
I0729 23:15:16.566491 29396 main.go:141] libmachine: (ha-238496) <pae/>
I0729 23:15:16.566507 29396 main.go:141] libmachine: (ha-238496)
I0729 23:15:16.566515 29396 main.go:141] libmachine: (ha-238496) </features>
I0729 23:15:16.566520 29396 main.go:141] libmachine: (ha-238496) <cpu mode='host-passthrough'>
I0729 23:15:16.566525 29396 main.go:141] libmachine: (ha-238496)
I0729 23:15:16.566530 29396 main.go:141] libmachine: (ha-238496) </cpu>
I0729 23:15:16.566536 29396 main.go:141] libmachine: (ha-238496) <os>
I0729 23:15:16.566541 29396 main.go:141] libmachine: (ha-238496) <type>hvm</type>
I0729 23:15:16.566548 29396 main.go:141] libmachine: (ha-238496) <boot dev='cdrom'/>
I0729 23:15:16.566553 29396 main.go:141] libmachine: (ha-238496) <boot dev='hd'/>
I0729 23:15:16.566561 29396 main.go:141] libmachine: (ha-238496) <bootmenu enable='no'/>
I0729 23:15:16.566565 29396 main.go:141] libmachine: (ha-238496) </os>
I0729 23:15:16.566572 29396 main.go:141] libmachine: (ha-238496) <devices>
I0729 23:15:16.566577 29396 main.go:141] libmachine: (ha-238496) <disk type='file' device='cdrom'>
I0729 23:15:16.566612 29396 main.go:141] libmachine: (ha-238496) <source file='/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/boot2docker.iso'/>
I0729 23:15:16.566636 29396 main.go:141] libmachine: (ha-238496) <target dev='hdc' bus='scsi'/>
I0729 23:15:16.566649 29396 main.go:141] libmachine: (ha-238496) <readonly/>
I0729 23:15:16.566665 29396 main.go:141] libmachine: (ha-238496) </disk>
I0729 23:15:16.566678 29396 main.go:141] libmachine: (ha-238496) <disk type='file' device='disk'>
I0729 23:15:16.566708 29396 main.go:141] libmachine: (ha-238496) <driver name='qemu' type='raw' cache='default' io='threads' />
I0729 23:15:16.566730 29396 main.go:141] libmachine: (ha-238496) <source file='/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/ha-238496.rawdisk'/>
I0729 23:15:16.566745 29396 main.go:141] libmachine: (ha-238496) <target dev='hda' bus='virtio'/>
I0729 23:15:16.566756 29396 main.go:141] libmachine: (ha-238496) </disk>
I0729 23:15:16.566767 29396 main.go:141] libmachine: (ha-238496) <interface type='network'>
I0729 23:15:16.566780 29396 main.go:141] libmachine: (ha-238496) <source network='mk-ha-238496'/>
I0729 23:15:16.566791 29396 main.go:141] libmachine: (ha-238496) <model type='virtio'/>
I0729 23:15:16.566803 29396 main.go:141] libmachine: (ha-238496) </interface>
I0729 23:15:16.566814 29396 main.go:141] libmachine: (ha-238496) <interface type='network'>
I0729 23:15:16.566835 29396 main.go:141] libmachine: (ha-238496) <source network='default'/>
I0729 23:15:16.566852 29396 main.go:141] libmachine: (ha-238496) <model type='virtio'/>
I0729 23:15:16.566865 29396 main.go:141] libmachine: (ha-238496) </interface>
I0729 23:15:16.566879 29396 main.go:141] libmachine: (ha-238496) <serial type='pty'>
I0729 23:15:16.566885 29396 main.go:141] libmachine: (ha-238496) <target port='0'/>
I0729 23:15:16.566890 29396 main.go:141] libmachine: (ha-238496) </serial>
I0729 23:15:16.566896 29396 main.go:141] libmachine: (ha-238496) <console type='pty'>
I0729 23:15:16.566910 29396 main.go:141] libmachine: (ha-238496) <target type='serial' port='0'/>
I0729 23:15:16.566919 29396 main.go:141] libmachine: (ha-238496) </console>
I0729 23:15:16.566923 29396 main.go:141] libmachine: (ha-238496) <rng model='virtio'>
I0729 23:15:16.566932 29396 main.go:141] libmachine: (ha-238496) <backend model='random'>/dev/random</backend>
I0729 23:15:16.566936 29396 main.go:141] libmachine: (ha-238496) </rng>
I0729 23:15:16.566941 29396 main.go:141] libmachine: (ha-238496)
I0729 23:15:16.566947 29396 main.go:141] libmachine: (ha-238496)
I0729 23:15:16.566973 29396 main.go:141] libmachine: (ha-238496) </devices>
I0729 23:15:16.566988 29396 main.go:141] libmachine: (ha-238496) </domain>
I0729 23:15:16.566999 29396 main.go:141] libmachine: (ha-238496)
I0729 23:15:16.571593 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:12:68:36 in network default
I0729 23:15:16.572154 29396 main.go:141] libmachine: (ha-238496) Ensuring networks are active...
I0729 23:15:16.572168 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:16.572900 29396 main.go:141] libmachine: (ha-238496) Ensuring network default is active
I0729 23:15:16.573143 29396 main.go:141] libmachine: (ha-238496) Ensuring network mk-ha-238496 is active
I0729 23:15:16.573572 29396 main.go:141] libmachine: (ha-238496) Getting domain xml...
I0729 23:15:16.574187 29396 main.go:141] libmachine: (ha-238496) Creating domain...
I0729 23:15:17.757845 29396 main.go:141] libmachine: (ha-238496) Waiting to get IP...
I0729 23:15:17.759681 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:17.760105 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:17.760131 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:17.760076 29419 retry.go:31] will retry after 209.245228ms: waiting for machine to come up
I0729 23:15:17.970437 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:17.970829 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:17.970853 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:17.970813 29419 retry.go:31] will retry after 283.092243ms: waiting for machine to come up
I0729 23:15:18.255348 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:18.255731 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:18.255755 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:18.255694 29419 retry.go:31] will retry after 359.08307ms: waiting for machine to come up
I0729 23:15:18.616174 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:18.616649 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:18.616677 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:18.616605 29419 retry.go:31] will retry after 467.932022ms: waiting for machine to come up
I0729 23:15:19.086305 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:19.086757 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:19.086782 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:19.086720 29419 retry.go:31] will retry after 530.040761ms: waiting for machine to come up
I0729 23:15:19.618323 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:19.618874 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:19.618898 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:19.618841 29419 retry.go:31] will retry after 750.123731ms: waiting for machine to come up
I0729 23:15:20.370740 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:20.371168 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:20.371208 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:20.371132 29419 retry.go:31] will retry after 910.01431ms: waiting for machine to come up
I0729 23:15:21.282557 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:21.283069 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:21.283093 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:21.283007 29419 retry.go:31] will retry after 1.475852847s: waiting for machine to come up
I0729 23:15:22.760548 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:22.760976 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:22.761002 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:22.760929 29419 retry.go:31] will retry after 1.358011717s: waiting for machine to come up
I0729 23:15:24.120772 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:24.121105 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:24.121126 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:24.121063 29419 retry.go:31] will retry after 2.051676006s: waiting for machine to come up
I0729 23:15:26.174118 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:26.174517 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:26.174544 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:26.174467 29419 retry.go:31] will retry after 1.794194493s: waiting for machine to come up
I0729 23:15:27.971315 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:27.971709 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:27.971737 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:27.971664 29419 retry.go:31] will retry after 3.105101795s: waiting for machine to come up
I0729 23:15:31.080782 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:31.081278 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:31.081309 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:31.081209 29419 retry.go:31] will retry after 2.85435641s: waiting for machine to come up
I0729 23:15:33.936818 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:33.937191 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find current IP address of domain ha-238496 in network mk-ha-238496
I0729 23:15:33.937214 29396 main.go:141] libmachine: (ha-238496) DBG | I0729 23:15:33.937147 29419 retry.go:31] will retry after 5.319541558s: waiting for machine to come up
I0729 23:15:39.260319 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.260809 29396 main.go:141] libmachine: (ha-238496) Found IP for machine: 192.168.39.113
I0729 23:15:39.260831 29396 main.go:141] libmachine: (ha-238496) Reserving static IP address...
I0729 23:15:39.260860 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has current primary IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.261182 29396 main.go:141] libmachine: (ha-238496) DBG | unable to find host DHCP lease matching {name: "ha-238496", mac: "52:54:00:4c:48:55", ip: "192.168.39.113"} in network mk-ha-238496
I0729 23:15:39.335455 29396 main.go:141] libmachine: (ha-238496) DBG | Getting to WaitForSSH function...
I0729 23:15:39.335480 29396 main.go:141] libmachine: (ha-238496) Reserved static IP address: 192.168.39.113
I0729 23:15:39.335503 29396 main.go:141] libmachine: (ha-238496) Waiting for SSH to be available...
I0729 23:15:39.337991 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.338365 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:minikube Clientid:01:52:54:00:4c:48:55}
I0729 23:15:39.338394 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.338622 29396 main.go:141] libmachine: (ha-238496) DBG | Using SSH client type: external
I0729 23:15:39.338667 29396 main.go:141] libmachine: (ha-238496) DBG | Using SSH private key: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/id_rsa (-rw-------)
I0729 23:15:39.338712 29396 main.go:141] libmachine: (ha-238496) DBG | &{[-F /dev/null -o ConnectionAttempts=3 -o ConnectTimeout=10 -o ControlMaster=no -o ControlPath=none -o LogLevel=quiet -o PasswordAuthentication=no -o ServerAliveInterval=60 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null docker@192.168.39.113 -o IdentitiesOnly=yes -i /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/id_rsa -p 22] /usr/bin/ssh <nil>}
I0729 23:15:39.338731 29396 main.go:141] libmachine: (ha-238496) DBG | About to run SSH command:
I0729 23:15:39.338743 29396 main.go:141] libmachine: (ha-238496) DBG | exit 0
I0729 23:15:39.466724 29396 main.go:141] libmachine: (ha-238496) DBG | SSH cmd err, output: <nil>:
I0729 23:15:39.466990 29396 main.go:141] libmachine: (ha-238496) KVM machine creation complete!
I0729 23:15:39.467279 29396 main.go:141] libmachine: (ha-238496) Calling .GetConfigRaw
I0729 23:15:39.467833 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:15:39.468014 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:15:39.468168 29396 main.go:141] libmachine: Waiting for machine to be running, this may take a few minutes...
I0729 23:15:39.468181 29396 main.go:141] libmachine: (ha-238496) Calling .GetState
I0729 23:15:39.469401 29396 main.go:141] libmachine: Detecting operating system of created instance...
I0729 23:15:39.469417 29396 main.go:141] libmachine: Waiting for SSH to be available...
I0729 23:15:39.469425 29396 main.go:141] libmachine: Getting to WaitForSSH function...
I0729 23:15:39.469431 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:15:39.471628 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.471976 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:39.472007 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.472163 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:15:39.472323 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:39.472474 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:39.472620 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:15:39.472769 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:15:39.472953 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.113 22 <nil> <nil>}
I0729 23:15:39.472964 29396 main.go:141] libmachine: About to run SSH command:
exit 0
I0729 23:15:39.586090 29396 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0729 23:15:39.586116 29396 main.go:141] libmachine: Detecting the provisioner...
I0729 23:15:39.586125 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:15:39.588745 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.589128 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:39.589193 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.589246 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:15:39.589420 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:39.589583 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:39.589754 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:15:39.589909 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:15:39.590080 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.113 22 <nil> <nil>}
I0729 23:15:39.590090 29396 main.go:141] libmachine: About to run SSH command:
cat /etc/os-release
I0729 23:15:39.703946 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: NAME=Buildroot
VERSION=2023.02.9-dirty
ID=buildroot
VERSION_ID=2023.02.9
PRETTY_NAME="Buildroot 2023.02.9"
I0729 23:15:39.704026 29396 main.go:141] libmachine: found compatible host: buildroot
I0729 23:15:39.704035 29396 main.go:141] libmachine: Provisioning with buildroot...
I0729 23:15:39.704042 29396 main.go:141] libmachine: (ha-238496) Calling .GetMachineName
I0729 23:15:39.704269 29396 buildroot.go:166] provisioning hostname "ha-238496"
I0729 23:15:39.704285 29396 main.go:141] libmachine: (ha-238496) Calling .GetMachineName
I0729 23:15:39.704444 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:15:39.707086 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.707402 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:39.707431 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.707571 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:15:39.707766 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:39.707912 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:39.708018 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:15:39.708169 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:15:39.708332 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.113 22 <nil> <nil>}
I0729 23:15:39.708345 29396 main.go:141] libmachine: About to run SSH command:
sudo hostname ha-238496 && echo "ha-238496" | sudo tee /etc/hostname
I0729 23:15:39.833009 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: ha-238496
I0729 23:15:39.833034 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:15:39.835707 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.836036 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:39.836064 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.836236 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:15:39.836439 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:39.836598 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:39.836747 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:15:39.836920 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:15:39.837122 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.113 22 <nil> <nil>}
I0729 23:15:39.837141 29396 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sha-238496' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 ha-238496/g' /etc/hosts;
else
echo '127.0.1.1 ha-238496' | sudo tee -a /etc/hosts;
fi
fi
I0729 23:15:39.955917 29396 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0729 23:15:39.955943 29396 buildroot.go:172] set auth options {CertDir:/home/jenkins/minikube-integration/19347-12221/.minikube CaCertPath:/home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19347-12221/.minikube}
I0729 23:15:39.955960 29396 buildroot.go:174] setting up certificates
I0729 23:15:39.955968 29396 provision.go:84] configureAuth start
I0729 23:15:39.955977 29396 main.go:141] libmachine: (ha-238496) Calling .GetMachineName
I0729 23:15:39.956235 29396 main.go:141] libmachine: (ha-238496) Calling .GetIP
I0729 23:15:39.958684 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.959026 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:39.959052 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.959163 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:15:39.961045 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.961447 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:39.961474 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:39.961623 29396 provision.go:143] copyHostCerts
I0729 23:15:39.961650 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem -> /home/jenkins/minikube-integration/19347-12221/.minikube/ca.pem
I0729 23:15:39.961683 29396 exec_runner.go:144] found /home/jenkins/minikube-integration/19347-12221/.minikube/ca.pem, removing ...
I0729 23:15:39.961691 29396 exec_runner.go:203] rm: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.pem
I0729 23:15:39.961763 29396 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19347-12221/.minikube/ca.pem (1078 bytes)
I0729 23:15:39.961832 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem -> /home/jenkins/minikube-integration/19347-12221/.minikube/cert.pem
I0729 23:15:39.961850 29396 exec_runner.go:144] found /home/jenkins/minikube-integration/19347-12221/.minikube/cert.pem, removing ...
I0729 23:15:39.961854 29396 exec_runner.go:203] rm: /home/jenkins/minikube-integration/19347-12221/.minikube/cert.pem
I0729 23:15:39.961877 29396 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19347-12221/.minikube/cert.pem (1123 bytes)
I0729 23:15:39.961914 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/key.pem -> /home/jenkins/minikube-integration/19347-12221/.minikube/key.pem
I0729 23:15:39.961930 29396 exec_runner.go:144] found /home/jenkins/minikube-integration/19347-12221/.minikube/key.pem, removing ...
I0729 23:15:39.961937 29396 exec_runner.go:203] rm: /home/jenkins/minikube-integration/19347-12221/.minikube/key.pem
I0729 23:15:39.961957 29396 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19347-12221/.minikube/key.pem (1675 bytes)
I0729 23:15:39.962000 29396 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca-key.pem org=jenkins.ha-238496 san=[127.0.0.1 192.168.39.113 ha-238496 localhost minikube]
I0729 23:15:40.158265 29396 provision.go:177] copyRemoteCerts
I0729 23:15:40.158330 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0729 23:15:40.158351 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:15:40.160707 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:40.161020 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:40.161048 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:40.161164 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:15:40.161340 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:40.161493 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:15:40.161634 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.113 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/id_rsa Username:docker}
I0729 23:15:40.244831 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem -> /etc/docker/ca.pem
I0729 23:15:40.244915 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0729 23:15:40.270164 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server.pem -> /etc/docker/server.pem
I0729 23:15:40.270224 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server.pem --> /etc/docker/server.pem (1196 bytes)
I0729 23:15:40.294569 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server-key.pem -> /etc/docker/server-key.pem
I0729 23:15:40.294624 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
I0729 23:15:40.318683 29396 provision.go:87] duration metric: took 362.704805ms to configureAuth
I0729 23:15:40.318728 29396 buildroot.go:189] setting minikube options for container-runtime
I0729 23:15:40.318902 29396 config.go:182] Loaded profile config "ha-238496": Driver=kvm2, ContainerRuntime=docker, KubernetesVersion=v1.30.3
I0729 23:15:40.318927 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:15:40.319210 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:15:40.321636 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:40.322052 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:40.322071 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:40.322247 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:15:40.322438 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:40.322648 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:40.322792 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:15:40.322945 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:15:40.323111 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.113 22 <nil> <nil>}
I0729 23:15:40.323121 29396 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0729 23:15:40.436017 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: tmpfs
I0729 23:15:40.436041 29396 buildroot.go:70] root file system type: tmpfs
I0729 23:15:40.436132 29396 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0729 23:15:40.436149 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:15:40.438671 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:40.439024 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:40.439051 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:40.439247 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:15:40.439419 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:40.439554 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:40.439683 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:15:40.439821 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:15:40.439970 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.113 22 <nil> <nil>}
I0729 23:15:40.440026 29396 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %!s(MISSING) "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network.target minikube-automount.service docker.socket
Requires= minikube-automount.service docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=kvm2 --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0729 23:15:40.565241 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network.target minikube-automount.service docker.socket
Requires= minikube-automount.service docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=kvm2 --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0729 23:15:40.565269 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:15:40.567788 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:40.568077 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:40.568099 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:40.568318 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:15:40.568530 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:40.568695 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:40.568834 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:15:40.569029 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:15:40.569212 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.113 22 <nil> <nil>}
I0729 23:15:40.569235 29396 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0729 23:15:42.377799 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: diff: can't stat '/lib/systemd/system/docker.service': No such file or directory
Created symlink /etc/systemd/system/multi-user.target.wants/docker.service → /usr/lib/systemd/system/docker.service.
I0729 23:15:42.377830 29396 main.go:141] libmachine: Checking connection to Docker...
I0729 23:15:42.377841 29396 main.go:141] libmachine: (ha-238496) Calling .GetURL
I0729 23:15:42.379169 29396 main.go:141] libmachine: (ha-238496) DBG | Using libvirt version 6000000
I0729 23:15:42.381260 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.381560 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:42.381579 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.381732 29396 main.go:141] libmachine: Docker is up and running!
I0729 23:15:42.381747 29396 main.go:141] libmachine: Reticulating splines...
I0729 23:15:42.381755 29396 client.go:171] duration metric: took 26.283947034s to LocalClient.Create
I0729 23:15:42.381778 29396 start.go:167] duration metric: took 26.284000692s to libmachine.API.Create "ha-238496"
I0729 23:15:42.381791 29396 start.go:293] postStartSetup for "ha-238496" (driver="kvm2")
I0729 23:15:42.381803 29396 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0729 23:15:42.381827 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:15:42.382062 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0729 23:15:42.382082 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:15:42.384054 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.384328 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:42.384353 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.384454 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:15:42.384623 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:42.384755 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:15:42.384870 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.113 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/id_rsa Username:docker}
I0729 23:15:42.469528 29396 ssh_runner.go:195] Run: cat /etc/os-release
I0729 23:15:42.473609 29396 info.go:137] Remote host: Buildroot 2023.02.9
I0729 23:15:42.473629 29396 filesync.go:126] Scanning /home/jenkins/minikube-integration/19347-12221/.minikube/addons for local assets ...
I0729 23:15:42.473689 29396 filesync.go:126] Scanning /home/jenkins/minikube-integration/19347-12221/.minikube/files for local assets ...
I0729 23:15:42.473790 29396 filesync.go:149] local asset: /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem -> 194112.pem in /etc/ssl/certs
I0729 23:15:42.473805 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem -> /etc/ssl/certs/194112.pem
I0729 23:15:42.473926 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I0729 23:15:42.483711 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem --> /etc/ssl/certs/194112.pem (1708 bytes)
I0729 23:15:42.506318 29396 start.go:296] duration metric: took 124.51393ms for postStartSetup
I0729 23:15:42.506367 29396 main.go:141] libmachine: (ha-238496) Calling .GetConfigRaw
I0729 23:15:42.506923 29396 main.go:141] libmachine: (ha-238496) Calling .GetIP
I0729 23:15:42.509333 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.509632 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:42.509667 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.509802 29396 profile.go:143] Saving config to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/config.json ...
I0729 23:15:42.509962 29396 start.go:128] duration metric: took 26.429970449s to createHost
I0729 23:15:42.509982 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:15:42.511816 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.512118 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:42.512140 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.512277 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:15:42.512422 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:42.512578 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:42.512697 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:15:42.512842 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:15:42.513014 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.113 22 <nil> <nil>}
I0729 23:15:42.513033 29396 main.go:141] libmachine: About to run SSH command:
date +%!s(MISSING).%!N(MISSING)
I0729 23:15:42.623228 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: 1722294942.603042995
I0729 23:15:42.623249 29396 fix.go:216] guest clock: 1722294942.603042995
I0729 23:15:42.623257 29396 fix.go:229] Guest: 2024-07-29 23:15:42.603042995 +0000 UTC Remote: 2024-07-29 23:15:42.509972352 +0000 UTC m=+26.531920231 (delta=93.070643ms)
I0729 23:15:42.623290 29396 fix.go:200] guest clock delta is within tolerance: 93.070643ms
I0729 23:15:42.623295 29396 start.go:83] releasing machines lock for "ha-238496", held for 26.543370229s
I0729 23:15:42.623314 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:15:42.623578 29396 main.go:141] libmachine: (ha-238496) Calling .GetIP
I0729 23:15:42.625968 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.626314 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:42.626342 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.626428 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:15:42.626934 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:15:42.627089 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:15:42.627166 29396 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0729 23:15:42.627222 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:15:42.627317 29396 ssh_runner.go:195] Run: cat /version.json
I0729 23:15:42.627365 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:15:42.629771 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.630108 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:42.630134 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.630156 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.630291 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:15:42.630453 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:42.630511 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:42.630532 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:42.630615 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:15:42.630681 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:15:42.630757 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.113 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/id_rsa Username:docker}
I0729 23:15:42.630826 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:15:42.630932 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:15:42.631082 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.113 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/id_rsa Username:docker}
I0729 23:15:42.730752 29396 ssh_runner.go:195] Run: systemctl --version
I0729 23:15:42.736524 29396 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
W0729 23:15:42.742007 29396 cni.go:209] loopback cni configuration skipped: "/etc/cni/net.d/*loopback.conf*" not found
I0729 23:15:42.742057 29396 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%!p(MISSING), " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0729 23:15:42.759615 29396 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist] bridge cni config(s)
I0729 23:15:42.759637 29396 start.go:495] detecting cgroup driver to use...
I0729 23:15:42.759746 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0729 23:15:42.777633 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.9"|' /etc/containerd/config.toml"
I0729 23:15:42.788127 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0729 23:15:42.798493 29396 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0729 23:15:42.798537 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0729 23:15:42.808963 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0729 23:15:42.819514 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0729 23:15:42.829904 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0729 23:15:42.840528 29396 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0729 23:15:42.850970 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0729 23:15:42.861324 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0729 23:15:42.871772 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0729 23:15:42.882086 29396 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0729 23:15:42.891615 29396 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0729 23:15:42.901032 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:15:43.009973 29396 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0729 23:15:43.033091 29396 start.go:495] detecting cgroup driver to use...
I0729 23:15:43.033181 29396 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0729 23:15:43.048969 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I0729 23:15:43.062883 29396 ssh_runner.go:195] Run: sudo systemctl stop -f containerd
I0729 23:15:43.078870 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I0729 23:15:43.091552 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0729 23:15:43.104157 29396 ssh_runner.go:195] Run: sudo systemctl stop -f crio
I0729 23:15:43.136301 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0729 23:15:43.149567 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0729 23:15:43.168655 29396 ssh_runner.go:195] Run: which cri-dockerd
I0729 23:15:43.172499 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0729 23:15:43.182004 29396 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (189 bytes)
I0729 23:15:43.199138 29396 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0729 23:15:43.325878 29396 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0729 23:15:43.453009 29396 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0729 23:15:43.453127 29396 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0729 23:15:43.470991 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:15:43.581853 29396 ssh_runner.go:195] Run: sudo systemctl restart docker
I0729 23:15:45.932928 29396 ssh_runner.go:235] Completed: sudo systemctl restart docker: (2.351034324s)
I0729 23:15:45.933006 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0729 23:15:45.946192 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0729 23:15:45.959363 29396 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0729 23:15:46.071000 29396 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0729 23:15:46.191364 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:15:46.323471 29396 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0729 23:15:46.341239 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0729 23:15:46.354892 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:15:46.471610 29396 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0729 23:15:46.552045 29396 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0729 23:15:46.552118 29396 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0729 23:15:46.557982 29396 start.go:563] Will wait 60s for crictl version
I0729 23:15:46.558053 29396 ssh_runner.go:195] Run: which crictl
I0729 23:15:46.562207 29396 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0729 23:15:46.599562 29396 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.1.0
RuntimeApiVersion: v1
I0729 23:15:46.599682 29396 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0729 23:15:46.627294 29396 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0729 23:15:46.653596 29396 out.go:204] * Preparing Kubernetes v1.30.3 on Docker 27.1.0 ...
I0729 23:15:46.653641 29396 main.go:141] libmachine: (ha-238496) Calling .GetIP
I0729 23:15:46.655975 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:46.656328 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:15:46.656356 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:15:46.656605 29396 ssh_runner.go:195] Run: grep 192.168.39.1 host.minikube.internal$ /etc/hosts
I0729 23:15:46.660789 29396 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.39.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0729 23:15:46.674267 29396 kubeadm.go:883] updating cluster {Name:ha-238496 KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube-builds/iso/19319/minikube-v1.33.1-1721690939-19319-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721902582-19326@sha256:540fb5dc7f38be17ff5276a38dfe6c8a4b1d9ba1c27c62244e6eebd7e37696e7 Memory:2200 CPUs:2 DiskSize:20000 Driver:kvm2 HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.30.3 Cl
usterName:ha-238496 Namespace:default APIServerHAVIP:192.168.39.254 APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.39.113 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:true ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort
:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0729 23:15:46.674374 29396 preload.go:131] Checking if preload exists for k8s version v1.30.3 and runtime docker
I0729 23:15:46.674414 29396 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0729 23:15:46.691432 29396 docker.go:685] Got preloaded images:
I0729 23:15:46.691458 29396 docker.go:691] registry.k8s.io/kube-apiserver:v1.30.3 wasn't preloaded
I0729 23:15:46.691514 29396 ssh_runner.go:195] Run: sudo cat /var/lib/docker/image/overlay2/repositories.json
I0729 23:15:46.701601 29396 ssh_runner.go:195] Run: which lz4
I0729 23:15:46.705691 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.30.3-docker-overlay2-amd64.tar.lz4 -> /preloaded.tar.lz4
I0729 23:15:46.705795 29396 ssh_runner.go:195] Run: stat -c "%!s(MISSING) %!y(MISSING)" /preloaded.tar.lz4
I0729 23:15:46.709989 29396 ssh_runner.go:352] existence check for /preloaded.tar.lz4: stat -c "%!s(MISSING) %!y(MISSING)" /preloaded.tar.lz4: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/preloaded.tar.lz4': No such file or directory
I0729 23:15:46.710023 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.30.3-docker-overlay2-amd64.tar.lz4 --> /preloaded.tar.lz4 (359612007 bytes)
I0729 23:15:48.030772 29396 docker.go:649] duration metric: took 1.325008008s to copy over tarball
I0729 23:15:48.030851 29396 ssh_runner.go:195] Run: sudo tar --xattrs --xattrs-include security.capability -I lz4 -C /var -xf /preloaded.tar.lz4
I0729 23:15:49.909865 29396 ssh_runner.go:235] Completed: sudo tar --xattrs --xattrs-include security.capability -I lz4 -C /var -xf /preloaded.tar.lz4: (1.878989077s)
I0729 23:15:49.909896 29396 ssh_runner.go:146] rm: /preloaded.tar.lz4
I0729 23:15:49.946556 29396 ssh_runner.go:195] Run: sudo cat /var/lib/docker/image/overlay2/repositories.json
I0729 23:15:49.958041 29396 ssh_runner.go:362] scp memory --> /var/lib/docker/image/overlay2/repositories.json (2630 bytes)
I0729 23:15:49.976704 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:15:50.094931 29396 ssh_runner.go:195] Run: sudo systemctl restart docker
I0729 23:15:53.045588 29396 ssh_runner.go:235] Completed: sudo systemctl restart docker: (2.950621533s)
I0729 23:15:53.045678 29396 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0729 23:15:53.065699 29396 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.30.3
registry.k8s.io/kube-controller-manager:v1.30.3
registry.k8s.io/kube-scheduler:v1.30.3
registry.k8s.io/kube-proxy:v1.30.3
registry.k8s.io/etcd:3.5.12-0
registry.k8s.io/coredns/coredns:v1.11.1
registry.k8s.io/pause:3.9
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0729 23:15:53.065726 29396 cache_images.go:84] Images are preloaded, skipping loading
I0729 23:15:53.065753 29396 kubeadm.go:934] updating node { 192.168.39.113 8443 v1.30.3 docker true true} ...
I0729 23:15:53.065875 29396 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.30.3/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=ha-238496 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.39.113
[Install]
config:
{KubernetesVersion:v1.30.3 ClusterName:ha-238496 Namespace:default APIServerHAVIP:192.168.39.254 APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0729 23:15:53.065948 29396 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0729 23:15:53.124463 29396 cni.go:84] Creating CNI manager for ""
I0729 23:15:53.124487 29396 cni.go:136] multinode detected (1 nodes found), recommending kindnet
I0729 23:15:53.124500 29396 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0729 23:15:53.124531 29396 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.39.113 APIServerPort:8443 KubernetesVersion:v1.30.3 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:ha-238496 NodeName:ha-238496 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.39.113"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.39.113 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuberne
tes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0729 23:15:53.124724 29396 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.39.113
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "ha-238496"
kubeletExtraArgs:
node-ip: 192.168.39.113
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.39.113"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.30.3
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%!"(MISSING)
nodefs.inodesFree: "0%!"(MISSING)
imagefs.available: "0%!"(MISSING)
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0729 23:15:53.124755 29396 kube-vip.go:115] generating kube-vip config ...
I0729 23:15:53.124798 29396 ssh_runner.go:195] Run: sudo sh -c "modprobe --all ip_vs ip_vs_rr ip_vs_wrr ip_vs_sh nf_conntrack"
I0729 23:15:53.140148 29396 kube-vip.go:167] auto-enabling control-plane load-balancing in kube-vip
I0729 23:15:53.140256 29396 kube-vip.go:137] kube-vip config:
apiVersion: v1
kind: Pod
metadata:
creationTimestamp: null
name: kube-vip
namespace: kube-system
spec:
containers:
- args:
- manager
env:
- name: vip_arp
value: "true"
- name: port
value: "8443"
- name: vip_nodename
valueFrom:
fieldRef:
fieldPath: spec.nodeName
- name: vip_interface
value: eth0
- name: vip_cidr
value: "32"
- name: dns_mode
value: first
- name: cp_enable
value: "true"
- name: cp_namespace
value: kube-system
- name: vip_leaderelection
value: "true"
- name: vip_leasename
value: plndr-cp-lock
- name: vip_leaseduration
value: "5"
- name: vip_renewdeadline
value: "3"
- name: vip_retryperiod
value: "1"
- name: address
value: 192.168.39.254
- name: prometheus_server
value: :2112
- name : lb_enable
value: "true"
- name: lb_port
value: "8443"
image: ghcr.io/kube-vip/kube-vip:v0.8.0
imagePullPolicy: IfNotPresent
name: kube-vip
resources: {}
securityContext:
capabilities:
add:
- NET_ADMIN
- NET_RAW
volumeMounts:
- mountPath: /etc/kubernetes/admin.conf
name: kubeconfig
hostAliases:
- hostnames:
- kubernetes
ip: 127.0.0.1
hostNetwork: true
volumes:
- hostPath:
path: "/etc/kubernetes/super-admin.conf"
name: kubeconfig
status: {}
I0729 23:15:53.140321 29396 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.30.3
I0729 23:15:53.150652 29396 binaries.go:44] Found k8s binaries, skipping transfer
I0729 23:15:53.150754 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube /etc/kubernetes/manifests
I0729 23:15:53.160855 29396 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (310 bytes)
I0729 23:15:53.178094 29396 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0729 23:15:53.195033 29396 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2157 bytes)
I0729 23:15:53.211624 29396 ssh_runner.go:362] scp memory --> /etc/kubernetes/manifests/kube-vip.yaml (1447 bytes)
I0729 23:15:53.227725 29396 ssh_runner.go:195] Run: grep 192.168.39.254 control-plane.minikube.internal$ /etc/hosts
I0729 23:15:53.231464 29396 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.39.254 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0729 23:15:53.243680 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:15:53.359728 29396 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0729 23:15:53.378242 29396 certs.go:68] Setting up /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496 for IP: 192.168.39.113
I0729 23:15:53.378263 29396 certs.go:194] generating shared ca certs ...
I0729 23:15:53.378278 29396 certs.go:226] acquiring lock for ca certs: {Name:mk651b4a346cb6b65a98f292d471b5ea2ee1b039 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:15:53.378432 29396 certs.go:235] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.key
I0729 23:15:53.378498 29396 certs.go:235] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.key
I0729 23:15:53.378511 29396 certs.go:256] generating profile certs ...
I0729 23:15:53.378560 29396 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.key
I0729 23:15:53.378574 29396 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.crt with IP's: []
I0729 23:15:53.439745 29396 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.crt ...
I0729 23:15:53.439770 29396 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.crt: {Name:mk3680a79602e99b9ae91e80b8b2de160b5edb69 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:15:53.439957 29396 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.key ...
I0729 23:15:53.439970 29396 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.key: {Name:mk0e13c40ecb7926570f1b67b7773d1f6d768c18 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:15:53.440072 29396 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key.baf0f774
I0729 23:15:53.440088 29396 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt.baf0f774 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.39.113 192.168.39.254]
I0729 23:15:53.572007 29396 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt.baf0f774 ...
I0729 23:15:53.572036 29396 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt.baf0f774: {Name:mke75c2770be3b25f9eacede6606aa30a2dd64eb Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:15:53.572192 29396 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key.baf0f774 ...
I0729 23:15:53.572204 29396 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key.baf0f774: {Name:mke89460e234eb62312d64b4c7839272bd34a2fb Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:15:53.572269 29396 certs.go:381] copying /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt.baf0f774 -> /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt
I0729 23:15:53.572358 29396 certs.go:385] copying /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key.baf0f774 -> /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key
I0729 23:15:53.572414 29396 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.key
I0729 23:15:53.572429 29396 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.crt with IP's: []
I0729 23:15:53.966977 29396 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.crt ...
I0729 23:15:53.967010 29396 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.crt: {Name:mk18942f8388406e75b85f575e1d984b1dcf1e12 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:15:53.967185 29396 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.key ...
I0729 23:15:53.967198 29396 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.key: {Name:mkf68b432adec4f2b6ef250568be4da083135a86 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:15:53.967262 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt -> /var/lib/minikube/certs/ca.crt
I0729 23:15:53.967278 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.key -> /var/lib/minikube/certs/ca.key
I0729 23:15:53.967289 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.crt -> /var/lib/minikube/certs/proxy-client-ca.crt
I0729 23:15:53.967302 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.key -> /var/lib/minikube/certs/proxy-client-ca.key
I0729 23:15:53.967314 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt -> /var/lib/minikube/certs/apiserver.crt
I0729 23:15:53.967326 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key -> /var/lib/minikube/certs/apiserver.key
I0729 23:15:53.967349 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.crt -> /var/lib/minikube/certs/proxy-client.crt
I0729 23:15:53.967361 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.key -> /var/lib/minikube/certs/proxy-client.key
I0729 23:15:53.967411 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/19411.pem (1338 bytes)
W0729 23:15:53.967442 29396 certs.go:480] ignoring /home/jenkins/minikube-integration/19347-12221/.minikube/certs/19411_empty.pem, impossibly tiny 0 bytes
I0729 23:15:53.967450 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca-key.pem (1679 bytes)
I0729 23:15:53.967476 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem (1078 bytes)
I0729 23:15:53.967497 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem (1123 bytes)
I0729 23:15:53.967520 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/key.pem (1675 bytes)
I0729 23:15:53.967590 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem (1708 bytes)
I0729 23:15:53.967618 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem -> /usr/share/ca-certificates/194112.pem
I0729 23:15:53.967638 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt -> /usr/share/ca-certificates/minikubeCA.pem
I0729 23:15:53.967650 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/19411.pem -> /usr/share/ca-certificates/19411.pem
I0729 23:15:53.968167 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0729 23:15:53.998671 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I0729 23:15:54.024684 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0729 23:15:54.050581 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I0729 23:15:54.075445 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1424 bytes)
I0729 23:15:54.100891 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0729 23:15:54.125857 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0729 23:15:54.153786 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0729 23:15:54.178024 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem --> /usr/share/ca-certificates/194112.pem (1708 bytes)
I0729 23:15:54.202191 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0729 23:15:54.234228 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/certs/19411.pem --> /usr/share/ca-certificates/19411.pem (1338 bytes)
I0729 23:15:54.259155 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0729 23:15:54.279935 29396 ssh_runner.go:195] Run: openssl version
I0729 23:15:54.286143 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0729 23:15:54.297792 29396 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0729 23:15:54.302641 29396 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Jul 29 23:03 /usr/share/ca-certificates/minikubeCA.pem
I0729 23:15:54.302708 29396 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0729 23:15:54.309137 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0729 23:15:54.320807 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/19411.pem && ln -fs /usr/share/ca-certificates/19411.pem /etc/ssl/certs/19411.pem"
I0729 23:15:54.332599 29396 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/19411.pem
I0729 23:15:54.337569 29396 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 Jul 29 23:11 /usr/share/ca-certificates/19411.pem
I0729 23:15:54.337630 29396 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/19411.pem
I0729 23:15:54.343919 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/19411.pem /etc/ssl/certs/51391683.0"
I0729 23:15:54.355853 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/194112.pem && ln -fs /usr/share/ca-certificates/194112.pem /etc/ssl/certs/194112.pem"
I0729 23:15:54.367547 29396 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/194112.pem
I0729 23:15:54.372348 29396 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 Jul 29 23:11 /usr/share/ca-certificates/194112.pem
I0729 23:15:54.372401 29396 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/194112.pem
I0729 23:15:54.378585 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/194112.pem /etc/ssl/certs/3ec20f2e.0"
I0729 23:15:54.389969 29396 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0729 23:15:54.394421 29396 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0729 23:15:54.394477 29396 kubeadm.go:392] StartCluster: {Name:ha-238496 KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube-builds/iso/19319/minikube-v1.33.1-1721690939-19319-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721902582-19326@sha256:540fb5dc7f38be17ff5276a38dfe6c8a4b1d9ba1c27c62244e6eebd7e37696e7 Memory:2200 CPUs:2 DiskSize:20000 Driver:kvm2 HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.30.3 Clust
erName:ha-238496 Namespace:default APIServerHAVIP:192.168.39.254 APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.39.113 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:true ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0
MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0729 23:15:54.394575 29396 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0729 23:15:54.415998 29396 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0729 23:15:54.426580 29396 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0729 23:15:54.436898 29396 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0729 23:15:54.447198 29396 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0729 23:15:54.447220 29396 kubeadm.go:157] found existing configuration files:
I0729 23:15:54.447262 29396 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0729 23:15:54.456798 29396 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0729 23:15:54.456855 29396 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0729 23:15:54.467435 29396 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0729 23:15:54.477135 29396 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0729 23:15:54.477189 29396 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0729 23:15:54.487320 29396 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0729 23:15:54.497272 29396 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0729 23:15:54.497330 29396 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0729 23:15:54.507122 29396 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0729 23:15:54.516249 29396 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0729 23:15:54.516304 29396 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0729 23:15:54.526031 29396 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.30.3:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem"
I0729 23:15:54.762624 29396 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0729 23:16:06.821168 29396 kubeadm.go:310] [init] Using Kubernetes version: v1.30.3
I0729 23:16:06.821243 29396 kubeadm.go:310] [preflight] Running pre-flight checks
I0729 23:16:06.821329 29396 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0729 23:16:06.821461 29396 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0729 23:16:06.821619 29396 kubeadm.go:310] [preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
I0729 23:16:06.821719 29396 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0729 23:16:06.823115 29396 out.go:204] - Generating certificates and keys ...
I0729 23:16:06.823208 29396 kubeadm.go:310] [certs] Using existing ca certificate authority
I0729 23:16:06.823294 29396 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0729 23:16:06.823360 29396 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0729 23:16:06.823424 29396 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0729 23:16:06.823504 29396 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0729 23:16:06.823547 29396 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0729 23:16:06.823597 29396 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0729 23:16:06.823701 29396 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [ha-238496 localhost] and IPs [192.168.39.113 127.0.0.1 ::1]
I0729 23:16:06.823767 29396 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0729 23:16:06.823882 29396 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [ha-238496 localhost] and IPs [192.168.39.113 127.0.0.1 ::1]
I0729 23:16:06.823941 29396 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0729 23:16:06.823993 29396 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0729 23:16:06.824033 29396 kubeadm.go:310] [certs] Generating "sa" key and public key
I0729 23:16:06.824082 29396 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0729 23:16:06.824128 29396 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0729 23:16:06.824175 29396 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0729 23:16:06.824232 29396 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0729 23:16:06.824285 29396 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0729 23:16:06.824356 29396 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0729 23:16:06.824461 29396 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0729 23:16:06.824544 29396 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0729 23:16:06.825608 29396 out.go:204] - Booting up control plane ...
I0729 23:16:06.825690 29396 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0729 23:16:06.825768 29396 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0729 23:16:06.825823 29396 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0729 23:16:06.825908 29396 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0729 23:16:06.825995 29396 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0729 23:16:06.826045 29396 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0729 23:16:06.826170 29396 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0729 23:16:06.826238 29396 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet. This can take up to 4m0s
I0729 23:16:06.826287 29396 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 501.29539ms
I0729 23:16:06.826345 29396 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0729 23:16:06.826398 29396 kubeadm.go:310] [api-check] The API server is healthy after 6.502060506s
I0729 23:16:06.826538 29396 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0729 23:16:06.826742 29396 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0729 23:16:06.826829 29396 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0729 23:16:06.826991 29396 kubeadm.go:310] [mark-control-plane] Marking the node ha-238496 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0729 23:16:06.827039 29396 kubeadm.go:310] [bootstrap-token] Using token: 7wps4h.r9ujmgl0smjas4sr
I0729 23:16:06.829264 29396 out.go:204] - Configuring RBAC rules ...
I0729 23:16:06.829363 29396 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0729 23:16:06.829433 29396 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0729 23:16:06.829579 29396 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0729 23:16:06.829845 29396 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0729 23:16:06.829982 29396 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0729 23:16:06.830088 29396 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0729 23:16:06.830228 29396 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0729 23:16:06.830291 29396 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0729 23:16:06.830365 29396 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0729 23:16:06.830381 29396 kubeadm.go:310]
I0729 23:16:06.830443 29396 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0729 23:16:06.830450 29396 kubeadm.go:310]
I0729 23:16:06.830522 29396 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0729 23:16:06.830529 29396 kubeadm.go:310]
I0729 23:16:06.830558 29396 kubeadm.go:310] mkdir -p $HOME/.kube
I0729 23:16:06.830627 29396 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0729 23:16:06.830705 29396 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0729 23:16:06.830715 29396 kubeadm.go:310]
I0729 23:16:06.830769 29396 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0729 23:16:06.830776 29396 kubeadm.go:310]
I0729 23:16:06.830814 29396 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0729 23:16:06.830820 29396 kubeadm.go:310]
I0729 23:16:06.830861 29396 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0729 23:16:06.830952 29396 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0729 23:16:06.831064 29396 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0729 23:16:06.831078 29396 kubeadm.go:310]
I0729 23:16:06.831177 29396 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0729 23:16:06.831277 29396 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0729 23:16:06.831285 29396 kubeadm.go:310]
I0729 23:16:06.831394 29396 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 7wps4h.r9ujmgl0smjas4sr \
I0729 23:16:06.831519 29396 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:da4124175dbd4d7966590c68bf3c2627d9fda969ee89096732ee7fd4a463dd4a \
I0729 23:16:06.831550 29396 kubeadm.go:310] --control-plane
I0729 23:16:06.831564 29396 kubeadm.go:310]
I0729 23:16:06.831681 29396 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0729 23:16:06.831690 29396 kubeadm.go:310]
I0729 23:16:06.831778 29396 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 7wps4h.r9ujmgl0smjas4sr \
I0729 23:16:06.831917 29396 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:da4124175dbd4d7966590c68bf3c2627d9fda969ee89096732ee7fd4a463dd4a
I0729 23:16:06.831929 29396 cni.go:84] Creating CNI manager for ""
I0729 23:16:06.831939 29396 cni.go:136] multinode detected (1 nodes found), recommending kindnet
I0729 23:16:06.833425 29396 out.go:177] * Configuring CNI (Container Networking Interface) ...
I0729 23:16:06.834468 29396 ssh_runner.go:195] Run: stat /opt/cni/bin/portmap
I0729 23:16:06.840942 29396 cni.go:182] applying CNI manifest using /var/lib/minikube/binaries/v1.30.3/kubectl ...
I0729 23:16:06.840959 29396 ssh_runner.go:362] scp memory --> /var/tmp/minikube/cni.yaml (2438 bytes)
I0729 23:16:06.859861 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml
I0729 23:16:07.234421 29396 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0729 23:16:07.234510 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:07.234551 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes ha-238496 minikube.k8s.io/updated_at=2024_07_29T23_16_07_0700 minikube.k8s.io/version=v1.33.1 minikube.k8s.io/commit=b13baeaf4895dcc6a8c5d0ab64a27ff86dff4ae3 minikube.k8s.io/name=ha-238496 minikube.k8s.io/primary=true
I0729 23:16:07.385204 29396 ops.go:34] apiserver oom_adj: -16
I0729 23:16:07.385357 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:07.885788 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:08.385838 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:08.885780 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:09.386269 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:09.886246 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:10.385618 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:10.886204 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:11.385914 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:11.885786 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:12.385963 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:12.886207 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:13.386172 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:13.886192 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:14.386169 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:14.885669 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:15.385587 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:15.886172 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:16.385515 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:16.886237 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:17.385987 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:17.885376 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:18.386069 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0729 23:16:18.536657 29396 kubeadm.go:1113] duration metric: took 11.302209455s to wait for elevateKubeSystemPrivileges
I0729 23:16:18.536711 29396 kubeadm.go:394] duration metric: took 24.142230868s to StartCluster
I0729 23:16:18.536734 29396 settings.go:142] acquiring lock: {Name:mk17e1ab030b9e2103931d17b9ef30ea797bca5d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:16:18.536844 29396 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19347-12221/kubeconfig
I0729 23:16:18.537722 29396 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19347-12221/kubeconfig: {Name:mkcda89ba949a6d5877faacf6424d912f9a0066b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:16:18.538196 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.30.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0729 23:16:18.538213 29396 start.go:233] HA (multi-control plane) cluster: will skip waiting for primary control-plane node &{Name: IP:192.168.39.113 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}
I0729 23:16:18.538246 29396 start.go:241] waiting for startup goroutines ...
I0729 23:16:18.538255 29396 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false helm-tiller:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:false volcano:false volumesnapshots:false yakd:false]
I0729 23:16:18.538327 29396 addons.go:69] Setting storage-provisioner=true in profile "ha-238496"
I0729 23:16:18.538338 29396 addons.go:69] Setting default-storageclass=true in profile "ha-238496"
I0729 23:16:18.538360 29396 addons.go:234] Setting addon storage-provisioner=true in "ha-238496"
I0729 23:16:18.538387 29396 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "ha-238496"
I0729 23:16:18.538395 29396 host.go:66] Checking if "ha-238496" exists ...
I0729 23:16:18.538427 29396 config.go:182] Loaded profile config "ha-238496": Driver=kvm2, ContainerRuntime=docker, KubernetesVersion=v1.30.3
I0729 23:16:18.538830 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:16:18.538858 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:16:18.538862 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:16:18.538893 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:16:18.554878 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:38993
I0729 23:16:18.554965 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:44877
I0729 23:16:18.555294 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:16:18.555345 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:16:18.555846 29396 main.go:141] libmachine: Using API Version 1
I0729 23:16:18.555872 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:16:18.555977 29396 main.go:141] libmachine: Using API Version 1
I0729 23:16:18.555996 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:16:18.556210 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:16:18.556388 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:16:18.556437 29396 main.go:141] libmachine: (ha-238496) Calling .GetState
I0729 23:16:18.556923 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:16:18.556956 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:16:18.558602 29396 loader.go:395] Config loaded from file: /home/jenkins/minikube-integration/19347-12221/kubeconfig
I0729 23:16:18.558947 29396 kapi.go:59] client config for ha-238496: &rest.Config{Host:"https://192.168.39.254:8443", APIPath:"", ContentConfig:rest.ContentConfig{AcceptContentTypes:"", ContentType:"", GroupVersion:(*schema.GroupVersion)(nil), NegotiatedSerializer:runtime.NegotiatedSerializer(nil)}, Username:"", Password:"", BearerToken:"", BearerTokenFile:"", Impersonate:rest.ImpersonationConfig{UserName:"", UID:"", Groups:[]string(nil), Extra:map[string][]string(nil)}, AuthProvider:<nil>, AuthConfigPersister:rest.AuthProviderConfigPersister(nil), ExecProvider:<nil>, TLSClientConfig:rest.sanitizedTLSClientConfig{Insecure:false, ServerName:"", CertFile:"/home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.crt", KeyFile:"/home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.key", CAFile:"/home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt", CertData:[]uint8(nil), KeyData:[]uint8(nil), CAData:[]uint8(nil), NextProtos:[]string(nil)
}, UserAgent:"", DisableCompression:false, Transport:http.RoundTripper(nil), WrapTransport:(transport.WrapperFunc)(0x1d02de0), QPS:0, Burst:0, RateLimiter:flowcontrol.RateLimiter(nil), WarningHandler:rest.WarningHandler(nil), Timeout:0, Dial:(func(context.Context, string, string) (net.Conn, error))(nil), Proxy:(func(*http.Request) (*url.URL, error))(nil)}
I0729 23:16:18.559506 29396 cert_rotation.go:137] Starting client certificate rotation controller
I0729 23:16:18.559711 29396 addons.go:234] Setting addon default-storageclass=true in "ha-238496"
I0729 23:16:18.559753 29396 host.go:66] Checking if "ha-238496" exists ...
I0729 23:16:18.560124 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:16:18.560157 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:16:18.573011 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:35147
I0729 23:16:18.573508 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:16:18.574041 29396 main.go:141] libmachine: Using API Version 1
I0729 23:16:18.574064 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:16:18.574405 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:16:18.574600 29396 main.go:141] libmachine: (ha-238496) Calling .GetState
I0729 23:16:18.574987 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:36847
I0729 23:16:18.575303 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:16:18.575763 29396 main.go:141] libmachine: Using API Version 1
I0729 23:16:18.575790 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:16:18.576170 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:16:18.576559 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:16:18.576789 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:16:18.576832 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:16:18.578472 29396 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0729 23:16:18.579868 29396 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0729 23:16:18.579890 29396 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0729 23:16:18.579912 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:16:18.583493 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:16:18.583967 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:16:18.584002 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:16:18.584142 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:16:18.584345 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:16:18.584537 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:16:18.584699 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.113 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/id_rsa Username:docker}
I0729 23:16:18.592521 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:39903
I0729 23:16:18.592939 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:16:18.593360 29396 main.go:141] libmachine: Using API Version 1
I0729 23:16:18.593380 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:16:18.593709 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:16:18.593904 29396 main.go:141] libmachine: (ha-238496) Calling .GetState
I0729 23:16:18.595554 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:16:18.595801 29396 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0729 23:16:18.595819 29396 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0729 23:16:18.595836 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:16:18.598908 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:16:18.599463 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:16:18.599487 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:16:18.599655 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:16:18.599843 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:16:18.600010 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:16:18.600114 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.113 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/id_rsa Username:docker}
I0729 23:16:18.672751 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.30.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.39.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.30.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0729 23:16:18.744494 29396 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.3/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0729 23:16:18.790001 29396 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.3/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0729 23:16:19.024537 29396 start.go:971] {"host.minikube.internal": 192.168.39.1} host record injected into CoreDNS's ConfigMap
I0729 23:16:19.336491 29396 main.go:141] libmachine: Making call to close driver server
I0729 23:16:19.336518 29396 main.go:141] libmachine: (ha-238496) Calling .Close
I0729 23:16:19.336625 29396 main.go:141] libmachine: Making call to close driver server
I0729 23:16:19.336646 29396 main.go:141] libmachine: (ha-238496) Calling .Close
I0729 23:16:19.336824 29396 main.go:141] libmachine: Successfully made call to close driver server
I0729 23:16:19.336842 29396 main.go:141] libmachine: Making call to close connection to plugin binary
I0729 23:16:19.336850 29396 main.go:141] libmachine: Making call to close driver server
I0729 23:16:19.336857 29396 main.go:141] libmachine: (ha-238496) Calling .Close
I0729 23:16:19.336884 29396 main.go:141] libmachine: Successfully made call to close driver server
I0729 23:16:19.336905 29396 main.go:141] libmachine: Making call to close connection to plugin binary
I0729 23:16:19.336938 29396 main.go:141] libmachine: Making call to close driver server
I0729 23:16:19.336950 29396 main.go:141] libmachine: (ha-238496) Calling .Close
I0729 23:16:19.337040 29396 main.go:141] libmachine: Successfully made call to close driver server
I0729 23:16:19.337052 29396 main.go:141] libmachine: Making call to close connection to plugin binary
I0729 23:16:19.337159 29396 round_trippers.go:463] GET https://192.168.39.254:8443/apis/storage.k8s.io/v1/storageclasses
I0729 23:16:19.337165 29396 round_trippers.go:469] Request Headers:
I0729 23:16:19.337176 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:16:19.337182 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:16:19.337298 29396 main.go:141] libmachine: (ha-238496) DBG | Closing plugin on server side
I0729 23:16:19.337357 29396 main.go:141] libmachine: Successfully made call to close driver server
I0729 23:16:19.337379 29396 main.go:141] libmachine: Making call to close connection to plugin binary
I0729 23:16:19.358829 29396 round_trippers.go:574] Response Status: 200 OK in 21 milliseconds
I0729 23:16:19.359997 29396 round_trippers.go:463] PUT https://192.168.39.254:8443/apis/storage.k8s.io/v1/storageclasses/standard
I0729 23:16:19.360011 29396 round_trippers.go:469] Request Headers:
I0729 23:16:19.360019 29396 round_trippers.go:473] Content-Type: application/json
I0729 23:16:19.360024 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:16:19.360029 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:16:19.365893 29396 round_trippers.go:574] Response Status: 200 OK in 5 milliseconds
I0729 23:16:19.366053 29396 main.go:141] libmachine: Making call to close driver server
I0729 23:16:19.366070 29396 main.go:141] libmachine: (ha-238496) Calling .Close
I0729 23:16:19.366340 29396 main.go:141] libmachine: Successfully made call to close driver server
I0729 23:16:19.366357 29396 main.go:141] libmachine: Making call to close connection to plugin binary
I0729 23:16:19.368339 29396 out.go:177] * Enabled addons: storage-provisioner, default-storageclass
I0729 23:16:19.369578 29396 addons.go:510] duration metric: took 831.320462ms for enable addons: enabled=[storage-provisioner default-storageclass]
I0729 23:16:19.369612 29396 start.go:246] waiting for cluster config update ...
I0729 23:16:19.369627 29396 start.go:255] writing updated cluster config ...
I0729 23:16:19.371229 29396 out.go:177]
I0729 23:16:19.372506 29396 config.go:182] Loaded profile config "ha-238496": Driver=kvm2, ContainerRuntime=docker, KubernetesVersion=v1.30.3
I0729 23:16:19.372569 29396 profile.go:143] Saving config to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/config.json ...
I0729 23:16:19.374125 29396 out.go:177] * Starting "ha-238496-m02" control-plane node in "ha-238496" cluster
I0729 23:16:19.375196 29396 preload.go:131] Checking if preload exists for k8s version v1.30.3 and runtime docker
I0729 23:16:19.375218 29396 cache.go:56] Caching tarball of preloaded images
I0729 23:16:19.375315 29396 preload.go:172] Found /home/jenkins/minikube-integration/19347-12221/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.30.3-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0729 23:16:19.375329 29396 cache.go:59] Finished verifying existence of preloaded tar for v1.30.3 on docker
I0729 23:16:19.375392 29396 profile.go:143] Saving config to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/config.json ...
I0729 23:16:19.375562 29396 start.go:360] acquireMachinesLock for ha-238496-m02: {Name:mk79fbc287386032c39e512567e9786663e657a9 Clock:{} Delay:500ms Timeout:13m0s Cancel:<nil>}
I0729 23:16:19.375605 29396 start.go:364] duration metric: took 24.398µs to acquireMachinesLock for "ha-238496-m02"
I0729 23:16:19.375625 29396 start.go:93] Provisioning new machine with config: &{Name:ha-238496 KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube-builds/iso/19319/minikube-v1.33.1-1721690939-19319-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721902582-19326@sha256:540fb5dc7f38be17ff5276a38dfe6c8a4b1d9ba1c27c62244e6eebd7e37696e7 Memory:2200 CPUs:2 DiskSize:20000 Driver:kvm2 HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{Kubernete
sVersion:v1.30.3 ClusterName:ha-238496 Namespace:default APIServerHAVIP:192.168.39.254 APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.39.113 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true} {Name:m02 IP: Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:true ExtraDisks
:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name:m02 IP: Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}
I0729 23:16:19.375695 29396 start.go:125] createHost starting for "m02" (driver="kvm2")
I0729 23:16:19.377072 29396 out.go:204] * Creating kvm2 VM (CPUs=2, Memory=2200MB, Disk=20000MB) ...
I0729 23:16:19.377141 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:16:19.377163 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:16:19.391837 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:36879
I0729 23:16:19.392321 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:16:19.392882 29396 main.go:141] libmachine: Using API Version 1
I0729 23:16:19.392905 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:16:19.393200 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:16:19.393416 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetMachineName
I0729 23:16:19.393559 29396 main.go:141] libmachine: (ha-238496-m02) Calling .DriverName
I0729 23:16:19.393723 29396 start.go:159] libmachine.API.Create for "ha-238496" (driver="kvm2")
I0729 23:16:19.393744 29396 client.go:168] LocalClient.Create starting
I0729 23:16:19.393768 29396 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem
I0729 23:16:19.393798 29396 main.go:141] libmachine: Decoding PEM data...
I0729 23:16:19.393811 29396 main.go:141] libmachine: Parsing certificate...
I0729 23:16:19.393860 29396 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem
I0729 23:16:19.393878 29396 main.go:141] libmachine: Decoding PEM data...
I0729 23:16:19.393887 29396 main.go:141] libmachine: Parsing certificate...
I0729 23:16:19.393902 29396 main.go:141] libmachine: Running pre-create checks...
I0729 23:16:19.393909 29396 main.go:141] libmachine: (ha-238496-m02) Calling .PreCreateCheck
I0729 23:16:19.394061 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetConfigRaw
I0729 23:16:19.394404 29396 main.go:141] libmachine: Creating machine...
I0729 23:16:19.394417 29396 main.go:141] libmachine: (ha-238496-m02) Calling .Create
I0729 23:16:19.394564 29396 main.go:141] libmachine: (ha-238496-m02) Creating KVM machine...
I0729 23:16:19.395809 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found existing default KVM network
I0729 23:16:19.395989 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found existing private KVM network mk-ha-238496
I0729 23:16:19.396145 29396 main.go:141] libmachine: (ha-238496-m02) Setting up store path in /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02 ...
I0729 23:16:19.396168 29396 main.go:141] libmachine: (ha-238496-m02) Building disk image from file:///home/jenkins/minikube-integration/19347-12221/.minikube/cache/iso/amd64/minikube-v1.33.1-1721690939-19319-amd64.iso
I0729 23:16:19.396233 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:19.396145 29816 common.go:145] Making disk image using store path: /home/jenkins/minikube-integration/19347-12221/.minikube
I0729 23:16:19.396350 29396 main.go:141] libmachine: (ha-238496-m02) Downloading /home/jenkins/minikube-integration/19347-12221/.minikube/cache/boot2docker.iso from file:///home/jenkins/minikube-integration/19347-12221/.minikube/cache/iso/amd64/minikube-v1.33.1-1721690939-19319-amd64.iso...
I0729 23:16:19.672295 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:19.672179 29816 common.go:152] Creating ssh key: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02/id_rsa...
I0729 23:16:19.900328 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:19.900217 29816 common.go:158] Creating raw disk image: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02/ha-238496-m02.rawdisk...
I0729 23:16:19.900362 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Writing magic tar header
I0729 23:16:19.900378 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Writing SSH key tar header
I0729 23:16:19.900389 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:19.900329 29816 common.go:172] Fixing permissions on /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02 ...
I0729 23:16:19.900474 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Checking permissions on dir: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02
I0729 23:16:19.900496 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Checking permissions on dir: /home/jenkins/minikube-integration/19347-12221/.minikube/machines
I0729 23:16:19.900510 29396 main.go:141] libmachine: (ha-238496-m02) Setting executable bit set on /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02 (perms=drwx------)
I0729 23:16:19.900557 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Checking permissions on dir: /home/jenkins/minikube-integration/19347-12221/.minikube
I0729 23:16:19.900590 29396 main.go:141] libmachine: (ha-238496-m02) Setting executable bit set on /home/jenkins/minikube-integration/19347-12221/.minikube/machines (perms=drwxr-xr-x)
I0729 23:16:19.900603 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Checking permissions on dir: /home/jenkins/minikube-integration/19347-12221
I0729 23:16:19.900616 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Checking permissions on dir: /home/jenkins/minikube-integration
I0729 23:16:19.900627 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Checking permissions on dir: /home/jenkins
I0729 23:16:19.900638 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Checking permissions on dir: /home
I0729 23:16:19.900646 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Skipping /home - not owner
I0729 23:16:19.900664 29396 main.go:141] libmachine: (ha-238496-m02) Setting executable bit set on /home/jenkins/minikube-integration/19347-12221/.minikube (perms=drwxr-xr-x)
I0729 23:16:19.900681 29396 main.go:141] libmachine: (ha-238496-m02) Setting executable bit set on /home/jenkins/minikube-integration/19347-12221 (perms=drwxrwxr-x)
I0729 23:16:19.900695 29396 main.go:141] libmachine: (ha-238496-m02) Setting executable bit set on /home/jenkins/minikube-integration (perms=drwxrwxr-x)
I0729 23:16:19.900706 29396 main.go:141] libmachine: (ha-238496-m02) Setting executable bit set on /home/jenkins (perms=drwxr-xr-x)
I0729 23:16:19.900738 29396 main.go:141] libmachine: (ha-238496-m02) Creating domain...
I0729 23:16:19.901507 29396 main.go:141] libmachine: (ha-238496-m02) define libvirt domain using xml:
I0729 23:16:19.901527 29396 main.go:141] libmachine: (ha-238496-m02) <domain type='kvm'>
I0729 23:16:19.901553 29396 main.go:141] libmachine: (ha-238496-m02) <name>ha-238496-m02</name>
I0729 23:16:19.901576 29396 main.go:141] libmachine: (ha-238496-m02) <memory unit='MiB'>2200</memory>
I0729 23:16:19.901588 29396 main.go:141] libmachine: (ha-238496-m02) <vcpu>2</vcpu>
I0729 23:16:19.901596 29396 main.go:141] libmachine: (ha-238496-m02) <features>
I0729 23:16:19.901614 29396 main.go:141] libmachine: (ha-238496-m02) <acpi/>
I0729 23:16:19.901625 29396 main.go:141] libmachine: (ha-238496-m02) <apic/>
I0729 23:16:19.901631 29396 main.go:141] libmachine: (ha-238496-m02) <pae/>
I0729 23:16:19.901637 29396 main.go:141] libmachine: (ha-238496-m02)
I0729 23:16:19.901642 29396 main.go:141] libmachine: (ha-238496-m02) </features>
I0729 23:16:19.901647 29396 main.go:141] libmachine: (ha-238496-m02) <cpu mode='host-passthrough'>
I0729 23:16:19.901652 29396 main.go:141] libmachine: (ha-238496-m02)
I0729 23:16:19.901656 29396 main.go:141] libmachine: (ha-238496-m02) </cpu>
I0729 23:16:19.901661 29396 main.go:141] libmachine: (ha-238496-m02) <os>
I0729 23:16:19.901666 29396 main.go:141] libmachine: (ha-238496-m02) <type>hvm</type>
I0729 23:16:19.901671 29396 main.go:141] libmachine: (ha-238496-m02) <boot dev='cdrom'/>
I0729 23:16:19.901676 29396 main.go:141] libmachine: (ha-238496-m02) <boot dev='hd'/>
I0729 23:16:19.901681 29396 main.go:141] libmachine: (ha-238496-m02) <bootmenu enable='no'/>
I0729 23:16:19.901685 29396 main.go:141] libmachine: (ha-238496-m02) </os>
I0729 23:16:19.901690 29396 main.go:141] libmachine: (ha-238496-m02) <devices>
I0729 23:16:19.901695 29396 main.go:141] libmachine: (ha-238496-m02) <disk type='file' device='cdrom'>
I0729 23:16:19.901703 29396 main.go:141] libmachine: (ha-238496-m02) <source file='/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02/boot2docker.iso'/>
I0729 23:16:19.901712 29396 main.go:141] libmachine: (ha-238496-m02) <target dev='hdc' bus='scsi'/>
I0729 23:16:19.901717 29396 main.go:141] libmachine: (ha-238496-m02) <readonly/>
I0729 23:16:19.901721 29396 main.go:141] libmachine: (ha-238496-m02) </disk>
I0729 23:16:19.901730 29396 main.go:141] libmachine: (ha-238496-m02) <disk type='file' device='disk'>
I0729 23:16:19.901735 29396 main.go:141] libmachine: (ha-238496-m02) <driver name='qemu' type='raw' cache='default' io='threads' />
I0729 23:16:19.901743 29396 main.go:141] libmachine: (ha-238496-m02) <source file='/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02/ha-238496-m02.rawdisk'/>
I0729 23:16:19.901747 29396 main.go:141] libmachine: (ha-238496-m02) <target dev='hda' bus='virtio'/>
I0729 23:16:19.901752 29396 main.go:141] libmachine: (ha-238496-m02) </disk>
I0729 23:16:19.901756 29396 main.go:141] libmachine: (ha-238496-m02) <interface type='network'>
I0729 23:16:19.901761 29396 main.go:141] libmachine: (ha-238496-m02) <source network='mk-ha-238496'/>
I0729 23:16:19.901766 29396 main.go:141] libmachine: (ha-238496-m02) <model type='virtio'/>
I0729 23:16:19.901771 29396 main.go:141] libmachine: (ha-238496-m02) </interface>
I0729 23:16:19.901775 29396 main.go:141] libmachine: (ha-238496-m02) <interface type='network'>
I0729 23:16:19.901782 29396 main.go:141] libmachine: (ha-238496-m02) <source network='default'/>
I0729 23:16:19.901793 29396 main.go:141] libmachine: (ha-238496-m02) <model type='virtio'/>
I0729 23:16:19.901801 29396 main.go:141] libmachine: (ha-238496-m02) </interface>
I0729 23:16:19.901808 29396 main.go:141] libmachine: (ha-238496-m02) <serial type='pty'>
I0729 23:16:19.901816 29396 main.go:141] libmachine: (ha-238496-m02) <target port='0'/>
I0729 23:16:19.901821 29396 main.go:141] libmachine: (ha-238496-m02) </serial>
I0729 23:16:19.901826 29396 main.go:141] libmachine: (ha-238496-m02) <console type='pty'>
I0729 23:16:19.901831 29396 main.go:141] libmachine: (ha-238496-m02) <target type='serial' port='0'/>
I0729 23:16:19.901836 29396 main.go:141] libmachine: (ha-238496-m02) </console>
I0729 23:16:19.901840 29396 main.go:141] libmachine: (ha-238496-m02) <rng model='virtio'>
I0729 23:16:19.901850 29396 main.go:141] libmachine: (ha-238496-m02) <backend model='random'>/dev/random</backend>
I0729 23:16:19.901854 29396 main.go:141] libmachine: (ha-238496-m02) </rng>
I0729 23:16:19.901858 29396 main.go:141] libmachine: (ha-238496-m02)
I0729 23:16:19.901864 29396 main.go:141] libmachine: (ha-238496-m02)
I0729 23:16:19.901872 29396 main.go:141] libmachine: (ha-238496-m02) </devices>
I0729 23:16:19.901878 29396 main.go:141] libmachine: (ha-238496-m02) </domain>
I0729 23:16:19.901892 29396 main.go:141] libmachine: (ha-238496-m02)
I0729 23:16:19.908728 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:2a:9b:e1 in network default
I0729 23:16:19.909244 29396 main.go:141] libmachine: (ha-238496-m02) Ensuring networks are active...
I0729 23:16:19.909288 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:19.910028 29396 main.go:141] libmachine: (ha-238496-m02) Ensuring network default is active
I0729 23:16:19.910449 29396 main.go:141] libmachine: (ha-238496-m02) Ensuring network mk-ha-238496 is active
I0729 23:16:19.910816 29396 main.go:141] libmachine: (ha-238496-m02) Getting domain xml...
I0729 23:16:19.911569 29396 main.go:141] libmachine: (ha-238496-m02) Creating domain...
I0729 23:16:21.122194 29396 main.go:141] libmachine: (ha-238496-m02) Waiting to get IP...
I0729 23:16:21.123143 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:21.123515 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:21.123536 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:21.123490 29816 retry.go:31] will retry after 210.542983ms: waiting for machine to come up
I0729 23:16:21.335988 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:21.336378 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:21.336408 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:21.336354 29816 retry.go:31] will retry after 291.309738ms: waiting for machine to come up
I0729 23:16:21.628887 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:21.629304 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:21.629331 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:21.629272 29816 retry.go:31] will retry after 460.631998ms: waiting for machine to come up
I0729 23:16:22.092069 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:22.092559 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:22.092587 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:22.092518 29816 retry.go:31] will retry after 374.861132ms: waiting for machine to come up
I0729 23:16:22.469027 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:22.469518 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:22.469546 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:22.469478 29816 retry.go:31] will retry after 604.947482ms: waiting for machine to come up
I0729 23:16:23.076290 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:23.076777 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:23.076802 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:23.076724 29816 retry.go:31] will retry after 806.329173ms: waiting for machine to come up
I0729 23:16:23.884406 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:23.884817 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:23.884846 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:23.884808 29816 retry.go:31] will retry after 803.379339ms: waiting for machine to come up
I0729 23:16:24.689636 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:24.690078 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:24.690108 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:24.690039 29816 retry.go:31] will retry after 1.280518832s: waiting for machine to come up
I0729 23:16:25.972490 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:25.972999 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:25.973030 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:25.972949 29816 retry.go:31] will retry after 1.549162667s: waiting for machine to come up
I0729 23:16:27.523891 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:27.524360 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:27.524384 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:27.524311 29816 retry.go:31] will retry after 1.581798428s: waiting for machine to come up
I0729 23:16:29.107873 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:29.108320 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:29.108343 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:29.108277 29816 retry.go:31] will retry after 1.968794912s: waiting for machine to come up
I0729 23:16:31.078415 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:31.078906 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:31.078936 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:31.078857 29816 retry.go:31] will retry after 2.58499227s: waiting for machine to come up
I0729 23:16:33.665171 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:33.665534 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:33.665559 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:33.665508 29816 retry.go:31] will retry after 4.074814902s: waiting for machine to come up
I0729 23:16:37.743773 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:37.744169 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find current IP address of domain ha-238496-m02 in network mk-ha-238496
I0729 23:16:37.744199 29396 main.go:141] libmachine: (ha-238496-m02) DBG | I0729 23:16:37.744119 29816 retry.go:31] will retry after 4.097801489s: waiting for machine to come up
I0729 23:16:41.845420 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:41.845844 29396 main.go:141] libmachine: (ha-238496-m02) Found IP for machine: 192.168.39.226
I0729 23:16:41.845892 29396 main.go:141] libmachine: (ha-238496-m02) Reserving static IP address...
I0729 23:16:41.845907 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has current primary IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:41.846301 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find host DHCP lease matching {name: "ha-238496-m02", mac: "52:54:00:15:f5:ca", ip: "192.168.39.226"} in network mk-ha-238496
I0729 23:16:41.924834 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Getting to WaitForSSH function...
I0729 23:16:41.924862 29396 main.go:141] libmachine: (ha-238496-m02) Reserved static IP address: 192.168.39.226
I0729 23:16:41.924875 29396 main.go:141] libmachine: (ha-238496-m02) Waiting for SSH to be available...
I0729 23:16:41.926998 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:41.927263 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496
I0729 23:16:41.927302 29396 main.go:141] libmachine: (ha-238496-m02) DBG | unable to find defined IP address of network mk-ha-238496 interface with MAC address 52:54:00:15:f5:ca
I0729 23:16:41.927427 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Using SSH client type: external
I0729 23:16:41.927457 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Using SSH private key: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02/id_rsa (-rw-------)
I0729 23:16:41.927501 29396 main.go:141] libmachine: (ha-238496-m02) DBG | &{[-F /dev/null -o ConnectionAttempts=3 -o ConnectTimeout=10 -o ControlMaster=no -o ControlPath=none -o LogLevel=quiet -o PasswordAuthentication=no -o ServerAliveInterval=60 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null docker@ -o IdentitiesOnly=yes -i /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02/id_rsa -p 22] /usr/bin/ssh <nil>}
I0729 23:16:41.927520 29396 main.go:141] libmachine: (ha-238496-m02) DBG | About to run SSH command:
I0729 23:16:41.927537 29396 main.go:141] libmachine: (ha-238496-m02) DBG | exit 0
I0729 23:16:41.931361 29396 main.go:141] libmachine: (ha-238496-m02) DBG | SSH cmd err, output: exit status 255:
I0729 23:16:41.931392 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Error getting ssh command 'exit 0' : ssh command error:
I0729 23:16:41.931404 29396 main.go:141] libmachine: (ha-238496-m02) DBG | command : exit 0
I0729 23:16:41.931412 29396 main.go:141] libmachine: (ha-238496-m02) DBG | err : exit status 255
I0729 23:16:41.931422 29396 main.go:141] libmachine: (ha-238496-m02) DBG | output :
I0729 23:16:44.931808 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Getting to WaitForSSH function...
I0729 23:16:44.934574 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:44.934999 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:44.935029 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:44.935130 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Using SSH client type: external
I0729 23:16:44.935160 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Using SSH private key: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02/id_rsa (-rw-------)
I0729 23:16:44.935214 29396 main.go:141] libmachine: (ha-238496-m02) DBG | &{[-F /dev/null -o ConnectionAttempts=3 -o ConnectTimeout=10 -o ControlMaster=no -o ControlPath=none -o LogLevel=quiet -o PasswordAuthentication=no -o ServerAliveInterval=60 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null docker@192.168.39.226 -o IdentitiesOnly=yes -i /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02/id_rsa -p 22] /usr/bin/ssh <nil>}
I0729 23:16:44.935254 29396 main.go:141] libmachine: (ha-238496-m02) DBG | About to run SSH command:
I0729 23:16:44.935283 29396 main.go:141] libmachine: (ha-238496-m02) DBG | exit 0
I0729 23:16:45.058642 29396 main.go:141] libmachine: (ha-238496-m02) DBG | SSH cmd err, output: <nil>:
I0729 23:16:45.058949 29396 main.go:141] libmachine: (ha-238496-m02) KVM machine creation complete!
I0729 23:16:45.059234 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetConfigRaw
I0729 23:16:45.059788 29396 main.go:141] libmachine: (ha-238496-m02) Calling .DriverName
I0729 23:16:45.059995 29396 main.go:141] libmachine: (ha-238496-m02) Calling .DriverName
I0729 23:16:45.060119 29396 main.go:141] libmachine: Waiting for machine to be running, this may take a few minutes...
I0729 23:16:45.060135 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetState
I0729 23:16:45.061649 29396 main.go:141] libmachine: Detecting operating system of created instance...
I0729 23:16:45.061663 29396 main.go:141] libmachine: Waiting for SSH to be available...
I0729 23:16:45.061668 29396 main.go:141] libmachine: Getting to WaitForSSH function...
I0729 23:16:45.061674 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHHostname
I0729 23:16:45.064202 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.064562 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:45.064588 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.064706 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHPort
I0729 23:16:45.064883 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:45.065030 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:45.065154 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHUsername
I0729 23:16:45.065291 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:16:45.065502 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.226 22 <nil> <nil>}
I0729 23:16:45.065516 29396 main.go:141] libmachine: About to run SSH command:
exit 0
I0729 23:16:45.166119 29396 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0729 23:16:45.166141 29396 main.go:141] libmachine: Detecting the provisioner...
I0729 23:16:45.166149 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHHostname
I0729 23:16:45.168854 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.169205 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:45.169234 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.169346 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHPort
I0729 23:16:45.169578 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:45.169731 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:45.169874 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHUsername
I0729 23:16:45.170021 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:16:45.170177 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.226 22 <nil> <nil>}
I0729 23:16:45.170196 29396 main.go:141] libmachine: About to run SSH command:
cat /etc/os-release
I0729 23:16:45.275789 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: NAME=Buildroot
VERSION=2023.02.9-dirty
ID=buildroot
VERSION_ID=2023.02.9
PRETTY_NAME="Buildroot 2023.02.9"
I0729 23:16:45.275875 29396 main.go:141] libmachine: found compatible host: buildroot
I0729 23:16:45.275890 29396 main.go:141] libmachine: Provisioning with buildroot...
I0729 23:16:45.275904 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetMachineName
I0729 23:16:45.276160 29396 buildroot.go:166] provisioning hostname "ha-238496-m02"
I0729 23:16:45.276180 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetMachineName
I0729 23:16:45.276336 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHHostname
I0729 23:16:45.278898 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.279211 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:45.279232 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.279339 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHPort
I0729 23:16:45.279533 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:45.279682 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:45.279823 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHUsername
I0729 23:16:45.280005 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:16:45.280214 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.226 22 <nil> <nil>}
I0729 23:16:45.280232 29396 main.go:141] libmachine: About to run SSH command:
sudo hostname ha-238496-m02 && echo "ha-238496-m02" | sudo tee /etc/hostname
I0729 23:16:45.398012 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: ha-238496-m02
I0729 23:16:45.398033 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHHostname
I0729 23:16:45.400557 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.400983 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:45.401017 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.401186 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHPort
I0729 23:16:45.401376 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:45.401565 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:45.401720 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHUsername
I0729 23:16:45.401909 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:16:45.402091 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.226 22 <nil> <nil>}
I0729 23:16:45.402113 29396 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sha-238496-m02' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 ha-238496-m02/g' /etc/hosts;
else
echo '127.0.1.1 ha-238496-m02' | sudo tee -a /etc/hosts;
fi
fi
I0729 23:16:45.511766 29396 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0729 23:16:45.511797 29396 buildroot.go:172] set auth options {CertDir:/home/jenkins/minikube-integration/19347-12221/.minikube CaCertPath:/home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19347-12221/.minikube}
I0729 23:16:45.511811 29396 buildroot.go:174] setting up certificates
I0729 23:16:45.511819 29396 provision.go:84] configureAuth start
I0729 23:16:45.511827 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetMachineName
I0729 23:16:45.512093 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetIP
I0729 23:16:45.514793 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.515163 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:45.515185 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.515367 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHHostname
I0729 23:16:45.517391 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.517686 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:45.517713 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.517853 29396 provision.go:143] copyHostCerts
I0729 23:16:45.517886 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/key.pem -> /home/jenkins/minikube-integration/19347-12221/.minikube/key.pem
I0729 23:16:45.517926 29396 exec_runner.go:144] found /home/jenkins/minikube-integration/19347-12221/.minikube/key.pem, removing ...
I0729 23:16:45.517936 29396 exec_runner.go:203] rm: /home/jenkins/minikube-integration/19347-12221/.minikube/key.pem
I0729 23:16:45.518017 29396 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19347-12221/.minikube/key.pem (1675 bytes)
I0729 23:16:45.518105 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem -> /home/jenkins/minikube-integration/19347-12221/.minikube/ca.pem
I0729 23:16:45.518130 29396 exec_runner.go:144] found /home/jenkins/minikube-integration/19347-12221/.minikube/ca.pem, removing ...
I0729 23:16:45.518139 29396 exec_runner.go:203] rm: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.pem
I0729 23:16:45.518175 29396 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19347-12221/.minikube/ca.pem (1078 bytes)
I0729 23:16:45.518231 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem -> /home/jenkins/minikube-integration/19347-12221/.minikube/cert.pem
I0729 23:16:45.518256 29396 exec_runner.go:144] found /home/jenkins/minikube-integration/19347-12221/.minikube/cert.pem, removing ...
I0729 23:16:45.518264 29396 exec_runner.go:203] rm: /home/jenkins/minikube-integration/19347-12221/.minikube/cert.pem
I0729 23:16:45.518293 29396 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19347-12221/.minikube/cert.pem (1123 bytes)
I0729 23:16:45.518359 29396 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca-key.pem org=jenkins.ha-238496-m02 san=[127.0.0.1 192.168.39.226 ha-238496-m02 localhost minikube]
I0729 23:16:45.694511 29396 provision.go:177] copyRemoteCerts
I0729 23:16:45.694577 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0729 23:16:45.694604 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHHostname
I0729 23:16:45.697058 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.697459 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:45.697485 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.697707 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHPort
I0729 23:16:45.697873 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:45.698021 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHUsername
I0729 23:16:45.698146 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.226 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02/id_rsa Username:docker}
I0729 23:16:45.781576 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem -> /etc/docker/ca.pem
I0729 23:16:45.781638 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0729 23:16:45.808822 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server.pem -> /etc/docker/server.pem
I0729 23:16:45.808896 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0729 23:16:45.834554 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server-key.pem -> /etc/docker/server-key.pem
I0729 23:16:45.834635 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
I0729 23:16:45.859821 29396 provision.go:87] duration metric: took 347.99007ms to configureAuth
I0729 23:16:45.859854 29396 buildroot.go:189] setting minikube options for container-runtime
I0729 23:16:45.860021 29396 config.go:182] Loaded profile config "ha-238496": Driver=kvm2, ContainerRuntime=docker, KubernetesVersion=v1.30.3
I0729 23:16:45.860044 29396 main.go:141] libmachine: (ha-238496-m02) Calling .DriverName
I0729 23:16:45.860315 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHHostname
I0729 23:16:45.862902 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.863203 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:45.863232 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.863404 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHPort
I0729 23:16:45.863595 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:45.863762 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:45.863928 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHUsername
I0729 23:16:45.864072 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:16:45.864287 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.226 22 <nil> <nil>}
I0729 23:16:45.864298 29396 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0729 23:16:45.968371 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: tmpfs
I0729 23:16:45.968416 29396 buildroot.go:70] root file system type: tmpfs
I0729 23:16:45.968559 29396 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0729 23:16:45.968584 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHHostname
I0729 23:16:45.971687 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.972162 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:45.972195 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:45.972355 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHPort
I0729 23:16:45.972555 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:45.972746 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:45.972878 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHUsername
I0729 23:16:45.973043 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:16:45.973244 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.226 22 <nil> <nil>}
I0729 23:16:45.973307 29396 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %!s(MISSING) "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network.target minikube-automount.service docker.socket
Requires= minikube-automount.service docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
Environment="NO_PROXY=192.168.39.113"
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=kvm2 --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0729 23:16:46.094512 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network.target minikube-automount.service docker.socket
Requires= minikube-automount.service docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
Environment=NO_PROXY=192.168.39.113
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=kvm2 --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0729 23:16:46.094543 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHHostname
I0729 23:16:46.097313 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:46.097672 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:46.097700 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:46.097930 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHPort
I0729 23:16:46.098131 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:46.098292 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:46.098423 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHUsername
I0729 23:16:46.098570 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:16:46.098775 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.226 22 <nil> <nil>}
I0729 23:16:46.098792 29396 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0729 23:16:47.867253 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: diff: can't stat '/lib/systemd/system/docker.service': No such file or directory
Created symlink /etc/systemd/system/multi-user.target.wants/docker.service → /usr/lib/systemd/system/docker.service.
I0729 23:16:47.867290 29396 main.go:141] libmachine: Checking connection to Docker...
I0729 23:16:47.867302 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetURL
I0729 23:16:47.868904 29396 main.go:141] libmachine: (ha-238496-m02) DBG | Using libvirt version 6000000
I0729 23:16:47.871705 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:47.872324 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:47.872350 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:47.872527 29396 main.go:141] libmachine: Docker is up and running!
I0729 23:16:47.872542 29396 main.go:141] libmachine: Reticulating splines...
I0729 23:16:47.872549 29396 client.go:171] duration metric: took 28.478798279s to LocalClient.Create
I0729 23:16:47.872571 29396 start.go:167] duration metric: took 28.478849364s to libmachine.API.Create "ha-238496"
I0729 23:16:47.872584 29396 start.go:293] postStartSetup for "ha-238496-m02" (driver="kvm2")
I0729 23:16:47.872596 29396 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0729 23:16:47.872614 29396 main.go:141] libmachine: (ha-238496-m02) Calling .DriverName
I0729 23:16:47.872875 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0729 23:16:47.872899 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHHostname
I0729 23:16:47.875125 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:47.875484 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:47.875511 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:47.875666 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHPort
I0729 23:16:47.875845 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:47.876016 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHUsername
I0729 23:16:47.876187 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.226 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02/id_rsa Username:docker}
I0729 23:16:47.957455 29396 ssh_runner.go:195] Run: cat /etc/os-release
I0729 23:16:47.961907 29396 info.go:137] Remote host: Buildroot 2023.02.9
I0729 23:16:47.961934 29396 filesync.go:126] Scanning /home/jenkins/minikube-integration/19347-12221/.minikube/addons for local assets ...
I0729 23:16:47.962007 29396 filesync.go:126] Scanning /home/jenkins/minikube-integration/19347-12221/.minikube/files for local assets ...
I0729 23:16:47.962103 29396 filesync.go:149] local asset: /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem -> 194112.pem in /etc/ssl/certs
I0729 23:16:47.962115 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem -> /etc/ssl/certs/194112.pem
I0729 23:16:47.962226 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I0729 23:16:47.972087 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem --> /etc/ssl/certs/194112.pem (1708 bytes)
I0729 23:16:48.000323 29396 start.go:296] duration metric: took 127.725465ms for postStartSetup
I0729 23:16:48.000380 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetConfigRaw
I0729 23:16:48.000995 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetIP
I0729 23:16:48.003937 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:48.004303 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:48.004331 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:48.004613 29396 profile.go:143] Saving config to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/config.json ...
I0729 23:16:48.004818 29396 start.go:128] duration metric: took 28.62911238s to createHost
I0729 23:16:48.004843 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHHostname
I0729 23:16:48.007143 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:48.007517 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:48.007545 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:48.007747 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHPort
I0729 23:16:48.007948 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:48.008111 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:48.008270 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHUsername
I0729 23:16:48.008453 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:16:48.008662 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.226 22 <nil> <nil>}
I0729 23:16:48.008682 29396 main.go:141] libmachine: About to run SSH command:
date +%!s(MISSING).%!N(MISSING)
I0729 23:16:48.115978 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: 1722295008.094136082
I0729 23:16:48.115999 29396 fix.go:216] guest clock: 1722295008.094136082
I0729 23:16:48.116008 29396 fix.go:229] Guest: 2024-07-29 23:16:48.094136082 +0000 UTC Remote: 2024-07-29 23:16:48.004831359 +0000 UTC m=+92.026779250 (delta=89.304723ms)
I0729 23:16:48.116030 29396 fix.go:200] guest clock delta is within tolerance: 89.304723ms
I0729 23:16:48.116037 29396 start.go:83] releasing machines lock for "ha-238496-m02", held for 28.740421202s
I0729 23:16:48.116059 29396 main.go:141] libmachine: (ha-238496-m02) Calling .DriverName
I0729 23:16:48.116307 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetIP
I0729 23:16:48.118859 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:48.119193 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:48.119227 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:48.121181 29396 out.go:177] * Found network options:
I0729 23:16:48.122795 29396 out.go:177] - NO_PROXY=192.168.39.113
W0729 23:16:48.124384 29396 proxy.go:119] fail to check proxy env: Error ip not in block
I0729 23:16:48.124408 29396 main.go:141] libmachine: (ha-238496-m02) Calling .DriverName
I0729 23:16:48.124963 29396 main.go:141] libmachine: (ha-238496-m02) Calling .DriverName
I0729 23:16:48.125136 29396 main.go:141] libmachine: (ha-238496-m02) Calling .DriverName
I0729 23:16:48.125213 29396 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0729 23:16:48.125257 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHHostname
W0729 23:16:48.125325 29396 proxy.go:119] fail to check proxy env: Error ip not in block
I0729 23:16:48.125413 29396 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0729 23:16:48.125435 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHHostname
I0729 23:16:48.127820 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:48.128046 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:48.128181 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:48.128211 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:48.128326 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHPort
I0729 23:16:48.128488 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:48.128514 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:48.128534 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:48.128641 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHUsername
I0729 23:16:48.128709 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHPort
I0729 23:16:48.128790 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.226 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02/id_rsa Username:docker}
I0729 23:16:48.128837 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHKeyPath
I0729 23:16:48.128942 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetSSHUsername
I0729 23:16:48.129104 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.226 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m02/id_rsa Username:docker}
W0729 23:16:48.205896 29396 cni.go:209] loopback cni configuration skipped: "/etc/cni/net.d/*loopback.conf*" not found
I0729 23:16:48.205956 29396 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%!p(MISSING), " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0729 23:16:48.228230 29396 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist] bridge cni config(s)
I0729 23:16:48.228252 29396 start.go:495] detecting cgroup driver to use...
I0729 23:16:48.228343 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0729 23:16:48.248044 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.9"|' /etc/containerd/config.toml"
I0729 23:16:48.261247 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0729 23:16:48.274271 29396 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0729 23:16:48.274375 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0729 23:16:48.287313 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0729 23:16:48.300525 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0729 23:16:48.313542 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0729 23:16:48.326487 29396 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0729 23:16:48.339511 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0729 23:16:48.352454 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0729 23:16:48.365423 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0729 23:16:48.378632 29396 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0729 23:16:48.390528 29396 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0729 23:16:48.402554 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:16:48.537443 29396 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0729 23:16:48.562887 29396 start.go:495] detecting cgroup driver to use...
I0729 23:16:48.562978 29396 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0729 23:16:48.583195 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I0729 23:16:48.600886 29396 ssh_runner.go:195] Run: sudo systemctl stop -f containerd
I0729 23:16:48.626610 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I0729 23:16:48.640651 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0729 23:16:48.654131 29396 ssh_runner.go:195] Run: sudo systemctl stop -f crio
I0729 23:16:48.685200 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0729 23:16:48.699226 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0729 23:16:48.719673 29396 ssh_runner.go:195] Run: which cri-dockerd
I0729 23:16:48.723763 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0729 23:16:48.733499 29396 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (189 bytes)
I0729 23:16:48.751414 29396 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0729 23:16:48.878991 29396 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0729 23:16:49.003264 29396 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0729 23:16:49.003303 29396 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0729 23:16:49.021808 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:16:49.153978 29396 ssh_runner.go:195] Run: sudo systemctl restart docker
I0729 23:16:51.526166 29396 ssh_runner.go:235] Completed: sudo systemctl restart docker: (2.372153577s)
I0729 23:16:51.526240 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0729 23:16:51.540938 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0729 23:16:51.555750 29396 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0729 23:16:51.682013 29396 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0729 23:16:51.813640 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:16:51.945396 29396 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0729 23:16:51.965941 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0729 23:16:51.983182 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:16:52.109586 29396 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0729 23:16:52.189812 29396 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0729 23:16:52.189870 29396 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0729 23:16:52.196237 29396 start.go:563] Will wait 60s for crictl version
I0729 23:16:52.196309 29396 ssh_runner.go:195] Run: which crictl
I0729 23:16:52.202212 29396 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0729 23:16:52.244282 29396 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.1.0
RuntimeApiVersion: v1
I0729 23:16:52.244342 29396 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0729 23:16:52.273830 29396 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0729 23:16:52.298731 29396 out.go:204] * Preparing Kubernetes v1.30.3 on Docker 27.1.0 ...
I0729 23:16:52.300116 29396 out.go:177] - env NO_PROXY=192.168.39.113
I0729 23:16:52.301288 29396 main.go:141] libmachine: (ha-238496-m02) Calling .GetIP
I0729 23:16:52.304199 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:52.304576 29396 main.go:141] libmachine: (ha-238496-m02) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:15:f5:ca", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:16:34 +0000 UTC Type:0 Mac:52:54:00:15:f5:ca Iaid: IPaddr:192.168.39.226 Prefix:24 Hostname:ha-238496-m02 Clientid:01:52:54:00:15:f5:ca}
I0729 23:16:52.304610 29396 main.go:141] libmachine: (ha-238496-m02) DBG | domain ha-238496-m02 has defined IP address 192.168.39.226 and MAC address 52:54:00:15:f5:ca in network mk-ha-238496
I0729 23:16:52.304844 29396 ssh_runner.go:195] Run: grep 192.168.39.1 host.minikube.internal$ /etc/hosts
I0729 23:16:52.309522 29396 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.39.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0729 23:16:52.322555 29396 mustload.go:65] Loading cluster: ha-238496
I0729 23:16:52.322770 29396 config.go:182] Loaded profile config "ha-238496": Driver=kvm2, ContainerRuntime=docker, KubernetesVersion=v1.30.3
I0729 23:16:52.323008 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:16:52.323033 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:16:52.337685 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:43277
I0729 23:16:52.338156 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:16:52.338607 29396 main.go:141] libmachine: Using API Version 1
I0729 23:16:52.338626 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:16:52.338935 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:16:52.339112 29396 main.go:141] libmachine: (ha-238496) Calling .GetState
I0729 23:16:52.340721 29396 host.go:66] Checking if "ha-238496" exists ...
I0729 23:16:52.341000 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:16:52.341020 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:16:52.355343 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:33235
I0729 23:16:52.355792 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:16:52.356236 29396 main.go:141] libmachine: Using API Version 1
I0729 23:16:52.356257 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:16:52.356549 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:16:52.356710 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:16:52.356844 29396 certs.go:68] Setting up /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496 for IP: 192.168.39.226
I0729 23:16:52.356855 29396 certs.go:194] generating shared ca certs ...
I0729 23:16:52.356874 29396 certs.go:226] acquiring lock for ca certs: {Name:mk651b4a346cb6b65a98f292d471b5ea2ee1b039 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:16:52.357018 29396 certs.go:235] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.key
I0729 23:16:52.357071 29396 certs.go:235] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.key
I0729 23:16:52.357085 29396 certs.go:256] generating profile certs ...
I0729 23:16:52.357177 29396 certs.go:359] skipping valid signed profile cert regeneration for "minikube-user": /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.key
I0729 23:16:52.357209 29396 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key.ea71b923
I0729 23:16:52.357227 29396 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt.ea71b923 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.39.113 192.168.39.226 192.168.39.254]
I0729 23:16:52.438426 29396 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt.ea71b923 ...
I0729 23:16:52.438451 29396 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt.ea71b923: {Name:mk91f223539790286183d375e09336e0661489e8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:16:52.438609 29396 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key.ea71b923 ...
I0729 23:16:52.438621 29396 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key.ea71b923: {Name:mk0395716d5db1f731970b42562c431b469c4d3e Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:16:52.438718 29396 certs.go:381] copying /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt.ea71b923 -> /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt
I0729 23:16:52.438849 29396 certs.go:385] copying /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key.ea71b923 -> /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key
I0729 23:16:52.438963 29396 certs.go:359] skipping valid signed profile cert regeneration for "aggregator": /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.key
I0729 23:16:52.438978 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt -> /var/lib/minikube/certs/ca.crt
I0729 23:16:52.438991 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.key -> /var/lib/minikube/certs/ca.key
I0729 23:16:52.439004 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.crt -> /var/lib/minikube/certs/proxy-client-ca.crt
I0729 23:16:52.439018 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.key -> /var/lib/minikube/certs/proxy-client-ca.key
I0729 23:16:52.439030 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt -> /var/lib/minikube/certs/apiserver.crt
I0729 23:16:52.439042 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key -> /var/lib/minikube/certs/apiserver.key
I0729 23:16:52.439054 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.crt -> /var/lib/minikube/certs/proxy-client.crt
I0729 23:16:52.439066 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.key -> /var/lib/minikube/certs/proxy-client.key
I0729 23:16:52.439111 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/19411.pem (1338 bytes)
W0729 23:16:52.439137 29396 certs.go:480] ignoring /home/jenkins/minikube-integration/19347-12221/.minikube/certs/19411_empty.pem, impossibly tiny 0 bytes
I0729 23:16:52.439146 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca-key.pem (1679 bytes)
I0729 23:16:52.439166 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem (1078 bytes)
I0729 23:16:52.439186 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem (1123 bytes)
I0729 23:16:52.439206 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/key.pem (1675 bytes)
I0729 23:16:52.439245 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem (1708 bytes)
I0729 23:16:52.439270 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/19411.pem -> /usr/share/ca-certificates/19411.pem
I0729 23:16:52.439283 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem -> /usr/share/ca-certificates/194112.pem
I0729 23:16:52.439296 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt -> /usr/share/ca-certificates/minikubeCA.pem
I0729 23:16:52.439324 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:16:52.442047 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:16:52.442480 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:16:52.442503 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:16:52.442682 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:16:52.442905 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:16:52.443046 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:16:52.443158 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.113 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/id_rsa Username:docker}
I0729 23:16:52.518984 29396 ssh_runner.go:195] Run: stat -c %!s(MISSING) /var/lib/minikube/certs/sa.pub
I0729 23:16:52.523859 29396 ssh_runner.go:447] scp /var/lib/minikube/certs/sa.pub --> memory (451 bytes)
I0729 23:16:52.534773 29396 ssh_runner.go:195] Run: stat -c %!s(MISSING) /var/lib/minikube/certs/sa.key
I0729 23:16:52.538885 29396 ssh_runner.go:447] scp /var/lib/minikube/certs/sa.key --> memory (1675 bytes)
I0729 23:16:52.549218 29396 ssh_runner.go:195] Run: stat -c %!s(MISSING) /var/lib/minikube/certs/front-proxy-ca.crt
I0729 23:16:52.553615 29396 ssh_runner.go:447] scp /var/lib/minikube/certs/front-proxy-ca.crt --> memory (1123 bytes)
I0729 23:16:52.564716 29396 ssh_runner.go:195] Run: stat -c %!s(MISSING) /var/lib/minikube/certs/front-proxy-ca.key
I0729 23:16:52.568669 29396 ssh_runner.go:447] scp /var/lib/minikube/certs/front-proxy-ca.key --> memory (1679 bytes)
I0729 23:16:52.579330 29396 ssh_runner.go:195] Run: stat -c %!s(MISSING) /var/lib/minikube/certs/etcd/ca.crt
I0729 23:16:52.583704 29396 ssh_runner.go:447] scp /var/lib/minikube/certs/etcd/ca.crt --> memory (1094 bytes)
I0729 23:16:52.597894 29396 ssh_runner.go:195] Run: stat -c %!s(MISSING) /var/lib/minikube/certs/etcd/ca.key
I0729 23:16:52.602188 29396 ssh_runner.go:447] scp /var/lib/minikube/certs/etcd/ca.key --> memory (1679 bytes)
I0729 23:16:52.612958 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0729 23:16:52.638256 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I0729 23:16:52.661756 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0729 23:16:52.684792 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I0729 23:16:52.707756 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1436 bytes)
I0729 23:16:52.731576 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
I0729 23:16:52.754441 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0729 23:16:52.777771 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0729 23:16:52.801003 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/certs/19411.pem --> /usr/share/ca-certificates/19411.pem (1338 bytes)
I0729 23:16:52.825430 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem --> /usr/share/ca-certificates/194112.pem (1708 bytes)
I0729 23:16:52.848979 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0729 23:16:52.873731 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/certs/sa.pub (451 bytes)
I0729 23:16:52.890569 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/certs/sa.key (1675 bytes)
I0729 23:16:52.907546 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/certs/front-proxy-ca.crt (1123 bytes)
I0729 23:16:52.924240 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/certs/front-proxy-ca.key (1679 bytes)
I0729 23:16:52.940802 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/certs/etcd/ca.crt (1094 bytes)
I0729 23:16:52.956658 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/certs/etcd/ca.key (1679 bytes)
I0729 23:16:52.972800 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (744 bytes)
I0729 23:16:52.989957 29396 ssh_runner.go:195] Run: openssl version
I0729 23:16:52.995807 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/19411.pem && ln -fs /usr/share/ca-certificates/19411.pem /etc/ssl/certs/19411.pem"
I0729 23:16:53.006586 29396 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/19411.pem
I0729 23:16:53.011272 29396 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 Jul 29 23:11 /usr/share/ca-certificates/19411.pem
I0729 23:16:53.011333 29396 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/19411.pem
I0729 23:16:53.017395 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/19411.pem /etc/ssl/certs/51391683.0"
I0729 23:16:53.029003 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/194112.pem && ln -fs /usr/share/ca-certificates/194112.pem /etc/ssl/certs/194112.pem"
I0729 23:16:53.040176 29396 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/194112.pem
I0729 23:16:53.044895 29396 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 Jul 29 23:11 /usr/share/ca-certificates/194112.pem
I0729 23:16:53.044970 29396 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/194112.pem
I0729 23:16:53.051179 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/194112.pem /etc/ssl/certs/3ec20f2e.0"
I0729 23:16:53.063199 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0729 23:16:53.074351 29396 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0729 23:16:53.078889 29396 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Jul 29 23:03 /usr/share/ca-certificates/minikubeCA.pem
I0729 23:16:53.078951 29396 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0729 23:16:53.084785 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0729 23:16:53.095633 29396 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0729 23:16:53.099717 29396 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0729 23:16:53.099765 29396 kubeadm.go:934] updating node {m02 192.168.39.226 8443 v1.30.3 docker true true} ...
I0729 23:16:53.099848 29396 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.30.3/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=ha-238496-m02 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.39.226
[Install]
config:
{KubernetesVersion:v1.30.3 ClusterName:ha-238496 Namespace:default APIServerHAVIP:192.168.39.254 APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0729 23:16:53.099872 29396 kube-vip.go:115] generating kube-vip config ...
I0729 23:16:53.099904 29396 ssh_runner.go:195] Run: sudo sh -c "modprobe --all ip_vs ip_vs_rr ip_vs_wrr ip_vs_sh nf_conntrack"
I0729 23:16:53.115368 29396 kube-vip.go:167] auto-enabling control-plane load-balancing in kube-vip
I0729 23:16:53.115456 29396 kube-vip.go:137] kube-vip config:
apiVersion: v1
kind: Pod
metadata:
creationTimestamp: null
name: kube-vip
namespace: kube-system
spec:
containers:
- args:
- manager
env:
- name: vip_arp
value: "true"
- name: port
value: "8443"
- name: vip_nodename
valueFrom:
fieldRef:
fieldPath: spec.nodeName
- name: vip_interface
value: eth0
- name: vip_cidr
value: "32"
- name: dns_mode
value: first
- name: cp_enable
value: "true"
- name: cp_namespace
value: kube-system
- name: vip_leaderelection
value: "true"
- name: vip_leasename
value: plndr-cp-lock
- name: vip_leaseduration
value: "5"
- name: vip_renewdeadline
value: "3"
- name: vip_retryperiod
value: "1"
- name: address
value: 192.168.39.254
- name: prometheus_server
value: :2112
- name : lb_enable
value: "true"
- name: lb_port
value: "8443"
image: ghcr.io/kube-vip/kube-vip:v0.8.0
imagePullPolicy: IfNotPresent
name: kube-vip
resources: {}
securityContext:
capabilities:
add:
- NET_ADMIN
- NET_RAW
volumeMounts:
- mountPath: /etc/kubernetes/admin.conf
name: kubeconfig
hostAliases:
- hostnames:
- kubernetes
ip: 127.0.0.1
hostNetwork: true
volumes:
- hostPath:
path: "/etc/kubernetes/admin.conf"
name: kubeconfig
status: {}
I0729 23:16:53.115507 29396 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.30.3
I0729 23:16:53.125267 29396 binaries.go:47] Didn't find k8s binaries: sudo ls /var/lib/minikube/binaries/v1.30.3: Process exited with status 2
stdout:
stderr:
ls: cannot access '/var/lib/minikube/binaries/v1.30.3': No such file or directory
Initiating transfer...
I0729 23:16:53.125324 29396 ssh_runner.go:195] Run: sudo mkdir -p /var/lib/minikube/binaries/v1.30.3
I0729 23:16:53.134688 29396 download.go:107] Downloading: https://dl.k8s.io/release/v1.30.3/bin/linux/amd64/kubelet?checksum=file:https://dl.k8s.io/release/v1.30.3/bin/linux/amd64/kubelet.sha256 -> /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubelet
I0729 23:16:53.134720 29396 download.go:107] Downloading: https://dl.k8s.io/release/v1.30.3/bin/linux/amd64/kubeadm?checksum=file:https://dl.k8s.io/release/v1.30.3/bin/linux/amd64/kubeadm.sha256 -> /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubeadm
I0729 23:16:53.134688 29396 binary.go:74] Not caching binary, using https://dl.k8s.io/release/v1.30.3/bin/linux/amd64/kubectl?checksum=file:https://dl.k8s.io/release/v1.30.3/bin/linux/amd64/kubectl.sha256
I0729 23:16:53.134844 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubectl -> /var/lib/minikube/binaries/v1.30.3/kubectl
I0729 23:16:53.134934 29396 ssh_runner.go:195] Run: stat -c "%!s(MISSING) %!y(MISSING)" /var/lib/minikube/binaries/v1.30.3/kubectl
I0729 23:16:53.139246 29396 ssh_runner.go:352] existence check for /var/lib/minikube/binaries/v1.30.3/kubectl: stat -c "%!s(MISSING) %!y(MISSING)" /var/lib/minikube/binaries/v1.30.3/kubectl: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/binaries/v1.30.3/kubectl': No such file or directory
I0729 23:16:53.139276 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubectl --> /var/lib/minikube/binaries/v1.30.3/kubectl (51454104 bytes)
I0729 23:17:00.942779 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubeadm -> /var/lib/minikube/binaries/v1.30.3/kubeadm
I0729 23:17:00.942853 29396 ssh_runner.go:195] Run: stat -c "%!s(MISSING) %!y(MISSING)" /var/lib/minikube/binaries/v1.30.3/kubeadm
I0729 23:17:00.947933 29396 ssh_runner.go:352] existence check for /var/lib/minikube/binaries/v1.30.3/kubeadm: stat -c "%!s(MISSING) %!y(MISSING)" /var/lib/minikube/binaries/v1.30.3/kubeadm: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/binaries/v1.30.3/kubeadm': No such file or directory
I0729 23:17:00.947959 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubeadm --> /var/lib/minikube/binaries/v1.30.3/kubeadm (50249880 bytes)
I0729 23:17:05.522395 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0729 23:17:05.538088 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubelet -> /var/lib/minikube/binaries/v1.30.3/kubelet
I0729 23:17:05.538185 29396 ssh_runner.go:195] Run: stat -c "%!s(MISSING) %!y(MISSING)" /var/lib/minikube/binaries/v1.30.3/kubelet
I0729 23:17:05.542643 29396 ssh_runner.go:352] existence check for /var/lib/minikube/binaries/v1.30.3/kubelet: stat -c "%!s(MISSING) %!y(MISSING)" /var/lib/minikube/binaries/v1.30.3/kubelet: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/binaries/v1.30.3/kubelet': No such file or directory
I0729 23:17:05.542679 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubelet --> /var/lib/minikube/binaries/v1.30.3/kubelet (100125080 bytes)
I0729 23:17:05.957616 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /etc/kubernetes/manifests
I0729 23:17:05.967186 29396 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (314 bytes)
I0729 23:17:05.984857 29396 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0729 23:17:06.004743 29396 ssh_runner.go:362] scp memory --> /etc/kubernetes/manifests/kube-vip.yaml (1441 bytes)
I0729 23:17:06.022735 29396 ssh_runner.go:195] Run: grep 192.168.39.254 control-plane.minikube.internal$ /etc/hosts
I0729 23:17:06.026830 29396 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.39.254 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0729 23:17:06.039823 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:17:06.168431 29396 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0729 23:17:06.194567 29396 host.go:66] Checking if "ha-238496" exists ...
I0729 23:17:06.195059 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:17:06.195120 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:17:06.210453 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:46269
I0729 23:17:06.210933 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:17:06.211515 29396 main.go:141] libmachine: Using API Version 1
I0729 23:17:06.211544 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:17:06.211877 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:17:06.212057 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:17:06.212216 29396 start.go:317] joinCluster: &{Name:ha-238496 KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube-builds/iso/19319/minikube-v1.33.1-1721690939-19319-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721902582-19326@sha256:540fb5dc7f38be17ff5276a38dfe6c8a4b1d9ba1c27c62244e6eebd7e37696e7 Memory:2200 CPUs:2 DiskSize:20000 Driver:kvm2 HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.30.3 Cluster
Name:ha-238496 Namespace:default APIServerHAVIP:192.168.39.254 APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.39.113 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true} {Name:m02 IP:192.168.39.226 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[default-storageclass:true storage-provisioner:true] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:true ExtraDisks:0 CertExp
iration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0729 23:17:06.212386 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.30.3:$PATH" kubeadm token create --print-join-command --ttl=0"
I0729 23:17:06.212409 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:17:06.215839 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:17:06.216287 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:17:06.216321 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:17:06.216483 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:17:06.216659 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:17:06.216837 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:17:06.216979 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.113 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/id_rsa Username:docker}
I0729 23:17:06.426321 29396 start.go:343] trying to join control-plane node "m02" to cluster: &{Name:m02 IP:192.168.39.226 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}
I0729 23:17:06.426369 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.30.3:$PATH" kubeadm join control-plane.minikube.internal:8443 --token z9n94n.tldcwy00xngrv8wr --discovery-token-ca-cert-hash sha256:da4124175dbd4d7966590c68bf3c2627d9fda969ee89096732ee7fd4a463dd4a --ignore-preflight-errors=all --cri-socket unix:///var/run/cri-dockerd.sock --node-name=ha-238496-m02 --control-plane --apiserver-advertise-address=192.168.39.226 --apiserver-bind-port=8443"
I0729 23:17:29.346176 29396 ssh_runner.go:235] Completed: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.30.3:$PATH" kubeadm join control-plane.minikube.internal:8443 --token z9n94n.tldcwy00xngrv8wr --discovery-token-ca-cert-hash sha256:da4124175dbd4d7966590c68bf3c2627d9fda969ee89096732ee7fd4a463dd4a --ignore-preflight-errors=all --cri-socket unix:///var/run/cri-dockerd.sock --node-name=ha-238496-m02 --control-plane --apiserver-advertise-address=192.168.39.226 --apiserver-bind-port=8443": (22.919780887s)
I0729 23:17:29.346215 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo systemctl daemon-reload && sudo systemctl enable kubelet && sudo systemctl start kubelet"
I0729 23:17:29.882187 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes ha-238496-m02 minikube.k8s.io/updated_at=2024_07_29T23_17_29_0700 minikube.k8s.io/version=v1.33.1 minikube.k8s.io/commit=b13baeaf4895dcc6a8c5d0ab64a27ff86dff4ae3 minikube.k8s.io/name=ha-238496 minikube.k8s.io/primary=false
I0729 23:17:30.030849 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig taint nodes ha-238496-m02 node-role.kubernetes.io/control-plane:NoSchedule-
I0729 23:17:30.159567 29396 start.go:319] duration metric: took 23.947350002s to joinCluster
I0729 23:17:30.159629 29396 start.go:235] Will wait 6m0s for node &{Name:m02 IP:192.168.39.226 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}
I0729 23:17:30.159910 29396 config.go:182] Loaded profile config "ha-238496": Driver=kvm2, ContainerRuntime=docker, KubernetesVersion=v1.30.3
I0729 23:17:30.161150 29396 out.go:177] * Verifying Kubernetes components...
I0729 23:17:30.162759 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:17:30.452541 29396 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0729 23:17:30.485926 29396 loader.go:395] Config loaded from file: /home/jenkins/minikube-integration/19347-12221/kubeconfig
I0729 23:17:30.486245 29396 kapi.go:59] client config for ha-238496: &rest.Config{Host:"https://192.168.39.254:8443", APIPath:"", ContentConfig:rest.ContentConfig{AcceptContentTypes:"", ContentType:"", GroupVersion:(*schema.GroupVersion)(nil), NegotiatedSerializer:runtime.NegotiatedSerializer(nil)}, Username:"", Password:"", BearerToken:"", BearerTokenFile:"", Impersonate:rest.ImpersonationConfig{UserName:"", UID:"", Groups:[]string(nil), Extra:map[string][]string(nil)}, AuthProvider:<nil>, AuthConfigPersister:rest.AuthProviderConfigPersister(nil), ExecProvider:<nil>, TLSClientConfig:rest.sanitizedTLSClientConfig{Insecure:false, ServerName:"", CertFile:"/home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.crt", KeyFile:"/home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.key", CAFile:"/home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt", CertData:[]uint8(nil), KeyData:[]uint8(nil), CAData:[]uint8(nil), NextProtos:[]string(nil)
}, UserAgent:"", DisableCompression:false, Transport:http.RoundTripper(nil), WrapTransport:(transport.WrapperFunc)(0x1d02de0), QPS:0, Burst:0, RateLimiter:flowcontrol.RateLimiter(nil), WarningHandler:rest.WarningHandler(nil), Timeout:0, Dial:(func(context.Context, string, string) (net.Conn, error))(nil), Proxy:(func(*http.Request) (*url.URL, error))(nil)}
W0729 23:17:30.486335 29396 kubeadm.go:483] Overriding stale ClientConfig host https://192.168.39.254:8443 with https://192.168.39.113:8443
I0729 23:17:30.486630 29396 node_ready.go:35] waiting up to 6m0s for node "ha-238496-m02" to be "Ready" ...
I0729 23:17:30.486743 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:30.486754 29396 round_trippers.go:469] Request Headers:
I0729 23:17:30.486765 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:30.486774 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:30.498725 29396 round_trippers.go:574] Response Status: 200 OK in 11 milliseconds
I0729 23:17:30.986874 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:30.986899 29396 round_trippers.go:469] Request Headers:
I0729 23:17:30.986910 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:30.986916 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:31.037211 29396 round_trippers.go:574] Response Status: 200 OK in 50 milliseconds
I0729 23:17:31.487111 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:31.487138 29396 round_trippers.go:469] Request Headers:
I0729 23:17:31.487149 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:31.487158 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:31.505988 29396 round_trippers.go:574] Response Status: 200 OK in 18 milliseconds
I0729 23:17:31.986880 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:31.986903 29396 round_trippers.go:469] Request Headers:
I0729 23:17:31.986914 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:31.986923 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:31.995565 29396 round_trippers.go:574] Response Status: 200 OK in 8 milliseconds
I0729 23:17:32.487330 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:32.487360 29396 round_trippers.go:469] Request Headers:
I0729 23:17:32.487369 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:32.487374 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:32.493333 29396 round_trippers.go:574] Response Status: 200 OK in 5 milliseconds
I0729 23:17:32.493741 29396 node_ready.go:53] node "ha-238496-m02" has status "Ready":"False"
I0729 23:17:32.986894 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:32.986923 29396 round_trippers.go:469] Request Headers:
I0729 23:17:32.986934 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:32.986940 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:32.990578 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:33.486980 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:33.487007 29396 round_trippers.go:469] Request Headers:
I0729 23:17:33.487018 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:33.487026 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:33.490073 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:33.986899 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:33.986918 29396 round_trippers.go:469] Request Headers:
I0729 23:17:33.986926 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:33.986930 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:33.990326 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:34.487551 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:34.487571 29396 round_trippers.go:469] Request Headers:
I0729 23:17:34.487578 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:34.487582 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:34.491541 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:34.987623 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:34.987644 29396 round_trippers.go:469] Request Headers:
I0729 23:17:34.987652 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:34.987656 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:34.991325 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:34.991873 29396 node_ready.go:53] node "ha-238496-m02" has status "Ready":"False"
I0729 23:17:35.487183 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:35.487210 29396 round_trippers.go:469] Request Headers:
I0729 23:17:35.487223 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:35.487229 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:35.491177 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:35.987223 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:35.987248 29396 round_trippers.go:469] Request Headers:
I0729 23:17:35.987259 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:35.987266 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:35.990525 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:36.486820 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:36.486843 29396 round_trippers.go:469] Request Headers:
I0729 23:17:36.486855 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:36.486859 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:36.490221 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:36.987151 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:36.987172 29396 round_trippers.go:469] Request Headers:
I0729 23:17:36.987183 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:36.987189 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:36.990147 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:17:37.486819 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:37.486842 29396 round_trippers.go:469] Request Headers:
I0729 23:17:37.486849 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:37.486853 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:37.490362 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:37.490858 29396 node_ready.go:53] node "ha-238496-m02" has status "Ready":"False"
I0729 23:17:37.987231 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:37.987266 29396 round_trippers.go:469] Request Headers:
I0729 23:17:37.987278 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:37.987282 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:37.991172 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:38.487456 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:38.487480 29396 round_trippers.go:469] Request Headers:
I0729 23:17:38.487488 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:38.487492 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:38.491185 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:38.987722 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:38.987746 29396 round_trippers.go:469] Request Headers:
I0729 23:17:38.987757 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:38.987760 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:38.990953 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:39.487485 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:39.487507 29396 round_trippers.go:469] Request Headers:
I0729 23:17:39.487515 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:39.487519 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:39.490615 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:39.491413 29396 node_ready.go:53] node "ha-238496-m02" has status "Ready":"False"
I0729 23:17:39.987861 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:39.987885 29396 round_trippers.go:469] Request Headers:
I0729 23:17:39.987894 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:39.987899 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:39.991871 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:40.487060 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:40.487089 29396 round_trippers.go:469] Request Headers:
I0729 23:17:40.487099 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:40.487104 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:40.490670 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:40.986848 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:40.986878 29396 round_trippers.go:469] Request Headers:
I0729 23:17:40.986888 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:40.986892 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:40.990180 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:41.487838 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:41.487861 29396 round_trippers.go:469] Request Headers:
I0729 23:17:41.487868 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:41.487872 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:41.491386 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:41.492362 29396 node_ready.go:53] node "ha-238496-m02" has status "Ready":"False"
I0729 23:17:41.987822 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:41.987844 29396 round_trippers.go:469] Request Headers:
I0729 23:17:41.987852 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:41.987856 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:41.991266 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:42.487015 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:42.487037 29396 round_trippers.go:469] Request Headers:
I0729 23:17:42.487045 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:42.487049 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:42.491106 29396 round_trippers.go:574] Response Status: 200 OK in 4 milliseconds
I0729 23:17:42.986869 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:42.986888 29396 round_trippers.go:469] Request Headers:
I0729 23:17:42.986896 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:42.986901 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:42.990460 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:43.487128 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:43.487155 29396 round_trippers.go:469] Request Headers:
I0729 23:17:43.487166 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:43.487171 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:43.490615 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:43.987811 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:43.987833 29396 round_trippers.go:469] Request Headers:
I0729 23:17:43.987841 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:43.987845 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:43.991933 29396 round_trippers.go:574] Response Status: 200 OK in 4 milliseconds
I0729 23:17:43.992489 29396 node_ready.go:53] node "ha-238496-m02" has status "Ready":"False"
I0729 23:17:44.487362 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:44.487384 29396 round_trippers.go:469] Request Headers:
I0729 23:17:44.487392 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:44.487396 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:44.490237 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:17:44.986868 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:44.986890 29396 round_trippers.go:469] Request Headers:
I0729 23:17:44.986898 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:44.986902 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:44.990865 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:45.487027 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:45.487057 29396 round_trippers.go:469] Request Headers:
I0729 23:17:45.487067 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:45.487072 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:45.490556 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:45.987818 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:45.987840 29396 round_trippers.go:469] Request Headers:
I0729 23:17:45.987850 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:45.987857 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:45.993195 29396 round_trippers.go:574] Response Status: 200 OK in 5 milliseconds
I0729 23:17:45.993828 29396 node_ready.go:53] node "ha-238496-m02" has status "Ready":"False"
I0729 23:17:46.487019 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:46.487042 29396 round_trippers.go:469] Request Headers:
I0729 23:17:46.487054 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:46.487059 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:46.490270 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:46.987227 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:46.987251 29396 round_trippers.go:469] Request Headers:
I0729 23:17:46.987259 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:46.987263 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:46.990459 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:47.487319 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:47.487349 29396 round_trippers.go:469] Request Headers:
I0729 23:17:47.487358 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:47.487363 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:47.490818 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:47.986997 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:47.987019 29396 round_trippers.go:469] Request Headers:
I0729 23:17:47.987027 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:47.987033 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:47.990157 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:48.486848 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:48.486887 29396 round_trippers.go:469] Request Headers:
I0729 23:17:48.486898 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:48.486904 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:48.490233 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:48.490756 29396 node_ready.go:53] node "ha-238496-m02" has status "Ready":"False"
I0729 23:17:48.987333 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:48.987365 29396 round_trippers.go:469] Request Headers:
I0729 23:17:48.987373 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:48.987377 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:48.990768 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:49.487593 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:49.487616 29396 round_trippers.go:469] Request Headers:
I0729 23:17:49.487624 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:49.487628 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:49.490948 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:49.986845 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:49.986868 29396 round_trippers.go:469] Request Headers:
I0729 23:17:49.986876 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:49.986880 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:49.990586 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:50.487187 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:50.487210 29396 round_trippers.go:469] Request Headers:
I0729 23:17:50.487219 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:50.487226 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:50.491614 29396 round_trippers.go:574] Response Status: 200 OK in 4 milliseconds
I0729 23:17:50.492061 29396 node_ready.go:53] node "ha-238496-m02" has status "Ready":"False"
I0729 23:17:50.987524 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:50.987546 29396 round_trippers.go:469] Request Headers:
I0729 23:17:50.987554 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:50.987559 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:50.990568 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:17:51.487617 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:51.487650 29396 round_trippers.go:469] Request Headers:
I0729 23:17:51.487661 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:51.487666 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:51.491397 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:51.987402 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:51.987423 29396 round_trippers.go:469] Request Headers:
I0729 23:17:51.987431 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:51.987435 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:51.991650 29396 round_trippers.go:574] Response Status: 200 OK in 4 milliseconds
I0729 23:17:52.487727 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:52.487753 29396 round_trippers.go:469] Request Headers:
I0729 23:17:52.487764 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:52.487770 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:52.491368 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:52.491966 29396 node_ready.go:49] node "ha-238496-m02" has status "Ready":"True"
I0729 23:17:52.491985 29396 node_ready.go:38] duration metric: took 22.005334623s for node "ha-238496-m02" to be "Ready" ...
I0729 23:17:52.491993 29396 pod_ready.go:35] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0729 23:17:52.492046 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods
I0729 23:17:52.492054 29396 round_trippers.go:469] Request Headers:
I0729 23:17:52.492061 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:52.492064 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:52.504110 29396 round_trippers.go:574] Response Status: 200 OK in 12 milliseconds
I0729 23:17:52.510523 29396 pod_ready.go:78] waiting up to 6m0s for pod "coredns-7db6d8ff4d-p8nps" in "kube-system" namespace to be "Ready" ...
I0729 23:17:52.510635 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/coredns-7db6d8ff4d-p8nps
I0729 23:17:52.510646 29396 round_trippers.go:469] Request Headers:
I0729 23:17:52.510656 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:52.510665 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:52.517610 29396 round_trippers.go:574] Response Status: 200 OK in 6 milliseconds
I0729 23:17:52.518186 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:17:52.518201 29396 round_trippers.go:469] Request Headers:
I0729 23:17:52.518208 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:52.518211 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:52.526545 29396 round_trippers.go:574] Response Status: 200 OK in 8 milliseconds
I0729 23:17:52.527004 29396 pod_ready.go:92] pod "coredns-7db6d8ff4d-p8nps" in "kube-system" namespace has status "Ready":"True"
I0729 23:17:52.527023 29396 pod_ready.go:81] duration metric: took 16.46732ms for pod "coredns-7db6d8ff4d-p8nps" in "kube-system" namespace to be "Ready" ...
I0729 23:17:52.527035 29396 pod_ready.go:78] waiting up to 6m0s for pod "coredns-7db6d8ff4d-tjplq" in "kube-system" namespace to be "Ready" ...
I0729 23:17:52.527102 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/coredns-7db6d8ff4d-tjplq
I0729 23:17:52.527112 29396 round_trippers.go:469] Request Headers:
I0729 23:17:52.527122 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:52.527129 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:52.532180 29396 round_trippers.go:574] Response Status: 200 OK in 5 milliseconds
I0729 23:17:52.532980 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:17:52.532994 29396 round_trippers.go:469] Request Headers:
I0729 23:17:52.533003 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:52.533010 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:52.540147 29396 round_trippers.go:574] Response Status: 200 OK in 7 milliseconds
I0729 23:17:52.540734 29396 pod_ready.go:92] pod "coredns-7db6d8ff4d-tjplq" in "kube-system" namespace has status "Ready":"True"
I0729 23:17:52.540750 29396 pod_ready.go:81] duration metric: took 13.707118ms for pod "coredns-7db6d8ff4d-tjplq" in "kube-system" namespace to be "Ready" ...
I0729 23:17:52.540763 29396 pod_ready.go:78] waiting up to 6m0s for pod "etcd-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:17:52.540850 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/etcd-ha-238496
I0729 23:17:52.540860 29396 round_trippers.go:469] Request Headers:
I0729 23:17:52.540868 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:52.540875 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:52.546093 29396 round_trippers.go:574] Response Status: 200 OK in 5 milliseconds
I0729 23:17:52.546851 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:17:52.546869 29396 round_trippers.go:469] Request Headers:
I0729 23:17:52.546880 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:52.546886 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:52.552117 29396 round_trippers.go:574] Response Status: 200 OK in 5 milliseconds
I0729 23:17:52.552716 29396 pod_ready.go:92] pod "etcd-ha-238496" in "kube-system" namespace has status "Ready":"True"
I0729 23:17:52.552732 29396 pod_ready.go:81] duration metric: took 11.96269ms for pod "etcd-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:17:52.552741 29396 pod_ready.go:78] waiting up to 6m0s for pod "etcd-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:17:52.552797 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/etcd-ha-238496-m02
I0729 23:17:52.552808 29396 round_trippers.go:469] Request Headers:
I0729 23:17:52.552818 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:52.552826 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:52.558334 29396 round_trippers.go:574] Response Status: 200 OK in 5 milliseconds
I0729 23:17:52.559013 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:52.559032 29396 round_trippers.go:469] Request Headers:
I0729 23:17:52.559042 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:52.559047 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:52.561908 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:17:52.562357 29396 pod_ready.go:92] pod "etcd-ha-238496-m02" in "kube-system" namespace has status "Ready":"True"
I0729 23:17:52.562377 29396 pod_ready.go:81] duration metric: took 9.630355ms for pod "etcd-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:17:52.562391 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-apiserver-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:17:52.688744 29396 request.go:629] Waited for 126.281844ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-apiserver-ha-238496
I0729 23:17:52.688810 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-apiserver-ha-238496
I0729 23:17:52.688817 29396 round_trippers.go:469] Request Headers:
I0729 23:17:52.688828 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:52.688833 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:52.692265 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:52.888364 29396 request.go:629] Waited for 195.376381ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:17:52.888439 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:17:52.888445 29396 round_trippers.go:469] Request Headers:
I0729 23:17:52.888452 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:52.888457 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:52.891026 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:17:52.891567 29396 pod_ready.go:92] pod "kube-apiserver-ha-238496" in "kube-system" namespace has status "Ready":"True"
I0729 23:17:52.891584 29396 pod_ready.go:81] duration metric: took 329.186355ms for pod "kube-apiserver-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:17:52.891593 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-apiserver-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:17:53.087981 29396 request.go:629] Waited for 196.334679ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-apiserver-ha-238496-m02
I0729 23:17:53.088056 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-apiserver-ha-238496-m02
I0729 23:17:53.088064 29396 round_trippers.go:469] Request Headers:
I0729 23:17:53.088072 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:53.088079 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:53.091485 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:53.288474 29396 request.go:629] Waited for 196.352331ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:53.288520 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:53.288525 29396 round_trippers.go:469] Request Headers:
I0729 23:17:53.288532 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:53.288536 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:53.292168 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:53.292688 29396 pod_ready.go:92] pod "kube-apiserver-ha-238496-m02" in "kube-system" namespace has status "Ready":"True"
I0729 23:17:53.292709 29396 pod_ready.go:81] duration metric: took 401.108898ms for pod "kube-apiserver-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:17:53.292721 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-controller-manager-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:17:53.487802 29396 request.go:629] Waited for 195.017883ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-controller-manager-ha-238496
I0729 23:17:53.487875 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-controller-manager-ha-238496
I0729 23:17:53.487880 29396 round_trippers.go:469] Request Headers:
I0729 23:17:53.487888 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:53.487896 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:53.491238 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:53.688436 29396 request.go:629] Waited for 196.401022ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:17:53.688492 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:17:53.688509 29396 round_trippers.go:469] Request Headers:
I0729 23:17:53.688517 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:53.688521 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:53.691835 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:53.692301 29396 pod_ready.go:92] pod "kube-controller-manager-ha-238496" in "kube-system" namespace has status "Ready":"True"
I0729 23:17:53.692325 29396 pod_ready.go:81] duration metric: took 399.593595ms for pod "kube-controller-manager-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:17:53.692339 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-controller-manager-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:17:53.888409 29396 request.go:629] Waited for 196.00174ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-controller-manager-ha-238496-m02
I0729 23:17:53.888471 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-controller-manager-ha-238496-m02
I0729 23:17:53.888477 29396 round_trippers.go:469] Request Headers:
I0729 23:17:53.888484 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:53.888492 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:53.891692 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:54.088679 29396 request.go:629] Waited for 196.402331ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:54.088739 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:54.088747 29396 round_trippers.go:469] Request Headers:
I0729 23:17:54.088758 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:54.088763 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:54.092410 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:54.093200 29396 pod_ready.go:92] pod "kube-controller-manager-ha-238496-m02" in "kube-system" namespace has status "Ready":"True"
I0729 23:17:54.093219 29396 pod_ready.go:81] duration metric: took 400.872004ms for pod "kube-controller-manager-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:17:54.093229 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-proxy-m6vdn" in "kube-system" namespace to be "Ready" ...
I0729 23:17:54.288425 29396 request.go:629] Waited for 195.10707ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-proxy-m6vdn
I0729 23:17:54.288476 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-proxy-m6vdn
I0729 23:17:54.288482 29396 round_trippers.go:469] Request Headers:
I0729 23:17:54.288490 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:54.288493 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:54.292185 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:54.488193 29396 request.go:629] Waited for 195.421802ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:54.488248 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:54.488253 29396 round_trippers.go:469] Request Headers:
I0729 23:17:54.488260 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:54.488265 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:54.491550 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:54.492131 29396 pod_ready.go:92] pod "kube-proxy-m6vdn" in "kube-system" namespace has status "Ready":"True"
I0729 23:17:54.492151 29396 pod_ready.go:81] duration metric: took 398.917263ms for pod "kube-proxy-m6vdn" in "kube-system" namespace to be "Ready" ...
I0729 23:17:54.492160 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-proxy-nrvw6" in "kube-system" namespace to be "Ready" ...
I0729 23:17:54.688359 29396 request.go:629] Waited for 196.138725ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-proxy-nrvw6
I0729 23:17:54.688434 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-proxy-nrvw6
I0729 23:17:54.688440 29396 round_trippers.go:469] Request Headers:
I0729 23:17:54.688447 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:54.688451 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:54.691559 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:54.888419 29396 request.go:629] Waited for 196.328075ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:17:54.888482 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:17:54.888487 29396 round_trippers.go:469] Request Headers:
I0729 23:17:54.888494 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:54.888499 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:54.891566 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:54.892146 29396 pod_ready.go:92] pod "kube-proxy-nrvw6" in "kube-system" namespace has status "Ready":"True"
I0729 23:17:54.892169 29396 pod_ready.go:81] duration metric: took 400.003082ms for pod "kube-proxy-nrvw6" in "kube-system" namespace to be "Ready" ...
I0729 23:17:54.892179 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-scheduler-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:17:55.088353 29396 request.go:629] Waited for 196.106382ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-scheduler-ha-238496
I0729 23:17:55.088408 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-scheduler-ha-238496
I0729 23:17:55.088413 29396 round_trippers.go:469] Request Headers:
I0729 23:17:55.088423 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:55.088429 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:55.091946 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:55.287817 29396 request.go:629] Waited for 195.275329ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:17:55.287881 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:17:55.287904 29396 round_trippers.go:469] Request Headers:
I0729 23:17:55.287913 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:55.287918 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:55.291966 29396 round_trippers.go:574] Response Status: 200 OK in 4 milliseconds
I0729 23:17:55.292571 29396 pod_ready.go:92] pod "kube-scheduler-ha-238496" in "kube-system" namespace has status "Ready":"True"
I0729 23:17:55.292591 29396 pod_ready.go:81] duration metric: took 400.405235ms for pod "kube-scheduler-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:17:55.292601 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-scheduler-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:17:55.488732 29396 request.go:629] Waited for 196.075875ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-scheduler-ha-238496-m02
I0729 23:17:55.488803 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-scheduler-ha-238496-m02
I0729 23:17:55.488824 29396 round_trippers.go:469] Request Headers:
I0729 23:17:55.488835 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:55.488840 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:55.492759 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:55.687724 29396 request.go:629] Waited for 194.327667ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:55.687806 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:17:55.687815 29396 round_trippers.go:469] Request Headers:
I0729 23:17:55.687828 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:55.687837 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:55.691900 29396 round_trippers.go:574] Response Status: 200 OK in 4 milliseconds
I0729 23:17:55.692304 29396 pod_ready.go:92] pod "kube-scheduler-ha-238496-m02" in "kube-system" namespace has status "Ready":"True"
I0729 23:17:55.692319 29396 pod_ready.go:81] duration metric: took 399.71201ms for pod "kube-scheduler-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:17:55.692329 29396 pod_ready.go:38] duration metric: took 3.200327096s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0729 23:17:55.692348 29396 api_server.go:52] waiting for apiserver process to appear ...
I0729 23:17:55.692407 29396 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0729 23:17:55.707796 29396 api_server.go:72] duration metric: took 25.548134132s to wait for apiserver process to appear ...
I0729 23:17:55.707815 29396 api_server.go:88] waiting for apiserver healthz status ...
I0729 23:17:55.707828 29396 api_server.go:253] Checking apiserver healthz at https://192.168.39.113:8443/healthz ...
I0729 23:17:55.714746 29396 api_server.go:279] https://192.168.39.113:8443/healthz returned 200:
ok
I0729 23:17:55.714807 29396 round_trippers.go:463] GET https://192.168.39.113:8443/version
I0729 23:17:55.714813 29396 round_trippers.go:469] Request Headers:
I0729 23:17:55.714823 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:55.714831 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:55.716204 29396 round_trippers.go:574] Response Status: 200 OK in 1 milliseconds
I0729 23:17:55.716341 29396 api_server.go:141] control plane version: v1.30.3
I0729 23:17:55.716364 29396 api_server.go:131] duration metric: took 8.542521ms to wait for apiserver health ...
I0729 23:17:55.716374 29396 system_pods.go:43] waiting for kube-system pods to appear ...
I0729 23:17:55.888804 29396 request.go:629] Waited for 172.357533ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods
I0729 23:17:55.888877 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods
I0729 23:17:55.888883 29396 round_trippers.go:469] Request Headers:
I0729 23:17:55.888891 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:55.888894 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:55.894267 29396 round_trippers.go:574] Response Status: 200 OK in 5 milliseconds
I0729 23:17:55.898781 29396 system_pods.go:59] 17 kube-system pods found
I0729 23:17:55.898806 29396 system_pods.go:61] "coredns-7db6d8ff4d-p8nps" [af3f5c7b-1996-497f-95f7-4bfc87392dc7] Running
I0729 23:17:55.898811 29396 system_pods.go:61] "coredns-7db6d8ff4d-tjplq" [db7a6b8c-bfe3-4291-bf9a-9ce96bb5b0b7] Running
I0729 23:17:55.898815 29396 system_pods.go:61] "etcd-ha-238496" [ed3a1237-a4c1-4e3f-b7d6-6b5237f7a18b] Running
I0729 23:17:55.898819 29396 system_pods.go:61] "etcd-ha-238496-m02" [0a4d5ebc-a7be-445f-bdfc-47b3b1c01803] Running
I0729 23:17:55.898822 29396 system_pods.go:61] "kindnet-55jmm" [7ddd1f82-1105-4694-b8d6-5198fdbd1f86] Running
I0729 23:17:55.898827 29396 system_pods.go:61] "kindnet-xvzff" [400a9d4f-d218-443e-b001-edd5e5fd5af7] Running
I0729 23:17:55.898830 29396 system_pods.go:61] "kube-apiserver-ha-238496" [54eebf95-2bd3-4c57-9794-170fccda1dbb] Running
I0729 23:17:55.898834 29396 system_pods.go:61] "kube-apiserver-ha-238496-m02" [66429444-6c99-474c-9294-c569e1a5cc46] Running
I0729 23:17:55.898838 29396 system_pods.go:61] "kube-controller-manager-ha-238496" [bb6bc2ad-54ec-42fa-8f18-e33cb50a8ce8] Running
I0729 23:17:55.898842 29396 system_pods.go:61] "kube-controller-manager-ha-238496-m02" [8836c211-ee9d-403a-8383-333c22f1b945] Running
I0729 23:17:55.898845 29396 system_pods.go:61] "kube-proxy-m6vdn" [f3731d91-d919-4f7f-a7b9-2bf7ba93569b] Running
I0729 23:17:55.898848 29396 system_pods.go:61] "kube-proxy-nrvw6" [708cca57-5274-4ad9-871c-048f24b43a33] Running
I0729 23:17:55.898851 29396 system_pods.go:61] "kube-scheduler-ha-238496" [b4999631-2ffc-4684-ab41-7e065cbbe74b] Running
I0729 23:17:55.898857 29396 system_pods.go:61] "kube-scheduler-ha-238496-m02" [4eb7be71-6cad-4260-a4c0-6a97011e6ec5] Running
I0729 23:17:55.898859 29396 system_pods.go:61] "kube-vip-ha-238496" [f248f380-c48b-451a-82e7-0aeb1e0ba6eb] Running
I0729 23:17:55.898862 29396 system_pods.go:61] "kube-vip-ha-238496-m02" [39a50caf-f960-4d68-9235-d6dacace51c1] Running
I0729 23:17:55.898865 29396 system_pods.go:61] "storage-provisioner" [2feba04d-7105-41cd-b308-747ed0079849] Running
I0729 23:17:55.898871 29396 system_pods.go:74] duration metric: took 182.492062ms to wait for pod list to return data ...
I0729 23:17:55.898879 29396 default_sa.go:34] waiting for default service account to be created ...
I0729 23:17:56.088327 29396 request.go:629] Waited for 189.384723ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/default/serviceaccounts
I0729 23:17:56.088391 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/default/serviceaccounts
I0729 23:17:56.088397 29396 round_trippers.go:469] Request Headers:
I0729 23:17:56.088405 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:56.088409 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:56.091993 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:56.092247 29396 default_sa.go:45] found service account: "default"
I0729 23:17:56.092271 29396 default_sa.go:55] duration metric: took 193.384432ms for default service account to be created ...
I0729 23:17:56.092281 29396 system_pods.go:116] waiting for k8s-apps to be running ...
I0729 23:17:56.288411 29396 request.go:629] Waited for 196.051331ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods
I0729 23:17:56.288503 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods
I0729 23:17:56.288510 29396 round_trippers.go:469] Request Headers:
I0729 23:17:56.288519 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:56.288526 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:56.293596 29396 round_trippers.go:574] Response Status: 200 OK in 5 milliseconds
I0729 23:17:56.298060 29396 system_pods.go:86] 17 kube-system pods found
I0729 23:17:56.298088 29396 system_pods.go:89] "coredns-7db6d8ff4d-p8nps" [af3f5c7b-1996-497f-95f7-4bfc87392dc7] Running
I0729 23:17:56.298095 29396 system_pods.go:89] "coredns-7db6d8ff4d-tjplq" [db7a6b8c-bfe3-4291-bf9a-9ce96bb5b0b7] Running
I0729 23:17:56.298100 29396 system_pods.go:89] "etcd-ha-238496" [ed3a1237-a4c1-4e3f-b7d6-6b5237f7a18b] Running
I0729 23:17:56.298104 29396 system_pods.go:89] "etcd-ha-238496-m02" [0a4d5ebc-a7be-445f-bdfc-47b3b1c01803] Running
I0729 23:17:56.298110 29396 system_pods.go:89] "kindnet-55jmm" [7ddd1f82-1105-4694-b8d6-5198fdbd1f86] Running
I0729 23:17:56.298114 29396 system_pods.go:89] "kindnet-xvzff" [400a9d4f-d218-443e-b001-edd5e5fd5af7] Running
I0729 23:17:56.298120 29396 system_pods.go:89] "kube-apiserver-ha-238496" [54eebf95-2bd3-4c57-9794-170fccda1dbb] Running
I0729 23:17:56.298126 29396 system_pods.go:89] "kube-apiserver-ha-238496-m02" [66429444-6c99-474c-9294-c569e1a5cc46] Running
I0729 23:17:56.298132 29396 system_pods.go:89] "kube-controller-manager-ha-238496" [bb6bc2ad-54ec-42fa-8f18-e33cb50a8ce8] Running
I0729 23:17:56.298140 29396 system_pods.go:89] "kube-controller-manager-ha-238496-m02" [8836c211-ee9d-403a-8383-333c22f1b945] Running
I0729 23:17:56.298150 29396 system_pods.go:89] "kube-proxy-m6vdn" [f3731d91-d919-4f7f-a7b9-2bf7ba93569b] Running
I0729 23:17:56.298156 29396 system_pods.go:89] "kube-proxy-nrvw6" [708cca57-5274-4ad9-871c-048f24b43a33] Running
I0729 23:17:56.298162 29396 system_pods.go:89] "kube-scheduler-ha-238496" [b4999631-2ffc-4684-ab41-7e065cbbe74b] Running
I0729 23:17:56.298172 29396 system_pods.go:89] "kube-scheduler-ha-238496-m02" [4eb7be71-6cad-4260-a4c0-6a97011e6ec5] Running
I0729 23:17:56.298178 29396 system_pods.go:89] "kube-vip-ha-238496" [f248f380-c48b-451a-82e7-0aeb1e0ba6eb] Running
I0729 23:17:56.298182 29396 system_pods.go:89] "kube-vip-ha-238496-m02" [39a50caf-f960-4d68-9235-d6dacace51c1] Running
I0729 23:17:56.298186 29396 system_pods.go:89] "storage-provisioner" [2feba04d-7105-41cd-b308-747ed0079849] Running
I0729 23:17:56.298193 29396 system_pods.go:126] duration metric: took 205.904881ms to wait for k8s-apps to be running ...
I0729 23:17:56.298204 29396 system_svc.go:44] waiting for kubelet service to be running ....
I0729 23:17:56.298252 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0729 23:17:56.313300 29396 system_svc.go:56] duration metric: took 15.089221ms WaitForService to wait for kubelet
I0729 23:17:56.313325 29396 kubeadm.go:582] duration metric: took 26.153665426s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0729 23:17:56.313348 29396 node_conditions.go:102] verifying NodePressure condition ...
I0729 23:17:56.487722 29396 request.go:629] Waited for 174.298793ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes
I0729 23:17:56.487799 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes
I0729 23:17:56.487809 29396 round_trippers.go:469] Request Headers:
I0729 23:17:56.487820 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:17:56.487828 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:17:56.491492 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:17:56.492513 29396 node_conditions.go:122] node storage ephemeral capacity is 17734596Ki
I0729 23:17:56.492537 29396 node_conditions.go:123] node cpu capacity is 2
I0729 23:17:56.492550 29396 node_conditions.go:122] node storage ephemeral capacity is 17734596Ki
I0729 23:17:56.492555 29396 node_conditions.go:123] node cpu capacity is 2
I0729 23:17:56.492561 29396 node_conditions.go:105] duration metric: took 179.206939ms to run NodePressure ...
I0729 23:17:56.492573 29396 start.go:241] waiting for startup goroutines ...
I0729 23:17:56.492604 29396 start.go:255] writing updated cluster config ...
I0729 23:17:56.494762 29396 out.go:177]
I0729 23:17:56.496382 29396 config.go:182] Loaded profile config "ha-238496": Driver=kvm2, ContainerRuntime=docker, KubernetesVersion=v1.30.3
I0729 23:17:56.496476 29396 profile.go:143] Saving config to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/config.json ...
I0729 23:17:56.498380 29396 out.go:177] * Starting "ha-238496-m03" control-plane node in "ha-238496" cluster
I0729 23:17:56.499671 29396 preload.go:131] Checking if preload exists for k8s version v1.30.3 and runtime docker
I0729 23:17:56.499697 29396 cache.go:56] Caching tarball of preloaded images
I0729 23:17:56.499810 29396 preload.go:172] Found /home/jenkins/minikube-integration/19347-12221/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.30.3-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0729 23:17:56.499827 29396 cache.go:59] Finished verifying existence of preloaded tar for v1.30.3 on docker
I0729 23:17:56.499918 29396 profile.go:143] Saving config to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/config.json ...
I0729 23:17:56.500102 29396 start.go:360] acquireMachinesLock for ha-238496-m03: {Name:mk79fbc287386032c39e512567e9786663e657a9 Clock:{} Delay:500ms Timeout:13m0s Cancel:<nil>}
I0729 23:17:56.500164 29396 start.go:364] duration metric: took 35.548µs to acquireMachinesLock for "ha-238496-m03"
I0729 23:17:56.500188 29396 start.go:93] Provisioning new machine with config: &{Name:ha-238496 KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube-builds/iso/19319/minikube-v1.33.1-1721690939-19319-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721902582-19326@sha256:540fb5dc7f38be17ff5276a38dfe6c8a4b1d9ba1c27c62244e6eebd7e37696e7 Memory:2200 CPUs:2 DiskSize:20000 Driver:kvm2 HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{Kubernete
sVersion:v1.30.3 ClusterName:ha-238496 Namespace:default APIServerHAVIP:192.168.39.254 APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.39.113 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true} {Name:m02 IP:192.168.39.226 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true} {Name:m03 IP: Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[ambassador:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:false efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false helm-tiller:false inaccel:false ingress:false
ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:false storage-provisioner-gluster:false storage-provisioner-rancher:false volcano:false volumesnapshots:false yakd:false] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:true ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker Bina
ryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name:m03 IP: Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}
I0729 23:17:56.500322 29396 start.go:125] createHost starting for "m03" (driver="kvm2")
I0729 23:17:56.501844 29396 out.go:204] * Creating kvm2 VM (CPUs=2, Memory=2200MB, Disk=20000MB) ...
I0729 23:17:56.501937 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:17:56.501974 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:17:56.516619 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:45937
I0729 23:17:56.517059 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:17:56.517528 29396 main.go:141] libmachine: Using API Version 1
I0729 23:17:56.517548 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:17:56.517956 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:17:56.518148 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetMachineName
I0729 23:17:56.518310 29396 main.go:141] libmachine: (ha-238496-m03) Calling .DriverName
I0729 23:17:56.518469 29396 start.go:159] libmachine.API.Create for "ha-238496" (driver="kvm2")
I0729 23:17:56.518499 29396 client.go:168] LocalClient.Create starting
I0729 23:17:56.518534 29396 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem
I0729 23:17:56.518573 29396 main.go:141] libmachine: Decoding PEM data...
I0729 23:17:56.518591 29396 main.go:141] libmachine: Parsing certificate...
I0729 23:17:56.518657 29396 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem
I0729 23:17:56.518681 29396 main.go:141] libmachine: Decoding PEM data...
I0729 23:17:56.518711 29396 main.go:141] libmachine: Parsing certificate...
I0729 23:17:56.518737 29396 main.go:141] libmachine: Running pre-create checks...
I0729 23:17:56.518748 29396 main.go:141] libmachine: (ha-238496-m03) Calling .PreCreateCheck
I0729 23:17:56.518918 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetConfigRaw
I0729 23:17:56.519351 29396 main.go:141] libmachine: Creating machine...
I0729 23:17:56.519366 29396 main.go:141] libmachine: (ha-238496-m03) Calling .Create
I0729 23:17:56.519523 29396 main.go:141] libmachine: (ha-238496-m03) Creating KVM machine...
I0729 23:17:56.520838 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found existing default KVM network
I0729 23:17:56.521062 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found existing private KVM network mk-ha-238496
I0729 23:17:56.521175 29396 main.go:141] libmachine: (ha-238496-m03) Setting up store path in /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03 ...
I0729 23:17:56.521192 29396 main.go:141] libmachine: (ha-238496-m03) Building disk image from file:///home/jenkins/minikube-integration/19347-12221/.minikube/cache/iso/amd64/minikube-v1.33.1-1721690939-19319-amd64.iso
I0729 23:17:56.521250 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:17:56.521166 30320 common.go:145] Making disk image using store path: /home/jenkins/minikube-integration/19347-12221/.minikube
I0729 23:17:56.521335 29396 main.go:141] libmachine: (ha-238496-m03) Downloading /home/jenkins/minikube-integration/19347-12221/.minikube/cache/boot2docker.iso from file:///home/jenkins/minikube-integration/19347-12221/.minikube/cache/iso/amd64/minikube-v1.33.1-1721690939-19319-amd64.iso...
I0729 23:17:56.760034 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:17:56.759892 30320 common.go:152] Creating ssh key: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03/id_rsa...
I0729 23:17:56.853038 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:17:56.852885 30320 common.go:158] Creating raw disk image: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03/ha-238496-m03.rawdisk...
I0729 23:17:56.853082 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Writing magic tar header
I0729 23:17:56.853127 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Writing SSH key tar header
I0729 23:17:56.853159 29396 main.go:141] libmachine: (ha-238496-m03) Setting executable bit set on /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03 (perms=drwx------)
I0729 23:17:56.853175 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:17:56.852996 30320 common.go:172] Fixing permissions on /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03 ...
I0729 23:17:56.853200 29396 main.go:141] libmachine: (ha-238496-m03) Setting executable bit set on /home/jenkins/minikube-integration/19347-12221/.minikube/machines (perms=drwxr-xr-x)
I0729 23:17:56.853225 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Checking permissions on dir: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03
I0729 23:17:56.853237 29396 main.go:141] libmachine: (ha-238496-m03) Setting executable bit set on /home/jenkins/minikube-integration/19347-12221/.minikube (perms=drwxr-xr-x)
I0729 23:17:56.853253 29396 main.go:141] libmachine: (ha-238496-m03) Setting executable bit set on /home/jenkins/minikube-integration/19347-12221 (perms=drwxrwxr-x)
I0729 23:17:56.853291 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Checking permissions on dir: /home/jenkins/minikube-integration/19347-12221/.minikube/machines
I0729 23:17:56.853302 29396 main.go:141] libmachine: (ha-238496-m03) Setting executable bit set on /home/jenkins/minikube-integration (perms=drwxrwxr-x)
I0729 23:17:56.853313 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Checking permissions on dir: /home/jenkins/minikube-integration/19347-12221/.minikube
I0729 23:17:56.853338 29396 main.go:141] libmachine: (ha-238496-m03) Setting executable bit set on /home/jenkins (perms=drwxr-xr-x)
I0729 23:17:56.853354 29396 main.go:141] libmachine: (ha-238496-m03) Creating domain...
I0729 23:17:56.853368 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Checking permissions on dir: /home/jenkins/minikube-integration/19347-12221
I0729 23:17:56.853389 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Checking permissions on dir: /home/jenkins/minikube-integration
I0729 23:17:56.853401 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Checking permissions on dir: /home/jenkins
I0729 23:17:56.853413 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Checking permissions on dir: /home
I0729 23:17:56.853421 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Skipping /home - not owner
I0729 23:17:56.854341 29396 main.go:141] libmachine: (ha-238496-m03) define libvirt domain using xml:
I0729 23:17:56.854361 29396 main.go:141] libmachine: (ha-238496-m03) <domain type='kvm'>
I0729 23:17:56.854378 29396 main.go:141] libmachine: (ha-238496-m03) <name>ha-238496-m03</name>
I0729 23:17:56.854388 29396 main.go:141] libmachine: (ha-238496-m03) <memory unit='MiB'>2200</memory>
I0729 23:17:56.854399 29396 main.go:141] libmachine: (ha-238496-m03) <vcpu>2</vcpu>
I0729 23:17:56.854408 29396 main.go:141] libmachine: (ha-238496-m03) <features>
I0729 23:17:56.854419 29396 main.go:141] libmachine: (ha-238496-m03) <acpi/>
I0729 23:17:56.854429 29396 main.go:141] libmachine: (ha-238496-m03) <apic/>
I0729 23:17:56.854438 29396 main.go:141] libmachine: (ha-238496-m03) <pae/>
I0729 23:17:56.854449 29396 main.go:141] libmachine: (ha-238496-m03)
I0729 23:17:56.854459 29396 main.go:141] libmachine: (ha-238496-m03) </features>
I0729 23:17:56.854474 29396 main.go:141] libmachine: (ha-238496-m03) <cpu mode='host-passthrough'>
I0729 23:17:56.854483 29396 main.go:141] libmachine: (ha-238496-m03)
I0729 23:17:56.854489 29396 main.go:141] libmachine: (ha-238496-m03) </cpu>
I0729 23:17:56.854500 29396 main.go:141] libmachine: (ha-238496-m03) <os>
I0729 23:17:56.854510 29396 main.go:141] libmachine: (ha-238496-m03) <type>hvm</type>
I0729 23:17:56.854520 29396 main.go:141] libmachine: (ha-238496-m03) <boot dev='cdrom'/>
I0729 23:17:56.854534 29396 main.go:141] libmachine: (ha-238496-m03) <boot dev='hd'/>
I0729 23:17:56.854563 29396 main.go:141] libmachine: (ha-238496-m03) <bootmenu enable='no'/>
I0729 23:17:56.854583 29396 main.go:141] libmachine: (ha-238496-m03) </os>
I0729 23:17:56.854607 29396 main.go:141] libmachine: (ha-238496-m03) <devices>
I0729 23:17:56.854626 29396 main.go:141] libmachine: (ha-238496-m03) <disk type='file' device='cdrom'>
I0729 23:17:56.854667 29396 main.go:141] libmachine: (ha-238496-m03) <source file='/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03/boot2docker.iso'/>
I0729 23:17:56.854682 29396 main.go:141] libmachine: (ha-238496-m03) <target dev='hdc' bus='scsi'/>
I0729 23:17:56.854710 29396 main.go:141] libmachine: (ha-238496-m03) <readonly/>
I0729 23:17:56.854725 29396 main.go:141] libmachine: (ha-238496-m03) </disk>
I0729 23:17:56.854737 29396 main.go:141] libmachine: (ha-238496-m03) <disk type='file' device='disk'>
I0729 23:17:56.854747 29396 main.go:141] libmachine: (ha-238496-m03) <driver name='qemu' type='raw' cache='default' io='threads' />
I0729 23:17:56.854759 29396 main.go:141] libmachine: (ha-238496-m03) <source file='/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03/ha-238496-m03.rawdisk'/>
I0729 23:17:56.854766 29396 main.go:141] libmachine: (ha-238496-m03) <target dev='hda' bus='virtio'/>
I0729 23:17:56.854772 29396 main.go:141] libmachine: (ha-238496-m03) </disk>
I0729 23:17:56.854781 29396 main.go:141] libmachine: (ha-238496-m03) <interface type='network'>
I0729 23:17:56.854801 29396 main.go:141] libmachine: (ha-238496-m03) <source network='mk-ha-238496'/>
I0729 23:17:56.854815 29396 main.go:141] libmachine: (ha-238496-m03) <model type='virtio'/>
I0729 23:17:56.854824 29396 main.go:141] libmachine: (ha-238496-m03) </interface>
I0729 23:17:56.854830 29396 main.go:141] libmachine: (ha-238496-m03) <interface type='network'>
I0729 23:17:56.854838 29396 main.go:141] libmachine: (ha-238496-m03) <source network='default'/>
I0729 23:17:56.854845 29396 main.go:141] libmachine: (ha-238496-m03) <model type='virtio'/>
I0729 23:17:56.854850 29396 main.go:141] libmachine: (ha-238496-m03) </interface>
I0729 23:17:56.854860 29396 main.go:141] libmachine: (ha-238496-m03) <serial type='pty'>
I0729 23:17:56.854869 29396 main.go:141] libmachine: (ha-238496-m03) <target port='0'/>
I0729 23:17:56.854876 29396 main.go:141] libmachine: (ha-238496-m03) </serial>
I0729 23:17:56.854881 29396 main.go:141] libmachine: (ha-238496-m03) <console type='pty'>
I0729 23:17:56.854888 29396 main.go:141] libmachine: (ha-238496-m03) <target type='serial' port='0'/>
I0729 23:17:56.854903 29396 main.go:141] libmachine: (ha-238496-m03) </console>
I0729 23:17:56.854919 29396 main.go:141] libmachine: (ha-238496-m03) <rng model='virtio'>
I0729 23:17:56.854934 29396 main.go:141] libmachine: (ha-238496-m03) <backend model='random'>/dev/random</backend>
I0729 23:17:56.854943 29396 main.go:141] libmachine: (ha-238496-m03) </rng>
I0729 23:17:56.854951 29396 main.go:141] libmachine: (ha-238496-m03)
I0729 23:17:56.854960 29396 main.go:141] libmachine: (ha-238496-m03)
I0729 23:17:56.854969 29396 main.go:141] libmachine: (ha-238496-m03) </devices>
I0729 23:17:56.854995 29396 main.go:141] libmachine: (ha-238496-m03) </domain>
I0729 23:17:56.855005 29396 main.go:141] libmachine: (ha-238496-m03)
I0729 23:17:56.862233 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:cd:ef:f8 in network default
I0729 23:17:56.862925 29396 main.go:141] libmachine: (ha-238496-m03) Ensuring networks are active...
I0729 23:17:56.862951 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:17:56.863726 29396 main.go:141] libmachine: (ha-238496-m03) Ensuring network default is active
I0729 23:17:56.864012 29396 main.go:141] libmachine: (ha-238496-m03) Ensuring network mk-ha-238496 is active
I0729 23:17:56.864366 29396 main.go:141] libmachine: (ha-238496-m03) Getting domain xml...
I0729 23:17:56.865033 29396 main.go:141] libmachine: (ha-238496-m03) Creating domain...
I0729 23:17:58.116970 29396 main.go:141] libmachine: (ha-238496-m03) Waiting to get IP...
I0729 23:17:58.117886 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:17:58.118260 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find current IP address of domain ha-238496-m03 in network mk-ha-238496
I0729 23:17:58.118300 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:17:58.118258 30320 retry.go:31] will retry after 268.360451ms: waiting for machine to come up
I0729 23:17:58.388875 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:17:58.389411 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find current IP address of domain ha-238496-m03 in network mk-ha-238496
I0729 23:17:58.389435 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:17:58.389365 30320 retry.go:31] will retry after 348.144746ms: waiting for machine to come up
I0729 23:17:58.738773 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:17:58.739331 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find current IP address of domain ha-238496-m03 in network mk-ha-238496
I0729 23:17:58.739360 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:17:58.739272 30320 retry.go:31] will retry after 389.25045ms: waiting for machine to come up
I0729 23:17:59.129833 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:17:59.130375 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find current IP address of domain ha-238496-m03 in network mk-ha-238496
I0729 23:17:59.130401 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:17:59.130330 30320 retry.go:31] will retry after 474.496502ms: waiting for machine to come up
I0729 23:17:59.605919 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:17:59.606486 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find current IP address of domain ha-238496-m03 in network mk-ha-238496
I0729 23:17:59.606509 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:17:59.606424 30320 retry.go:31] will retry after 613.279938ms: waiting for machine to come up
I0729 23:18:00.221389 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:00.221827 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find current IP address of domain ha-238496-m03 in network mk-ha-238496
I0729 23:18:00.221851 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:18:00.221772 30320 retry.go:31] will retry after 600.582506ms: waiting for machine to come up
I0729 23:18:00.823549 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:00.823945 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find current IP address of domain ha-238496-m03 in network mk-ha-238496
I0729 23:18:00.823968 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:18:00.823898 30320 retry.go:31] will retry after 923.091946ms: waiting for machine to come up
I0729 23:18:01.748465 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:01.748887 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find current IP address of domain ha-238496-m03 in network mk-ha-238496
I0729 23:18:01.748914 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:18:01.748870 30320 retry.go:31] will retry after 1.165300062s: waiting for machine to come up
I0729 23:18:02.915681 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:02.916182 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find current IP address of domain ha-238496-m03 in network mk-ha-238496
I0729 23:18:02.916208 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:18:02.916141 30320 retry.go:31] will retry after 1.444012725s: waiting for machine to come up
I0729 23:18:04.361249 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:04.361757 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find current IP address of domain ha-238496-m03 in network mk-ha-238496
I0729 23:18:04.361779 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:18:04.361726 30320 retry.go:31] will retry after 2.185830021s: waiting for machine to come up
I0729 23:18:06.548999 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:06.549480 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find current IP address of domain ha-238496-m03 in network mk-ha-238496
I0729 23:18:06.549506 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:18:06.549438 30320 retry.go:31] will retry after 2.601246738s: waiting for machine to come up
I0729 23:18:09.154097 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:09.154517 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find current IP address of domain ha-238496-m03 in network mk-ha-238496
I0729 23:18:09.154538 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:18:09.154477 30320 retry.go:31] will retry after 3.18346416s: waiting for machine to come up
I0729 23:18:12.339329 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:12.339790 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find current IP address of domain ha-238496-m03 in network mk-ha-238496
I0729 23:18:12.339810 29396 main.go:141] libmachine: (ha-238496-m03) DBG | I0729 23:18:12.339749 30320 retry.go:31] will retry after 4.409983716s: waiting for machine to come up
I0729 23:18:16.750726 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:16.751193 29396 main.go:141] libmachine: (ha-238496-m03) Found IP for machine: 192.168.39.149
I0729 23:18:16.751226 29396 main.go:141] libmachine: (ha-238496-m03) Reserving static IP address...
I0729 23:18:16.751240 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has current primary IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:16.751590 29396 main.go:141] libmachine: (ha-238496-m03) DBG | unable to find host DHCP lease matching {name: "ha-238496-m03", mac: "52:54:00:34:73:00", ip: "192.168.39.149"} in network mk-ha-238496
I0729 23:18:16.821654 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Getting to WaitForSSH function...
I0729 23:18:16.821681 29396 main.go:141] libmachine: (ha-238496-m03) Reserved static IP address: 192.168.39.149
I0729 23:18:16.821693 29396 main.go:141] libmachine: (ha-238496-m03) Waiting for SSH to be available...
I0729 23:18:16.824621 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:16.825058 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:minikube Clientid:01:52:54:00:34:73:00}
I0729 23:18:16.825096 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:16.825253 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Using SSH client type: external
I0729 23:18:16.825282 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Using SSH private key: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03/id_rsa (-rw-------)
I0729 23:18:16.825312 29396 main.go:141] libmachine: (ha-238496-m03) DBG | &{[-F /dev/null -o ConnectionAttempts=3 -o ConnectTimeout=10 -o ControlMaster=no -o ControlPath=none -o LogLevel=quiet -o PasswordAuthentication=no -o ServerAliveInterval=60 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null docker@192.168.39.149 -o IdentitiesOnly=yes -i /home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03/id_rsa -p 22] /usr/bin/ssh <nil>}
I0729 23:18:16.825323 29396 main.go:141] libmachine: (ha-238496-m03) DBG | About to run SSH command:
I0729 23:18:16.825338 29396 main.go:141] libmachine: (ha-238496-m03) DBG | exit 0
I0729 23:18:16.946616 29396 main.go:141] libmachine: (ha-238496-m03) DBG | SSH cmd err, output: <nil>:
I0729 23:18:16.946922 29396 main.go:141] libmachine: (ha-238496-m03) KVM machine creation complete!
I0729 23:18:16.947196 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetConfigRaw
I0729 23:18:16.947801 29396 main.go:141] libmachine: (ha-238496-m03) Calling .DriverName
I0729 23:18:16.947974 29396 main.go:141] libmachine: (ha-238496-m03) Calling .DriverName
I0729 23:18:16.948102 29396 main.go:141] libmachine: Waiting for machine to be running, this may take a few minutes...
I0729 23:18:16.948118 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetState
I0729 23:18:16.949513 29396 main.go:141] libmachine: Detecting operating system of created instance...
I0729 23:18:16.949532 29396 main.go:141] libmachine: Waiting for SSH to be available...
I0729 23:18:16.949537 29396 main.go:141] libmachine: Getting to WaitForSSH function...
I0729 23:18:16.949543 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHHostname
I0729 23:18:16.951794 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:16.952180 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:16.952208 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:16.952339 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHPort
I0729 23:18:16.952505 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:16.952651 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:16.952805 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHUsername
I0729 23:18:16.952961 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:18:16.953205 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.149 22 <nil> <nil>}
I0729 23:18:16.953221 29396 main.go:141] libmachine: About to run SSH command:
exit 0
I0729 23:18:17.054089 29396 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0729 23:18:17.054114 29396 main.go:141] libmachine: Detecting the provisioner...
I0729 23:18:17.054126 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHHostname
I0729 23:18:17.057286 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.057884 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:17.057914 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.058180 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHPort
I0729 23:18:17.058466 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:17.058723 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:17.058900 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHUsername
I0729 23:18:17.059130 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:18:17.059333 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.149 22 <nil> <nil>}
I0729 23:18:17.059349 29396 main.go:141] libmachine: About to run SSH command:
cat /etc/os-release
I0729 23:18:17.167547 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: NAME=Buildroot
VERSION=2023.02.9-dirty
ID=buildroot
VERSION_ID=2023.02.9
PRETTY_NAME="Buildroot 2023.02.9"
I0729 23:18:17.167618 29396 main.go:141] libmachine: found compatible host: buildroot
I0729 23:18:17.167642 29396 main.go:141] libmachine: Provisioning with buildroot...
I0729 23:18:17.167653 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetMachineName
I0729 23:18:17.167880 29396 buildroot.go:166] provisioning hostname "ha-238496-m03"
I0729 23:18:17.167907 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetMachineName
I0729 23:18:17.168122 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHHostname
I0729 23:18:17.170867 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.171312 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:17.171345 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.171499 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHPort
I0729 23:18:17.171697 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:17.171868 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:17.172033 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHUsername
I0729 23:18:17.172229 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:18:17.172456 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.149 22 <nil> <nil>}
I0729 23:18:17.172482 29396 main.go:141] libmachine: About to run SSH command:
sudo hostname ha-238496-m03 && echo "ha-238496-m03" | sudo tee /etc/hostname
I0729 23:18:17.290998 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: ha-238496-m03
I0729 23:18:17.291025 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHHostname
I0729 23:18:17.294658 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.295379 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:17.295410 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.295626 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHPort
I0729 23:18:17.295831 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:17.296004 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:17.296193 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHUsername
I0729 23:18:17.296419 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:18:17.296659 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.149 22 <nil> <nil>}
I0729 23:18:17.296681 29396 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sha-238496-m03' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 ha-238496-m03/g' /etc/hosts;
else
echo '127.0.1.1 ha-238496-m03' | sudo tee -a /etc/hosts;
fi
fi
I0729 23:18:17.412566 29396 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0729 23:18:17.412594 29396 buildroot.go:172] set auth options {CertDir:/home/jenkins/minikube-integration/19347-12221/.minikube CaCertPath:/home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19347-12221/.minikube}
I0729 23:18:17.412611 29396 buildroot.go:174] setting up certificates
I0729 23:18:17.412623 29396 provision.go:84] configureAuth start
I0729 23:18:17.412631 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetMachineName
I0729 23:18:17.412886 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetIP
I0729 23:18:17.415634 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.415982 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:17.416009 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.416139 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHHostname
I0729 23:18:17.418291 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.418718 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:17.418745 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.418887 29396 provision.go:143] copyHostCerts
I0729 23:18:17.418920 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem -> /home/jenkins/minikube-integration/19347-12221/.minikube/ca.pem
I0729 23:18:17.418960 29396 exec_runner.go:144] found /home/jenkins/minikube-integration/19347-12221/.minikube/ca.pem, removing ...
I0729 23:18:17.418971 29396 exec_runner.go:203] rm: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.pem
I0729 23:18:17.419054 29396 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19347-12221/.minikube/ca.pem (1078 bytes)
I0729 23:18:17.419146 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem -> /home/jenkins/minikube-integration/19347-12221/.minikube/cert.pem
I0729 23:18:17.419169 29396 exec_runner.go:144] found /home/jenkins/minikube-integration/19347-12221/.minikube/cert.pem, removing ...
I0729 23:18:17.419178 29396 exec_runner.go:203] rm: /home/jenkins/minikube-integration/19347-12221/.minikube/cert.pem
I0729 23:18:17.419221 29396 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19347-12221/.minikube/cert.pem (1123 bytes)
I0729 23:18:17.419284 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/key.pem -> /home/jenkins/minikube-integration/19347-12221/.minikube/key.pem
I0729 23:18:17.419307 29396 exec_runner.go:144] found /home/jenkins/minikube-integration/19347-12221/.minikube/key.pem, removing ...
I0729 23:18:17.419316 29396 exec_runner.go:203] rm: /home/jenkins/minikube-integration/19347-12221/.minikube/key.pem
I0729 23:18:17.419353 29396 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19347-12221/.minikube/key.pem (1675 bytes)
I0729 23:18:17.419477 29396 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca-key.pem org=jenkins.ha-238496-m03 san=[127.0.0.1 192.168.39.149 ha-238496-m03 localhost minikube]
I0729 23:18:17.499973 29396 provision.go:177] copyRemoteCerts
I0729 23:18:17.500084 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0729 23:18:17.500119 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHHostname
I0729 23:18:17.502839 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.503240 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:17.503280 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.503394 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHPort
I0729 23:18:17.503579 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:17.503727 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHUsername
I0729 23:18:17.503848 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.149 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03/id_rsa Username:docker}
I0729 23:18:17.585149 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem -> /etc/docker/ca.pem
I0729 23:18:17.585226 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0729 23:18:17.611065 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server.pem -> /etc/docker/server.pem
I0729 23:18:17.611132 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0729 23:18:17.636605 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server-key.pem -> /etc/docker/server-key.pem
I0729 23:18:17.636684 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
I0729 23:18:17.663333 29396 provision.go:87] duration metric: took 250.697302ms to configureAuth
I0729 23:18:17.663364 29396 buildroot.go:189] setting minikube options for container-runtime
I0729 23:18:17.663658 29396 config.go:182] Loaded profile config "ha-238496": Driver=kvm2, ContainerRuntime=docker, KubernetesVersion=v1.30.3
I0729 23:18:17.663690 29396 main.go:141] libmachine: (ha-238496-m03) Calling .DriverName
I0729 23:18:17.664032 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHHostname
I0729 23:18:17.666831 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.667201 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:17.667236 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.667362 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHPort
I0729 23:18:17.667568 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:17.667730 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:17.667864 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHUsername
I0729 23:18:17.668006 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:18:17.668186 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.149 22 <nil> <nil>}
I0729 23:18:17.668197 29396 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0729 23:18:17.772931 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: tmpfs
I0729 23:18:17.772958 29396 buildroot.go:70] root file system type: tmpfs
I0729 23:18:17.773073 29396 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0729 23:18:17.773094 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHHostname
I0729 23:18:17.775780 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.776131 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:17.776168 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.776334 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHPort
I0729 23:18:17.776561 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:17.776761 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:17.776903 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHUsername
I0729 23:18:17.777084 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:18:17.777237 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.149 22 <nil> <nil>}
I0729 23:18:17.777296 29396 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %!s(MISSING) "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network.target minikube-automount.service docker.socket
Requires= minikube-automount.service docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
Environment="NO_PROXY=192.168.39.113"
Environment="NO_PROXY=192.168.39.113,192.168.39.226"
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=kvm2 --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0729 23:18:17.894853 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network.target minikube-automount.service docker.socket
Requires= minikube-automount.service docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
Environment=NO_PROXY=192.168.39.113
Environment=NO_PROXY=192.168.39.113,192.168.39.226
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=kvm2 --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0729 23:18:17.894900 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHHostname
I0729 23:18:17.897632 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.898057 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:17.898091 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:17.898268 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHPort
I0729 23:18:17.898549 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:17.898768 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:17.898941 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHUsername
I0729 23:18:17.899117 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:18:17.899270 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.149 22 <nil> <nil>}
I0729 23:18:17.899286 29396 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0729 23:18:19.710533 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: diff: can't stat '/lib/systemd/system/docker.service': No such file or directory
Created symlink /etc/systemd/system/multi-user.target.wants/docker.service → /usr/lib/systemd/system/docker.service.
I0729 23:18:19.710562 29396 main.go:141] libmachine: Checking connection to Docker...
I0729 23:18:19.710570 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetURL
I0729 23:18:19.712009 29396 main.go:141] libmachine: (ha-238496-m03) DBG | Using libvirt version 6000000
I0729 23:18:19.714322 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.714682 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:19.714714 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.714913 29396 main.go:141] libmachine: Docker is up and running!
I0729 23:18:19.714930 29396 main.go:141] libmachine: Reticulating splines...
I0729 23:18:19.714938 29396 client.go:171] duration metric: took 23.196430062s to LocalClient.Create
I0729 23:18:19.714965 29396 start.go:167] duration metric: took 23.196496052s to libmachine.API.Create "ha-238496"
I0729 23:18:19.714977 29396 start.go:293] postStartSetup for "ha-238496-m03" (driver="kvm2")
I0729 23:18:19.714992 29396 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0729 23:18:19.715025 29396 main.go:141] libmachine: (ha-238496-m03) Calling .DriverName
I0729 23:18:19.715294 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0729 23:18:19.715321 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHHostname
I0729 23:18:19.717345 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.717724 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:19.717752 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.717859 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHPort
I0729 23:18:19.718034 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:19.718204 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHUsername
I0729 23:18:19.718352 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.149 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03/id_rsa Username:docker}
I0729 23:18:19.801578 29396 ssh_runner.go:195] Run: cat /etc/os-release
I0729 23:18:19.806033 29396 info.go:137] Remote host: Buildroot 2023.02.9
I0729 23:18:19.806058 29396 filesync.go:126] Scanning /home/jenkins/minikube-integration/19347-12221/.minikube/addons for local assets ...
I0729 23:18:19.806117 29396 filesync.go:126] Scanning /home/jenkins/minikube-integration/19347-12221/.minikube/files for local assets ...
I0729 23:18:19.806183 29396 filesync.go:149] local asset: /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem -> 194112.pem in /etc/ssl/certs
I0729 23:18:19.806192 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem -> /etc/ssl/certs/194112.pem
I0729 23:18:19.806266 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I0729 23:18:19.815398 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem --> /etc/ssl/certs/194112.pem (1708 bytes)
I0729 23:18:19.841583 29396 start.go:296] duration metric: took 126.592964ms for postStartSetup
I0729 23:18:19.841628 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetConfigRaw
I0729 23:18:19.842233 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetIP
I0729 23:18:19.844912 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.845270 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:19.845300 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.845620 29396 profile.go:143] Saving config to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/config.json ...
I0729 23:18:19.845824 29396 start.go:128] duration metric: took 23.345491779s to createHost
I0729 23:18:19.845846 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHHostname
I0729 23:18:19.848061 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.848372 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:19.848398 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.848564 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHPort
I0729 23:18:19.848749 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:19.848909 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:19.849065 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHUsername
I0729 23:18:19.849191 29396 main.go:141] libmachine: Using SSH client type: native
I0729 23:18:19.849338 29396 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da80] 0x8307e0 <nil> [] 0s} 192.168.39.149 22 <nil> <nil>}
I0729 23:18:19.849348 29396 main.go:141] libmachine: About to run SSH command:
date +%!s(MISSING).%!N(MISSING)
I0729 23:18:19.951990 29396 main.go:141] libmachine: SSH cmd err, output: <nil>: 1722295099.917890654
I0729 23:18:19.952014 29396 fix.go:216] guest clock: 1722295099.917890654
I0729 23:18:19.952023 29396 fix.go:229] Guest: 2024-07-29 23:18:19.917890654 +0000 UTC Remote: 2024-07-29 23:18:19.84583587 +0000 UTC m=+183.867783749 (delta=72.054784ms)
I0729 23:18:19.952043 29396 fix.go:200] guest clock delta is within tolerance: 72.054784ms
I0729 23:18:19.952048 29396 start.go:83] releasing machines lock for "ha-238496-m03", held for 23.451873031s
I0729 23:18:19.952066 29396 main.go:141] libmachine: (ha-238496-m03) Calling .DriverName
I0729 23:18:19.952316 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetIP
I0729 23:18:19.955220 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.955681 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:19.955705 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.958083 29396 out.go:177] * Found network options:
I0729 23:18:19.959273 29396 out.go:177] - NO_PROXY=192.168.39.113,192.168.39.226
W0729 23:18:19.960362 29396 proxy.go:119] fail to check proxy env: Error ip not in block
W0729 23:18:19.960382 29396 proxy.go:119] fail to check proxy env: Error ip not in block
I0729 23:18:19.960394 29396 main.go:141] libmachine: (ha-238496-m03) Calling .DriverName
I0729 23:18:19.960914 29396 main.go:141] libmachine: (ha-238496-m03) Calling .DriverName
I0729 23:18:19.961090 29396 main.go:141] libmachine: (ha-238496-m03) Calling .DriverName
I0729 23:18:19.961186 29396 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0729 23:18:19.961222 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHHostname
W0729 23:18:19.961278 29396 proxy.go:119] fail to check proxy env: Error ip not in block
W0729 23:18:19.961299 29396 proxy.go:119] fail to check proxy env: Error ip not in block
I0729 23:18:19.961355 29396 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0729 23:18:19.961370 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHHostname
I0729 23:18:19.964054 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.964314 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.964345 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:19.964367 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.964513 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHPort
I0729 23:18:19.964682 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:19.964779 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:19.964808 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:19.964835 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHUsername
I0729 23:18:19.964951 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHPort
I0729 23:18:19.965021 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.149 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03/id_rsa Username:docker}
I0729 23:18:19.965091 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHKeyPath
I0729 23:18:19.965232 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetSSHUsername
I0729 23:18:19.965377 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.149 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496-m03/id_rsa Username:docker}
W0729 23:18:20.069265 29396 cni.go:209] loopback cni configuration skipped: "/etc/cni/net.d/*loopback.conf*" not found
I0729 23:18:20.069346 29396 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%!p(MISSING), " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0729 23:18:20.089100 29396 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist] bridge cni config(s)
I0729 23:18:20.089127 29396 start.go:495] detecting cgroup driver to use...
I0729 23:18:20.089226 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0729 23:18:20.109423 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.9"|' /etc/containerd/config.toml"
I0729 23:18:20.121185 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0729 23:18:20.131667 29396 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0729 23:18:20.131729 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0729 23:18:20.142107 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0729 23:18:20.152727 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0729 23:18:20.162722 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0729 23:18:20.173052 29396 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0729 23:18:20.183164 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0729 23:18:20.193942 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0729 23:18:20.204204 29396 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0729 23:18:20.216044 29396 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0729 23:18:20.225885 29396 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0729 23:18:20.235268 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:18:20.354306 29396 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0729 23:18:20.386442 29396 start.go:495] detecting cgroup driver to use...
I0729 23:18:20.386526 29396 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0729 23:18:20.403972 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I0729 23:18:20.436001 29396 ssh_runner.go:195] Run: sudo systemctl stop -f containerd
I0729 23:18:20.467406 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I0729 23:18:20.484807 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0729 23:18:20.506289 29396 ssh_runner.go:195] Run: sudo systemctl stop -f crio
I0729 23:18:20.539471 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0729 23:18:20.553836 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0729 23:18:20.573444 29396 ssh_runner.go:195] Run: which cri-dockerd
I0729 23:18:20.578089 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0729 23:18:20.588175 29396 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (189 bytes)
I0729 23:18:20.607230 29396 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0729 23:18:20.725875 29396 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0729 23:18:20.855232 29396 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0729 23:18:20.855280 29396 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0729 23:18:20.874090 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:18:20.994222 29396 ssh_runner.go:195] Run: sudo systemctl restart docker
I0729 23:18:23.368531 29396 ssh_runner.go:235] Completed: sudo systemctl restart docker: (2.374269914s)
I0729 23:18:23.368592 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0729 23:18:23.383584 29396 ssh_runner.go:195] Run: sudo systemctl stop cri-docker.socket
I0729 23:18:23.403170 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0729 23:18:23.418207 29396 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0729 23:18:23.541217 29396 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0729 23:18:23.670819 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:18:23.810131 29396 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0729 23:18:23.827814 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0729 23:18:23.842063 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:18:23.956037 29396 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0729 23:18:24.041700 29396 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0729 23:18:24.041767 29396 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0729 23:18:24.048237 29396 start.go:563] Will wait 60s for crictl version
I0729 23:18:24.048285 29396 ssh_runner.go:195] Run: which crictl
I0729 23:18:24.053205 29396 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0729 23:18:24.096189 29396 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.1.0
RuntimeApiVersion: v1
I0729 23:18:24.096276 29396 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0729 23:18:24.127515 29396 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0729 23:18:24.155063 29396 out.go:204] * Preparing Kubernetes v1.30.3 on Docker 27.1.0 ...
I0729 23:18:24.156415 29396 out.go:177] - env NO_PROXY=192.168.39.113
I0729 23:18:24.157636 29396 out.go:177] - env NO_PROXY=192.168.39.113,192.168.39.226
I0729 23:18:24.158734 29396 main.go:141] libmachine: (ha-238496-m03) Calling .GetIP
I0729 23:18:24.161448 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:24.161874 29396 main.go:141] libmachine: (ha-238496-m03) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:34:73:00", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:18:11 +0000 UTC Type:0 Mac:52:54:00:34:73:00 Iaid: IPaddr:192.168.39.149 Prefix:24 Hostname:ha-238496-m03 Clientid:01:52:54:00:34:73:00}
I0729 23:18:24.161903 29396 main.go:141] libmachine: (ha-238496-m03) DBG | domain ha-238496-m03 has defined IP address 192.168.39.149 and MAC address 52:54:00:34:73:00 in network mk-ha-238496
I0729 23:18:24.162093 29396 ssh_runner.go:195] Run: grep 192.168.39.1 host.minikube.internal$ /etc/hosts
I0729 23:18:24.166564 29396 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.39.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0729 23:18:24.180873 29396 mustload.go:65] Loading cluster: ha-238496
I0729 23:18:24.181150 29396 config.go:182] Loaded profile config "ha-238496": Driver=kvm2, ContainerRuntime=docker, KubernetesVersion=v1.30.3
I0729 23:18:24.181509 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:18:24.181550 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:18:24.196403 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:45253
I0729 23:18:24.196811 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:18:24.197247 29396 main.go:141] libmachine: Using API Version 1
I0729 23:18:24.197268 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:18:24.197582 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:18:24.197801 29396 main.go:141] libmachine: (ha-238496) Calling .GetState
I0729 23:18:24.199552 29396 host.go:66] Checking if "ha-238496" exists ...
I0729 23:18:24.199959 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:18:24.200004 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:18:24.215836 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:38527
I0729 23:18:24.216217 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:18:24.216659 29396 main.go:141] libmachine: Using API Version 1
I0729 23:18:24.216681 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:18:24.217013 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:18:24.217183 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:18:24.217358 29396 certs.go:68] Setting up /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496 for IP: 192.168.39.149
I0729 23:18:24.217378 29396 certs.go:194] generating shared ca certs ...
I0729 23:18:24.217391 29396 certs.go:226] acquiring lock for ca certs: {Name:mk651b4a346cb6b65a98f292d471b5ea2ee1b039 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:18:24.217531 29396 certs.go:235] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.key
I0729 23:18:24.217572 29396 certs.go:235] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.key
I0729 23:18:24.217581 29396 certs.go:256] generating profile certs ...
I0729 23:18:24.217646 29396 certs.go:359] skipping valid signed profile cert regeneration for "minikube-user": /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.key
I0729 23:18:24.217675 29396 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key.020f40f1
I0729 23:18:24.217695 29396 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt.020f40f1 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.39.113 192.168.39.226 192.168.39.149 192.168.39.254]
I0729 23:18:24.480368 29396 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt.020f40f1 ...
I0729 23:18:24.480396 29396 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt.020f40f1: {Name:mka76b94d67b685d0074ff48e14df385dc4b115d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:18:24.480554 29396 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key.020f40f1 ...
I0729 23:18:24.480564 29396 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key.020f40f1: {Name:mka13cda52a163fc14ec9c650a674d5e5ee78192 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0729 23:18:24.480632 29396 certs.go:381] copying /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt.020f40f1 -> /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt
I0729 23:18:24.480761 29396 certs.go:385] copying /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key.020f40f1 -> /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key
I0729 23:18:24.480882 29396 certs.go:359] skipping valid signed profile cert regeneration for "aggregator": /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.key
I0729 23:18:24.480896 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt -> /var/lib/minikube/certs/ca.crt
I0729 23:18:24.480909 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.key -> /var/lib/minikube/certs/ca.key
I0729 23:18:24.480922 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.crt -> /var/lib/minikube/certs/proxy-client-ca.crt
I0729 23:18:24.480934 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.key -> /var/lib/minikube/certs/proxy-client-ca.key
I0729 23:18:24.480950 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt -> /var/lib/minikube/certs/apiserver.crt
I0729 23:18:24.480962 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key -> /var/lib/minikube/certs/apiserver.key
I0729 23:18:24.480973 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.crt -> /var/lib/minikube/certs/proxy-client.crt
I0729 23:18:24.480987 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.key -> /var/lib/minikube/certs/proxy-client.key
I0729 23:18:24.481070 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/19411.pem (1338 bytes)
W0729 23:18:24.481110 29396 certs.go:480] ignoring /home/jenkins/minikube-integration/19347-12221/.minikube/certs/19411_empty.pem, impossibly tiny 0 bytes
I0729 23:18:24.481122 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca-key.pem (1679 bytes)
I0729 23:18:24.481155 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/ca.pem (1078 bytes)
I0729 23:18:24.481184 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/cert.pem (1123 bytes)
I0729 23:18:24.481218 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/key.pem (1675 bytes)
I0729 23:18:24.481273 29396 certs.go:484] found cert: /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem (1708 bytes)
I0729 23:18:24.481309 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem -> /usr/share/ca-certificates/194112.pem
I0729 23:18:24.481328 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt -> /usr/share/ca-certificates/minikubeCA.pem
I0729 23:18:24.481343 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/certs/19411.pem -> /usr/share/ca-certificates/19411.pem
I0729 23:18:24.481373 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:18:24.484304 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:18:24.484716 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:18:24.484740 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:18:24.484903 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:18:24.485204 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:18:24.485367 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:18:24.485521 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.113 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/id_rsa Username:docker}
I0729 23:18:24.563047 29396 ssh_runner.go:195] Run: stat -c %!s(MISSING) /var/lib/minikube/certs/sa.pub
I0729 23:18:24.572327 29396 ssh_runner.go:447] scp /var/lib/minikube/certs/sa.pub --> memory (451 bytes)
I0729 23:18:24.585774 29396 ssh_runner.go:195] Run: stat -c %!s(MISSING) /var/lib/minikube/certs/sa.key
I0729 23:18:24.591059 29396 ssh_runner.go:447] scp /var/lib/minikube/certs/sa.key --> memory (1675 bytes)
I0729 23:18:24.603264 29396 ssh_runner.go:195] Run: stat -c %!s(MISSING) /var/lib/minikube/certs/front-proxy-ca.crt
I0729 23:18:24.608189 29396 ssh_runner.go:447] scp /var/lib/minikube/certs/front-proxy-ca.crt --> memory (1123 bytes)
I0729 23:18:24.621107 29396 ssh_runner.go:195] Run: stat -c %!s(MISSING) /var/lib/minikube/certs/front-proxy-ca.key
I0729 23:18:24.625692 29396 ssh_runner.go:447] scp /var/lib/minikube/certs/front-proxy-ca.key --> memory (1679 bytes)
I0729 23:18:24.637355 29396 ssh_runner.go:195] Run: stat -c %!s(MISSING) /var/lib/minikube/certs/etcd/ca.crt
I0729 23:18:24.643721 29396 ssh_runner.go:447] scp /var/lib/minikube/certs/etcd/ca.crt --> memory (1094 bytes)
I0729 23:18:24.661989 29396 ssh_runner.go:195] Run: stat -c %!s(MISSING) /var/lib/minikube/certs/etcd/ca.key
I0729 23:18:24.668318 29396 ssh_runner.go:447] scp /var/lib/minikube/certs/etcd/ca.key --> memory (1679 bytes)
I0729 23:18:24.678812 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0729 23:18:24.708772 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I0729 23:18:24.734095 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0729 23:18:24.761797 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I0729 23:18:24.791993 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1444 bytes)
I0729 23:18:24.818897 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
I0729 23:18:24.843320 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0729 23:18:24.868133 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0729 23:18:24.894007 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/files/etc/ssl/certs/194112.pem --> /usr/share/ca-certificates/194112.pem (1708 bytes)
I0729 23:18:24.919386 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0729 23:18:24.944533 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/certs/19411.pem --> /usr/share/ca-certificates/19411.pem (1338 bytes)
I0729 23:18:24.969936 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/certs/sa.pub (451 bytes)
I0729 23:18:24.988206 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/certs/sa.key (1675 bytes)
I0729 23:18:25.005201 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/certs/front-proxy-ca.crt (1123 bytes)
I0729 23:18:25.024627 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/certs/front-proxy-ca.key (1679 bytes)
I0729 23:18:25.043229 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/certs/etcd/ca.crt (1094 bytes)
I0729 23:18:25.063048 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/certs/etcd/ca.key (1679 bytes)
I0729 23:18:25.082802 29396 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (744 bytes)
I0729 23:18:25.102348 29396 ssh_runner.go:195] Run: openssl version
I0729 23:18:25.109028 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/194112.pem && ln -fs /usr/share/ca-certificates/194112.pem /etc/ssl/certs/194112.pem"
I0729 23:18:25.121686 29396 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/194112.pem
I0729 23:18:25.126649 29396 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 Jul 29 23:11 /usr/share/ca-certificates/194112.pem
I0729 23:18:25.126728 29396 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/194112.pem
I0729 23:18:25.132950 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/194112.pem /etc/ssl/certs/3ec20f2e.0"
I0729 23:18:25.145264 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0729 23:18:25.156705 29396 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0729 23:18:25.161408 29396 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Jul 29 23:03 /usr/share/ca-certificates/minikubeCA.pem
I0729 23:18:25.161468 29396 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0729 23:18:25.167328 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0729 23:18:25.178351 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/19411.pem && ln -fs /usr/share/ca-certificates/19411.pem /etc/ssl/certs/19411.pem"
I0729 23:18:25.190126 29396 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/19411.pem
I0729 23:18:25.194860 29396 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 Jul 29 23:11 /usr/share/ca-certificates/19411.pem
I0729 23:18:25.194913 29396 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/19411.pem
I0729 23:18:25.200760 29396 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/19411.pem /etc/ssl/certs/51391683.0"
I0729 23:18:25.212002 29396 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0729 23:18:25.216439 29396 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0729 23:18:25.216486 29396 kubeadm.go:934] updating node {m03 192.168.39.149 8443 v1.30.3 docker true true} ...
I0729 23:18:25.216561 29396 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.30.3/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=ha-238496-m03 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.39.149
[Install]
config:
{KubernetesVersion:v1.30.3 ClusterName:ha-238496 Namespace:default APIServerHAVIP:192.168.39.254 APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0729 23:18:25.216583 29396 kube-vip.go:115] generating kube-vip config ...
I0729 23:18:25.216614 29396 ssh_runner.go:195] Run: sudo sh -c "modprobe --all ip_vs ip_vs_rr ip_vs_wrr ip_vs_sh nf_conntrack"
I0729 23:18:25.232406 29396 kube-vip.go:167] auto-enabling control-plane load-balancing in kube-vip
I0729 23:18:25.232476 29396 kube-vip.go:137] kube-vip config:
apiVersion: v1
kind: Pod
metadata:
creationTimestamp: null
name: kube-vip
namespace: kube-system
spec:
containers:
- args:
- manager
env:
- name: vip_arp
value: "true"
- name: port
value: "8443"
- name: vip_nodename
valueFrom:
fieldRef:
fieldPath: spec.nodeName
- name: vip_interface
value: eth0
- name: vip_cidr
value: "32"
- name: dns_mode
value: first
- name: cp_enable
value: "true"
- name: cp_namespace
value: kube-system
- name: vip_leaderelection
value: "true"
- name: vip_leasename
value: plndr-cp-lock
- name: vip_leaseduration
value: "5"
- name: vip_renewdeadline
value: "3"
- name: vip_retryperiod
value: "1"
- name: address
value: 192.168.39.254
- name: prometheus_server
value: :2112
- name : lb_enable
value: "true"
- name: lb_port
value: "8443"
image: ghcr.io/kube-vip/kube-vip:v0.8.0
imagePullPolicy: IfNotPresent
name: kube-vip
resources: {}
securityContext:
capabilities:
add:
- NET_ADMIN
- NET_RAW
volumeMounts:
- mountPath: /etc/kubernetes/admin.conf
name: kubeconfig
hostAliases:
- hostnames:
- kubernetes
ip: 127.0.0.1
hostNetwork: true
volumes:
- hostPath:
path: "/etc/kubernetes/admin.conf"
name: kubeconfig
status: {}
I0729 23:18:25.232536 29396 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.30.3
I0729 23:18:25.242069 29396 binaries.go:47] Didn't find k8s binaries: sudo ls /var/lib/minikube/binaries/v1.30.3: Process exited with status 2
stdout:
stderr:
ls: cannot access '/var/lib/minikube/binaries/v1.30.3': No such file or directory
Initiating transfer...
I0729 23:18:25.242137 29396 ssh_runner.go:195] Run: sudo mkdir -p /var/lib/minikube/binaries/v1.30.3
I0729 23:18:25.251750 29396 binary.go:74] Not caching binary, using https://dl.k8s.io/release/v1.30.3/bin/linux/amd64/kubectl?checksum=file:https://dl.k8s.io/release/v1.30.3/bin/linux/amd64/kubectl.sha256
I0729 23:18:25.251771 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubectl -> /var/lib/minikube/binaries/v1.30.3/kubectl
I0729 23:18:25.251797 29396 binary.go:74] Not caching binary, using https://dl.k8s.io/release/v1.30.3/bin/linux/amd64/kubelet?checksum=file:https://dl.k8s.io/release/v1.30.3/bin/linux/amd64/kubelet.sha256
I0729 23:18:25.251829 29396 binary.go:74] Not caching binary, using https://dl.k8s.io/release/v1.30.3/bin/linux/amd64/kubeadm?checksum=file:https://dl.k8s.io/release/v1.30.3/bin/linux/amd64/kubeadm.sha256
I0729 23:18:25.251839 29396 ssh_runner.go:195] Run: stat -c "%!s(MISSING) %!y(MISSING)" /var/lib/minikube/binaries/v1.30.3/kubectl
I0729 23:18:25.251840 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0729 23:18:25.251844 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubeadm -> /var/lib/minikube/binaries/v1.30.3/kubeadm
I0729 23:18:25.251900 29396 ssh_runner.go:195] Run: stat -c "%!s(MISSING) %!y(MISSING)" /var/lib/minikube/binaries/v1.30.3/kubeadm
I0729 23:18:25.257314 29396 ssh_runner.go:352] existence check for /var/lib/minikube/binaries/v1.30.3/kubectl: stat -c "%!s(MISSING) %!y(MISSING)" /var/lib/minikube/binaries/v1.30.3/kubectl: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/binaries/v1.30.3/kubectl': No such file or directory
I0729 23:18:25.257333 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubectl --> /var/lib/minikube/binaries/v1.30.3/kubectl (51454104 bytes)
I0729 23:18:25.279779 29396 ssh_runner.go:352] existence check for /var/lib/minikube/binaries/v1.30.3/kubeadm: stat -c "%!s(MISSING) %!y(MISSING)" /var/lib/minikube/binaries/v1.30.3/kubeadm: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/binaries/v1.30.3/kubeadm': No such file or directory
I0729 23:18:25.279809 29396 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubelet -> /var/lib/minikube/binaries/v1.30.3/kubelet
I0729 23:18:25.279816 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubeadm --> /var/lib/minikube/binaries/v1.30.3/kubeadm (50249880 bytes)
I0729 23:18:25.279914 29396 ssh_runner.go:195] Run: stat -c "%!s(MISSING) %!y(MISSING)" /var/lib/minikube/binaries/v1.30.3/kubelet
I0729 23:18:25.327896 29396 ssh_runner.go:352] existence check for /var/lib/minikube/binaries/v1.30.3/kubelet: stat -c "%!s(MISSING) %!y(MISSING)" /var/lib/minikube/binaries/v1.30.3/kubelet: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/binaries/v1.30.3/kubelet': No such file or directory
I0729 23:18:25.327938 29396 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19347-12221/.minikube/cache/linux/amd64/v1.30.3/kubelet --> /var/lib/minikube/binaries/v1.30.3/kubelet (100125080 bytes)
I0729 23:18:26.135622 29396 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /etc/kubernetes/manifests
I0729 23:18:26.145637 29396 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (314 bytes)
I0729 23:18:26.163912 29396 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0729 23:18:26.181593 29396 ssh_runner.go:362] scp memory --> /etc/kubernetes/manifests/kube-vip.yaml (1441 bytes)
I0729 23:18:26.199220 29396 ssh_runner.go:195] Run: grep 192.168.39.254 control-plane.minikube.internal$ /etc/hosts
I0729 23:18:26.203822 29396 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.39.254 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0729 23:18:26.217025 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:18:26.342359 29396 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0729 23:18:26.368665 29396 host.go:66] Checking if "ha-238496" exists ...
I0729 23:18:26.369015 29396 main.go:141] libmachine: Found binary path at /home/jenkins/workspace/KVM_Linux_integration/out/docker-machine-driver-kvm2
I0729 23:18:26.369067 29396 main.go:141] libmachine: Launching plugin server for driver kvm2
I0729 23:18:26.386170 29396 main.go:141] libmachine: Plugin server listening at address 127.0.0.1:41791
I0729 23:18:26.386689 29396 main.go:141] libmachine: () Calling .GetVersion
I0729 23:18:26.387240 29396 main.go:141] libmachine: Using API Version 1
I0729 23:18:26.387261 29396 main.go:141] libmachine: () Calling .SetConfigRaw
I0729 23:18:26.387672 29396 main.go:141] libmachine: () Calling .GetMachineName
I0729 23:18:26.387888 29396 main.go:141] libmachine: (ha-238496) Calling .DriverName
I0729 23:18:26.388031 29396 start.go:317] joinCluster: &{Name:ha-238496 KeepContext:false EmbedCerts:false MinikubeISO:https://storage.googleapis.com/minikube-builds/iso/19319/minikube-v1.33.1-1721690939-19319-amd64.iso KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721902582-19326@sha256:540fb5dc7f38be17ff5276a38dfe6c8a4b1d9ba1c27c62244e6eebd7e37696e7 Memory:2200 CPUs:2 DiskSize:20000 Driver:kvm2 HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.30.3 Cluster
Name:ha-238496 Namespace:default APIServerHAVIP:192.168.39.254 APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.39.113 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true} {Name:m02 IP:192.168.39.226 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true} {Name:m03 IP:192.168.39.149 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[ambassador:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:false efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false helm-tiller:false inaccel:false ingress:false ingress-dn
s:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:false storage-provisioner-gluster:false storage-provisioner-rancher:false volcano:false volumesnapshots:false yakd:false] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:true ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror:
DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0729 23:18:26.388163 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.30.3:$PATH" kubeadm token create --print-join-command --ttl=0"
I0729 23:18:26.388187 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHHostname
I0729 23:18:26.391336 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:18:26.391895 29396 main.go:141] libmachine: (ha-238496) DBG | found host DHCP lease matching {name: "", mac: "52:54:00:4c:48:55", ip: ""} in network mk-ha-238496: {Iface:virbr1 ExpiryTime:2024-07-30 00:15:30 +0000 UTC Type:0 Mac:52:54:00:4c:48:55 Iaid: IPaddr:192.168.39.113 Prefix:24 Hostname:ha-238496 Clientid:01:52:54:00:4c:48:55}
I0729 23:18:26.391918 29396 main.go:141] libmachine: (ha-238496) DBG | domain ha-238496 has defined IP address 192.168.39.113 and MAC address 52:54:00:4c:48:55 in network mk-ha-238496
I0729 23:18:26.392077 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHPort
I0729 23:18:26.392263 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHKeyPath
I0729 23:18:26.392415 29396 main.go:141] libmachine: (ha-238496) Calling .GetSSHUsername
I0729 23:18:26.392560 29396 sshutil.go:53] new ssh client: &{IP:192.168.39.113 Port:22 SSHKeyPath:/home/jenkins/minikube-integration/19347-12221/.minikube/machines/ha-238496/id_rsa Username:docker}
I0729 23:18:26.581813 29396 start.go:343] trying to join control-plane node "m03" to cluster: &{Name:m03 IP:192.168.39.149 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}
I0729 23:18:26.581869 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.30.3:$PATH" kubeadm join control-plane.minikube.internal:8443 --token 0ywfv2.rkmgx63wf72zwqqk --discovery-token-ca-cert-hash sha256:da4124175dbd4d7966590c68bf3c2627d9fda969ee89096732ee7fd4a463dd4a --ignore-preflight-errors=all --cri-socket unix:///var/run/cri-dockerd.sock --node-name=ha-238496-m03 --control-plane --apiserver-advertise-address=192.168.39.149 --apiserver-bind-port=8443"
I0729 23:18:51.475405 29396 ssh_runner.go:235] Completed: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.30.3:$PATH" kubeadm join control-plane.minikube.internal:8443 --token 0ywfv2.rkmgx63wf72zwqqk --discovery-token-ca-cert-hash sha256:da4124175dbd4d7966590c68bf3c2627d9fda969ee89096732ee7fd4a463dd4a --ignore-preflight-errors=all --cri-socket unix:///var/run/cri-dockerd.sock --node-name=ha-238496-m03 --control-plane --apiserver-advertise-address=192.168.39.149 --apiserver-bind-port=8443": (24.893510323s)
I0729 23:18:51.475451 29396 ssh_runner.go:195] Run: /bin/bash -c "sudo systemctl daemon-reload && sudo systemctl enable kubelet && sudo systemctl start kubelet"
I0729 23:18:52.014804 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes ha-238496-m03 minikube.k8s.io/updated_at=2024_07_29T23_18_52_0700 minikube.k8s.io/version=v1.33.1 minikube.k8s.io/commit=b13baeaf4895dcc6a8c5d0ab64a27ff86dff4ae3 minikube.k8s.io/name=ha-238496 minikube.k8s.io/primary=false
I0729 23:18:52.155300 29396 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig taint nodes ha-238496-m03 node-role.kubernetes.io/control-plane:NoSchedule-
I0729 23:18:52.304779 29396 start.go:319] duration metric: took 25.916745009s to joinCluster
I0729 23:18:52.304850 29396 start.go:235] Will wait 6m0s for node &{Name:m03 IP:192.168.39.149 Port:8443 KubernetesVersion:v1.30.3 ContainerRuntime:docker ControlPlane:true Worker:true}
I0729 23:18:52.305212 29396 config.go:182] Loaded profile config "ha-238496": Driver=kvm2, ContainerRuntime=docker, KubernetesVersion=v1.30.3
I0729 23:18:52.306430 29396 out.go:177] * Verifying Kubernetes components...
I0729 23:18:52.307833 29396 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0729 23:18:52.570958 29396 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0729 23:18:52.598459 29396 loader.go:395] Config loaded from file: /home/jenkins/minikube-integration/19347-12221/kubeconfig
I0729 23:18:52.598826 29396 kapi.go:59] client config for ha-238496: &rest.Config{Host:"https://192.168.39.254:8443", APIPath:"", ContentConfig:rest.ContentConfig{AcceptContentTypes:"", ContentType:"", GroupVersion:(*schema.GroupVersion)(nil), NegotiatedSerializer:runtime.NegotiatedSerializer(nil)}, Username:"", Password:"", BearerToken:"", BearerTokenFile:"", Impersonate:rest.ImpersonationConfig{UserName:"", UID:"", Groups:[]string(nil), Extra:map[string][]string(nil)}, AuthProvider:<nil>, AuthConfigPersister:rest.AuthProviderConfigPersister(nil), ExecProvider:<nil>, TLSClientConfig:rest.sanitizedTLSClientConfig{Insecure:false, ServerName:"", CertFile:"/home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.crt", KeyFile:"/home/jenkins/minikube-integration/19347-12221/.minikube/profiles/ha-238496/client.key", CAFile:"/home/jenkins/minikube-integration/19347-12221/.minikube/ca.crt", CertData:[]uint8(nil), KeyData:[]uint8(nil), CAData:[]uint8(nil), NextProtos:[]string(nil)
}, UserAgent:"", DisableCompression:false, Transport:http.RoundTripper(nil), WrapTransport:(transport.WrapperFunc)(0x1d02de0), QPS:0, Burst:0, RateLimiter:flowcontrol.RateLimiter(nil), WarningHandler:rest.WarningHandler(nil), Timeout:0, Dial:(func(context.Context, string, string) (net.Conn, error))(nil), Proxy:(func(*http.Request) (*url.URL, error))(nil)}
W0729 23:18:52.598909 29396 kubeadm.go:483] Overriding stale ClientConfig host https://192.168.39.254:8443 with https://192.168.39.113:8443
I0729 23:18:52.599174 29396 node_ready.go:35] waiting up to 6m0s for node "ha-238496-m03" to be "Ready" ...
I0729 23:18:52.599248 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:52.599255 29396 round_trippers.go:469] Request Headers:
I0729 23:18:52.599266 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:52.599273 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:52.602438 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:18:53.099412 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:53.099460 29396 round_trippers.go:469] Request Headers:
I0729 23:18:53.099472 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:53.099478 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:53.103215 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:18:53.600157 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:53.600181 29396 round_trippers.go:469] Request Headers:
I0729 23:18:53.600190 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:53.600197 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:53.605227 29396 round_trippers.go:574] Response Status: 200 OK in 5 milliseconds
I0729 23:18:54.099698 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:54.099721 29396 round_trippers.go:469] Request Headers:
I0729 23:18:54.099730 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:54.099734 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:54.103051 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:18:54.600168 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:54.600193 29396 round_trippers.go:469] Request Headers:
I0729 23:18:54.600204 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:54.600210 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:54.604181 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:18:54.604814 29396 node_ready.go:53] node "ha-238496-m03" has status "Ready":"False"
I0729 23:18:55.099901 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:55.099941 29396 round_trippers.go:469] Request Headers:
I0729 23:18:55.099951 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:55.099956 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:55.103175 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:18:55.599949 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:55.599969 29396 round_trippers.go:469] Request Headers:
I0729 23:18:55.599977 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:55.599981 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:55.604676 29396 round_trippers.go:574] Response Status: 200 OK in 4 milliseconds
I0729 23:18:56.099926 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:56.099949 29396 round_trippers.go:469] Request Headers:
I0729 23:18:56.099957 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:56.099962 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:56.103784 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:18:56.599928 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:56.599955 29396 round_trippers.go:469] Request Headers:
I0729 23:18:56.599964 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:56.599969 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:56.603571 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:18:57.099427 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:57.099449 29396 round_trippers.go:469] Request Headers:
I0729 23:18:57.099456 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:57.099460 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:57.103316 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:18:57.103854 29396 node_ready.go:53] node "ha-238496-m03" has status "Ready":"False"
I0729 23:18:57.599973 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:57.599999 29396 round_trippers.go:469] Request Headers:
I0729 23:18:57.600010 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:57.600016 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:57.603583 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:18:58.099476 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:58.099500 29396 round_trippers.go:469] Request Headers:
I0729 23:18:58.099510 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:58.099514 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:58.103276 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:18:58.600166 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:58.600188 29396 round_trippers.go:469] Request Headers:
I0729 23:18:58.600196 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:58.600201 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:58.604175 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:18:59.100227 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:59.100262 29396 round_trippers.go:469] Request Headers:
I0729 23:18:59.100271 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:59.100275 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:59.104347 29396 round_trippers.go:574] Response Status: 200 OK in 4 milliseconds
I0729 23:18:59.106420 29396 node_ready.go:53] node "ha-238496-m03" has status "Ready":"False"
I0729 23:18:59.600223 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:18:59.600244 29396 round_trippers.go:469] Request Headers:
I0729 23:18:59.600252 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:18:59.600257 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:18:59.603629 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:00.100154 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:00.100174 29396 round_trippers.go:469] Request Headers:
I0729 23:19:00.100182 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:00.100185 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:00.103877 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:00.599346 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:00.599368 29396 round_trippers.go:469] Request Headers:
I0729 23:19:00.599374 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:00.599378 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:00.603739 29396 round_trippers.go:574] Response Status: 200 OK in 4 milliseconds
I0729 23:19:01.100139 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:01.100165 29396 round_trippers.go:469] Request Headers:
I0729 23:19:01.100178 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:01.100183 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:01.103691 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:01.599585 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:01.599606 29396 round_trippers.go:469] Request Headers:
I0729 23:19:01.599614 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:01.599618 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:01.603036 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:01.603692 29396 node_ready.go:53] node "ha-238496-m03" has status "Ready":"False"
I0729 23:19:02.099930 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:02.099953 29396 round_trippers.go:469] Request Headers:
I0729 23:19:02.099961 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:02.099964 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:02.103792 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:02.600386 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:02.600405 29396 round_trippers.go:469] Request Headers:
I0729 23:19:02.600414 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:02.600423 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:02.604228 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:03.099945 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:03.099969 29396 round_trippers.go:469] Request Headers:
I0729 23:19:03.099979 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:03.099987 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:03.103141 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:03.599963 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:03.599984 29396 round_trippers.go:469] Request Headers:
I0729 23:19:03.599991 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:03.599997 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:03.603708 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:03.604435 29396 node_ready.go:53] node "ha-238496-m03" has status "Ready":"False"
I0729 23:19:04.099425 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:04.099446 29396 round_trippers.go:469] Request Headers:
I0729 23:19:04.099462 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:04.099467 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:04.102745 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:04.600170 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:04.600195 29396 round_trippers.go:469] Request Headers:
I0729 23:19:04.600205 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:04.600210 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:04.604071 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:05.099965 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:05.099992 29396 round_trippers.go:469] Request Headers:
I0729 23:19:05.100004 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:05.100010 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:05.103443 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:05.599703 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:05.599733 29396 round_trippers.go:469] Request Headers:
I0729 23:19:05.599741 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:05.599745 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:05.602758 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:19:06.100313 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:06.100339 29396 round_trippers.go:469] Request Headers:
I0729 23:19:06.100349 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:06.100356 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:06.104074 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:06.104560 29396 node_ready.go:53] node "ha-238496-m03" has status "Ready":"False"
I0729 23:19:06.599776 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:06.599799 29396 round_trippers.go:469] Request Headers:
I0729 23:19:06.599807 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:06.599811 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:06.603694 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:07.100397 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:07.100425 29396 round_trippers.go:469] Request Headers:
I0729 23:19:07.100436 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:07.100443 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:07.109921 29396 round_trippers.go:574] Response Status: 200 OK in 9 milliseconds
I0729 23:19:07.599360 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:07.599382 29396 round_trippers.go:469] Request Headers:
I0729 23:19:07.599390 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:07.599394 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:07.603023 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:08.099904 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:08.099925 29396 round_trippers.go:469] Request Headers:
I0729 23:19:08.099932 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:08.099936 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:08.103187 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:08.599540 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:08.599562 29396 round_trippers.go:469] Request Headers:
I0729 23:19:08.599570 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:08.599576 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:08.602493 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:19:08.603114 29396 node_ready.go:53] node "ha-238496-m03" has status "Ready":"False"
I0729 23:19:09.099505 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:09.099526 29396 round_trippers.go:469] Request Headers:
I0729 23:19:09.099533 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:09.099536 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:09.217759 29396 round_trippers.go:574] Response Status: 200 OK in 118 milliseconds
I0729 23:19:09.599507 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:09.599529 29396 round_trippers.go:469] Request Headers:
I0729 23:19:09.599536 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:09.599540 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:09.603233 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:09.603757 29396 node_ready.go:49] node "ha-238496-m03" has status "Ready":"True"
I0729 23:19:09.603776 29396 node_ready.go:38] duration metric: took 17.004589012s for node "ha-238496-m03" to be "Ready" ...
I0729 23:19:09.603784 29396 pod_ready.go:35] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0729 23:19:09.603836 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods
I0729 23:19:09.603846 29396 round_trippers.go:469] Request Headers:
I0729 23:19:09.603855 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:09.603863 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:09.611049 29396 round_trippers.go:574] Response Status: 200 OK in 7 milliseconds
I0729 23:19:09.617888 29396 pod_ready.go:78] waiting up to 6m0s for pod "coredns-7db6d8ff4d-p8nps" in "kube-system" namespace to be "Ready" ...
I0729 23:19:09.617981 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/coredns-7db6d8ff4d-p8nps
I0729 23:19:09.617989 29396 round_trippers.go:469] Request Headers:
I0729 23:19:09.617997 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:09.618001 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:09.621617 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:09.622359 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:19:09.622375 29396 round_trippers.go:469] Request Headers:
I0729 23:19:09.622383 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:09.622388 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:09.625417 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:09.626016 29396 pod_ready.go:92] pod "coredns-7db6d8ff4d-p8nps" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:09.626037 29396 pod_ready.go:81] duration metric: took 8.124272ms for pod "coredns-7db6d8ff4d-p8nps" in "kube-system" namespace to be "Ready" ...
I0729 23:19:09.626046 29396 pod_ready.go:78] waiting up to 6m0s for pod "coredns-7db6d8ff4d-tjplq" in "kube-system" namespace to be "Ready" ...
I0729 23:19:09.626117 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/coredns-7db6d8ff4d-tjplq
I0729 23:19:09.626125 29396 round_trippers.go:469] Request Headers:
I0729 23:19:09.626139 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:09.626146 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:09.628706 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:19:09.629438 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:19:09.629451 29396 round_trippers.go:469] Request Headers:
I0729 23:19:09.629464 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:09.629468 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:09.632161 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:19:09.632598 29396 pod_ready.go:92] pod "coredns-7db6d8ff4d-tjplq" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:09.632614 29396 pod_ready.go:81] duration metric: took 6.559167ms for pod "coredns-7db6d8ff4d-tjplq" in "kube-system" namespace to be "Ready" ...
I0729 23:19:09.632622 29396 pod_ready.go:78] waiting up to 6m0s for pod "etcd-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:19:09.632663 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/etcd-ha-238496
I0729 23:19:09.632671 29396 round_trippers.go:469] Request Headers:
I0729 23:19:09.632677 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:09.632683 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:09.635207 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:19:09.635748 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:19:09.635761 29396 round_trippers.go:469] Request Headers:
I0729 23:19:09.635769 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:09.635774 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:09.638168 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:19:09.638558 29396 pod_ready.go:92] pod "etcd-ha-238496" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:09.638574 29396 pod_ready.go:81] duration metric: took 5.946799ms for pod "etcd-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:19:09.638585 29396 pod_ready.go:78] waiting up to 6m0s for pod "etcd-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:19:09.638634 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/etcd-ha-238496-m02
I0729 23:19:09.638641 29396 round_trippers.go:469] Request Headers:
I0729 23:19:09.638648 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:09.638653 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:09.641255 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:19:09.641916 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:19:09.641934 29396 round_trippers.go:469] Request Headers:
I0729 23:19:09.641944 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:09.641952 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:09.644523 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:19:09.644949 29396 pod_ready.go:92] pod "etcd-ha-238496-m02" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:09.644964 29396 pod_ready.go:81] duration metric: took 6.369287ms for pod "etcd-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:19:09.644973 29396 pod_ready.go:78] waiting up to 6m0s for pod "etcd-ha-238496-m03" in "kube-system" namespace to be "Ready" ...
I0729 23:19:09.800329 29396 request.go:629] Waited for 155.300695ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/etcd-ha-238496-m03
I0729 23:19:09.800403 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/etcd-ha-238496-m03
I0729 23:19:09.800414 29396 round_trippers.go:469] Request Headers:
I0729 23:19:09.800423 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:09.800447 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:09.803688 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:09.999868 29396 request.go:629] Waited for 195.356437ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:09.999935 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:09.999940 29396 round_trippers.go:469] Request Headers:
I0729 23:19:09.999947 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:09.999951 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:10.003518 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:10.004201 29396 pod_ready.go:92] pod "etcd-ha-238496-m03" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:10.004220 29396 pod_ready.go:81] duration metric: took 359.241717ms for pod "etcd-ha-238496-m03" in "kube-system" namespace to be "Ready" ...
I0729 23:19:10.004248 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-apiserver-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:19:10.200290 29396 request.go:629] Waited for 195.933294ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-apiserver-ha-238496
I0729 23:19:10.200365 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-apiserver-ha-238496
I0729 23:19:10.200371 29396 round_trippers.go:469] Request Headers:
I0729 23:19:10.200379 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:10.200384 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:10.208712 29396 round_trippers.go:574] Response Status: 200 OK in 8 milliseconds
I0729 23:19:10.399884 29396 request.go:629] Waited for 190.181377ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:19:10.399943 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:19:10.399950 29396 round_trippers.go:469] Request Headers:
I0729 23:19:10.399960 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:10.399966 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:10.403257 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:10.403958 29396 pod_ready.go:92] pod "kube-apiserver-ha-238496" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:10.403980 29396 pod_ready.go:81] duration metric: took 399.717532ms for pod "kube-apiserver-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:19:10.403994 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-apiserver-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:19:10.600125 29396 request.go:629] Waited for 196.04647ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-apiserver-ha-238496-m02
I0729 23:19:10.600187 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-apiserver-ha-238496-m02
I0729 23:19:10.600192 29396 round_trippers.go:469] Request Headers:
I0729 23:19:10.600201 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:10.600206 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:10.603755 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:10.799857 29396 request.go:629] Waited for 195.304515ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:19:10.799951 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:19:10.799960 29396 round_trippers.go:469] Request Headers:
I0729 23:19:10.799968 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:10.799974 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:10.804771 29396 round_trippers.go:574] Response Status: 200 OK in 4 milliseconds
I0729 23:19:10.805297 29396 pod_ready.go:92] pod "kube-apiserver-ha-238496-m02" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:10.805316 29396 pod_ready.go:81] duration metric: took 401.313343ms for pod "kube-apiserver-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:19:10.805328 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-apiserver-ha-238496-m03" in "kube-system" namespace to be "Ready" ...
I0729 23:19:11.000424 29396 request.go:629] Waited for 195.030165ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-apiserver-ha-238496-m03
I0729 23:19:11.000489 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-apiserver-ha-238496-m03
I0729 23:19:11.000513 29396 round_trippers.go:469] Request Headers:
I0729 23:19:11.000523 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:11.000526 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:11.003767 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:11.200256 29396 request.go:629] Waited for 195.566068ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:11.200305 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:11.200310 29396 round_trippers.go:469] Request Headers:
I0729 23:19:11.200317 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:11.200321 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:11.207986 29396 round_trippers.go:574] Response Status: 200 OK in 7 milliseconds
I0729 23:19:11.208915 29396 pod_ready.go:92] pod "kube-apiserver-ha-238496-m03" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:11.208933 29396 pod_ready.go:81] duration metric: took 403.597736ms for pod "kube-apiserver-ha-238496-m03" in "kube-system" namespace to be "Ready" ...
I0729 23:19:11.208943 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-controller-manager-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:19:11.399869 29396 request.go:629] Waited for 190.871481ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-controller-manager-ha-238496
I0729 23:19:11.399925 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-controller-manager-ha-238496
I0729 23:19:11.399932 29396 round_trippers.go:469] Request Headers:
I0729 23:19:11.399942 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:11.399950 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:11.403414 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:11.600424 29396 request.go:629] Waited for 196.336161ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:19:11.600474 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:19:11.600480 29396 round_trippers.go:469] Request Headers:
I0729 23:19:11.600490 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:11.600500 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:11.603696 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:11.604331 29396 pod_ready.go:92] pod "kube-controller-manager-ha-238496" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:11.604348 29396 pod_ready.go:81] duration metric: took 395.398637ms for pod "kube-controller-manager-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:19:11.604357 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-controller-manager-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:19:11.799625 29396 request.go:629] Waited for 194.825123ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-controller-manager-ha-238496-m02
I0729 23:19:11.799685 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-controller-manager-ha-238496-m02
I0729 23:19:11.799690 29396 round_trippers.go:469] Request Headers:
I0729 23:19:11.799697 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:11.799702 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:11.805192 29396 round_trippers.go:574] Response Status: 200 OK in 5 milliseconds
I0729 23:19:12.000327 29396 request.go:629] Waited for 194.372047ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:19:12.000404 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:19:12.000412 29396 round_trippers.go:469] Request Headers:
I0729 23:19:12.000420 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:12.000429 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:12.003948 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:12.004664 29396 pod_ready.go:92] pod "kube-controller-manager-ha-238496-m02" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:12.004682 29396 pod_ready.go:81] duration metric: took 400.318127ms for pod "kube-controller-manager-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:19:12.004692 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-controller-manager-ha-238496-m03" in "kube-system" namespace to be "Ready" ...
I0729 23:19:12.199772 29396 request.go:629] Waited for 195.023091ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-controller-manager-ha-238496-m03
I0729 23:19:12.199839 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-controller-manager-ha-238496-m03
I0729 23:19:12.199845 29396 round_trippers.go:469] Request Headers:
I0729 23:19:12.199853 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:12.199861 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:12.208978 29396 round_trippers.go:574] Response Status: 200 OK in 9 milliseconds
I0729 23:19:12.399970 29396 request.go:629] Waited for 188.017146ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:12.400027 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:12.400034 29396 round_trippers.go:469] Request Headers:
I0729 23:19:12.400044 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:12.400049 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:12.403826 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:12.404648 29396 pod_ready.go:92] pod "kube-controller-manager-ha-238496-m03" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:12.404664 29396 pod_ready.go:81] duration metric: took 399.96617ms for pod "kube-controller-manager-ha-238496-m03" in "kube-system" namespace to be "Ready" ...
I0729 23:19:12.404675 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-proxy-84q2j" in "kube-system" namespace to be "Ready" ...
I0729 23:19:12.599837 29396 request.go:629] Waited for 195.088918ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-proxy-84q2j
I0729 23:19:12.599902 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-proxy-84q2j
I0729 23:19:12.599908 29396 round_trippers.go:469] Request Headers:
I0729 23:19:12.599916 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:12.599922 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:12.603764 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:12.799903 29396 request.go:629] Waited for 195.280664ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:12.799988 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:12.799999 29396 round_trippers.go:469] Request Headers:
I0729 23:19:12.800010 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:12.800017 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:12.804112 29396 round_trippers.go:574] Response Status: 200 OK in 4 milliseconds
I0729 23:19:12.804755 29396 pod_ready.go:92] pod "kube-proxy-84q2j" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:12.804771 29396 pod_ready.go:81] duration metric: took 400.090247ms for pod "kube-proxy-84q2j" in "kube-system" namespace to be "Ready" ...
I0729 23:19:12.804784 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-proxy-m6vdn" in "kube-system" namespace to be "Ready" ...
I0729 23:19:12.999973 29396 request.go:629] Waited for 195.096525ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-proxy-m6vdn
I0729 23:19:13.000025 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-proxy-m6vdn
I0729 23:19:13.000030 29396 round_trippers.go:469] Request Headers:
I0729 23:19:13.000038 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:13.000043 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:13.004206 29396 round_trippers.go:574] Response Status: 200 OK in 4 milliseconds
I0729 23:19:13.200315 29396 request.go:629] Waited for 195.344861ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:19:13.200387 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:19:13.200394 29396 round_trippers.go:469] Request Headers:
I0729 23:19:13.200403 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:13.200407 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:13.203669 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:13.204391 29396 pod_ready.go:92] pod "kube-proxy-m6vdn" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:13.204408 29396 pod_ready.go:81] duration metric: took 399.614059ms for pod "kube-proxy-m6vdn" in "kube-system" namespace to be "Ready" ...
I0729 23:19:13.204418 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-proxy-nrvw6" in "kube-system" namespace to be "Ready" ...
I0729 23:19:13.399561 29396 request.go:629] Waited for 195.057489ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-proxy-nrvw6
I0729 23:19:13.399619 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-proxy-nrvw6
I0729 23:19:13.399625 29396 round_trippers.go:469] Request Headers:
I0729 23:19:13.399633 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:13.399641 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:13.403311 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:13.600350 29396 request.go:629] Waited for 196.355182ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:19:13.600420 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:19:13.600426 29396 round_trippers.go:469] Request Headers:
I0729 23:19:13.600433 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:13.600438 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:13.603539 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:13.604101 29396 pod_ready.go:92] pod "kube-proxy-nrvw6" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:13.604130 29396 pod_ready.go:81] duration metric: took 399.697487ms for pod "kube-proxy-nrvw6" in "kube-system" namespace to be "Ready" ...
I0729 23:19:13.604154 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-scheduler-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:19:13.800215 29396 request.go:629] Waited for 196.003393ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-scheduler-ha-238496
I0729 23:19:13.800264 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-scheduler-ha-238496
I0729 23:19:13.800269 29396 round_trippers.go:469] Request Headers:
I0729 23:19:13.800276 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:13.800282 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:13.803643 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:14.000583 29396 request.go:629] Waited for 196.347088ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:19:14.000640 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496
I0729 23:19:14.000646 29396 round_trippers.go:469] Request Headers:
I0729 23:19:14.000656 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:14.000662 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:14.003455 29396 round_trippers.go:574] Response Status: 200 OK in 2 milliseconds
I0729 23:19:14.004025 29396 pod_ready.go:92] pod "kube-scheduler-ha-238496" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:14.004043 29396 pod_ready.go:81] duration metric: took 399.883328ms for pod "kube-scheduler-ha-238496" in "kube-system" namespace to be "Ready" ...
I0729 23:19:14.004051 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-scheduler-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:19:14.200236 29396 request.go:629] Waited for 196.127033ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-scheduler-ha-238496-m02
I0729 23:19:14.200346 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-scheduler-ha-238496-m02
I0729 23:19:14.200366 29396 round_trippers.go:469] Request Headers:
I0729 23:19:14.200377 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:14.200384 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:14.210764 29396 round_trippers.go:574] Response Status: 200 OK in 10 milliseconds
I0729 23:19:14.399682 29396 request.go:629] Waited for 188.173442ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:19:14.399751 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m02
I0729 23:19:14.399756 29396 round_trippers.go:469] Request Headers:
I0729 23:19:14.399764 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:14.399771 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:14.403486 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:14.404167 29396 pod_ready.go:92] pod "kube-scheduler-ha-238496-m02" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:14.404189 29396 pod_ready.go:81] duration metric: took 400.130141ms for pod "kube-scheduler-ha-238496-m02" in "kube-system" namespace to be "Ready" ...
I0729 23:19:14.404202 29396 pod_ready.go:78] waiting up to 6m0s for pod "kube-scheduler-ha-238496-m03" in "kube-system" namespace to be "Ready" ...
I0729 23:19:14.599520 29396 request.go:629] Waited for 195.261081ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-scheduler-ha-238496-m03
I0729 23:19:14.599572 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods/kube-scheduler-ha-238496-m03
I0729 23:19:14.599577 29396 round_trippers.go:469] Request Headers:
I0729 23:19:14.599584 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:14.599592 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:14.602991 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:14.800140 29396 request.go:629] Waited for 196.358765ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:14.800203 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes/ha-238496-m03
I0729 23:19:14.800215 29396 round_trippers.go:469] Request Headers:
I0729 23:19:14.800226 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:14.800232 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:14.803666 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:14.804301 29396 pod_ready.go:92] pod "kube-scheduler-ha-238496-m03" in "kube-system" namespace has status "Ready":"True"
I0729 23:19:14.804318 29396 pod_ready.go:81] duration metric: took 400.109552ms for pod "kube-scheduler-ha-238496-m03" in "kube-system" namespace to be "Ready" ...
I0729 23:19:14.804328 29396 pod_ready.go:38] duration metric: took 5.200535213s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0729 23:19:14.804344 29396 api_server.go:52] waiting for apiserver process to appear ...
I0729 23:19:14.804391 29396 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0729 23:19:14.822172 29396 api_server.go:72] duration metric: took 22.517280061s to wait for apiserver process to appear ...
I0729 23:19:14.822199 29396 api_server.go:88] waiting for apiserver healthz status ...
I0729 23:19:14.822218 29396 api_server.go:253] Checking apiserver healthz at https://192.168.39.113:8443/healthz ...
I0729 23:19:14.826236 29396 api_server.go:279] https://192.168.39.113:8443/healthz returned 200:
ok
I0729 23:19:14.826287 29396 round_trippers.go:463] GET https://192.168.39.113:8443/version
I0729 23:19:14.826292 29396 round_trippers.go:469] Request Headers:
I0729 23:19:14.826299 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:14.826304 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:14.827085 29396 round_trippers.go:574] Response Status: 200 OK in 0 milliseconds
I0729 23:19:14.827244 29396 api_server.go:141] control plane version: v1.30.3
I0729 23:19:14.827263 29396 api_server.go:131] duration metric: took 5.057172ms to wait for apiserver health ...
I0729 23:19:14.827272 29396 system_pods.go:43] waiting for kube-system pods to appear ...
I0729 23:19:14.999578 29396 request.go:629] Waited for 172.242999ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods
I0729 23:19:14.999628 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods
I0729 23:19:14.999655 29396 round_trippers.go:469] Request Headers:
I0729 23:19:14.999665 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:14.999671 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:15.006220 29396 round_trippers.go:574] Response Status: 200 OK in 6 milliseconds
I0729 23:19:15.012395 29396 system_pods.go:59] 24 kube-system pods found
I0729 23:19:15.012427 29396 system_pods.go:61] "coredns-7db6d8ff4d-p8nps" [af3f5c7b-1996-497f-95f7-4bfc87392dc7] Running
I0729 23:19:15.012434 29396 system_pods.go:61] "coredns-7db6d8ff4d-tjplq" [db7a6b8c-bfe3-4291-bf9a-9ce96bb5b0b7] Running
I0729 23:19:15.012439 29396 system_pods.go:61] "etcd-ha-238496" [ed3a1237-a4c1-4e3f-b7d6-6b5237f7a18b] Running
I0729 23:19:15.012443 29396 system_pods.go:61] "etcd-ha-238496-m02" [0a4d5ebc-a7be-445f-bdfc-47b3b1c01803] Running
I0729 23:19:15.012448 29396 system_pods.go:61] "etcd-ha-238496-m03" [8cc4bf64-609d-4cf2-b8f6-e0f660e4428c] Running
I0729 23:19:15.012452 29396 system_pods.go:61] "kindnet-55jmm" [7ddd1f82-1105-4694-b8d6-5198fdbd1f86] Running
I0729 23:19:15.012459 29396 system_pods.go:61] "kindnet-kb2hw" [ef875a41-530f-48ba-b034-d08a8a7acbbc] Running
I0729 23:19:15.012464 29396 system_pods.go:61] "kindnet-xvzff" [400a9d4f-d218-443e-b001-edd5e5fd5af7] Running
I0729 23:19:15.012470 29396 system_pods.go:61] "kube-apiserver-ha-238496" [54eebf95-2bd3-4c57-9794-170fccda1dbb] Running
I0729 23:19:15.012475 29396 system_pods.go:61] "kube-apiserver-ha-238496-m02" [66429444-6c99-474c-9294-c569e1a5cc46] Running
I0729 23:19:15.012483 29396 system_pods.go:61] "kube-apiserver-ha-238496-m03" [fe9eddc6-6bb4-4f78-891d-e5830247246f] Running
I0729 23:19:15.012490 29396 system_pods.go:61] "kube-controller-manager-ha-238496" [bb6bc2ad-54ec-42fa-8f18-e33cb50a8ce8] Running
I0729 23:19:15.012498 29396 system_pods.go:61] "kube-controller-manager-ha-238496-m02" [8836c211-ee9d-403a-8383-333c22f1b945] Running
I0729 23:19:15.012503 29396 system_pods.go:61] "kube-controller-manager-ha-238496-m03" [8e748a5e-733a-4be7-896f-0501c2d63ab9] Running
I0729 23:19:15.012508 29396 system_pods.go:61] "kube-proxy-84q2j" [4a6fb431-510a-4ecb-a8d3-e595512e0e52] Running
I0729 23:19:15.012514 29396 system_pods.go:61] "kube-proxy-m6vdn" [f3731d91-d919-4f7f-a7b9-2bf7ba93569b] Running
I0729 23:19:15.012521 29396 system_pods.go:61] "kube-proxy-nrvw6" [708cca57-5274-4ad9-871c-048f24b43a33] Running
I0729 23:19:15.012525 29396 system_pods.go:61] "kube-scheduler-ha-238496" [b4999631-2ffc-4684-ab41-7e065cbbe74b] Running
I0729 23:19:15.012531 29396 system_pods.go:61] "kube-scheduler-ha-238496-m02" [4eb7be71-6cad-4260-a4c0-6a97011e6ec5] Running
I0729 23:19:15.012536 29396 system_pods.go:61] "kube-scheduler-ha-238496-m03" [bc034dc2-6055-4edd-90f2-7c80b18c5842] Running
I0729 23:19:15.012541 29396 system_pods.go:61] "kube-vip-ha-238496" [f248f380-c48b-451a-82e7-0aeb1e0ba6eb] Running
I0729 23:19:15.012547 29396 system_pods.go:61] "kube-vip-ha-238496-m02" [39a50caf-f960-4d68-9235-d6dacace51c1] Running
I0729 23:19:15.012553 29396 system_pods.go:61] "kube-vip-ha-238496-m03" [317c63a1-f4c1-44b7-8f97-3a5c01f9a64e] Running
I0729 23:19:15.012558 29396 system_pods.go:61] "storage-provisioner" [2feba04d-7105-41cd-b308-747ed0079849] Running
I0729 23:19:15.012565 29396 system_pods.go:74] duration metric: took 185.282872ms to wait for pod list to return data ...
I0729 23:19:15.012577 29396 default_sa.go:34] waiting for default service account to be created ...
I0729 23:19:15.199847 29396 request.go:629] Waited for 187.195853ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/default/serviceaccounts
I0729 23:19:15.199896 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/default/serviceaccounts
I0729 23:19:15.199901 29396 round_trippers.go:469] Request Headers:
I0729 23:19:15.199908 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:15.199914 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:15.203366 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:15.203475 29396 default_sa.go:45] found service account: "default"
I0729 23:19:15.203489 29396 default_sa.go:55] duration metric: took 190.906753ms for default service account to be created ...
I0729 23:19:15.203497 29396 system_pods.go:116] waiting for k8s-apps to be running ...
I0729 23:19:15.399849 29396 request.go:629] Waited for 196.29656ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods
I0729 23:19:15.399915 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/namespaces/kube-system/pods
I0729 23:19:15.399920 29396 round_trippers.go:469] Request Headers:
I0729 23:19:15.399928 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:15.399933 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:15.406727 29396 round_trippers.go:574] Response Status: 200 OK in 6 milliseconds
I0729 23:19:15.413271 29396 system_pods.go:86] 24 kube-system pods found
I0729 23:19:15.413300 29396 system_pods.go:89] "coredns-7db6d8ff4d-p8nps" [af3f5c7b-1996-497f-95f7-4bfc87392dc7] Running
I0729 23:19:15.413306 29396 system_pods.go:89] "coredns-7db6d8ff4d-tjplq" [db7a6b8c-bfe3-4291-bf9a-9ce96bb5b0b7] Running
I0729 23:19:15.413310 29396 system_pods.go:89] "etcd-ha-238496" [ed3a1237-a4c1-4e3f-b7d6-6b5237f7a18b] Running
I0729 23:19:15.413314 29396 system_pods.go:89] "etcd-ha-238496-m02" [0a4d5ebc-a7be-445f-bdfc-47b3b1c01803] Running
I0729 23:19:15.413319 29396 system_pods.go:89] "etcd-ha-238496-m03" [8cc4bf64-609d-4cf2-b8f6-e0f660e4428c] Running
I0729 23:19:15.413322 29396 system_pods.go:89] "kindnet-55jmm" [7ddd1f82-1105-4694-b8d6-5198fdbd1f86] Running
I0729 23:19:15.413326 29396 system_pods.go:89] "kindnet-kb2hw" [ef875a41-530f-48ba-b034-d08a8a7acbbc] Running
I0729 23:19:15.413330 29396 system_pods.go:89] "kindnet-xvzff" [400a9d4f-d218-443e-b001-edd5e5fd5af7] Running
I0729 23:19:15.413334 29396 system_pods.go:89] "kube-apiserver-ha-238496" [54eebf95-2bd3-4c57-9794-170fccda1dbb] Running
I0729 23:19:15.413338 29396 system_pods.go:89] "kube-apiserver-ha-238496-m02" [66429444-6c99-474c-9294-c569e1a5cc46] Running
I0729 23:19:15.413343 29396 system_pods.go:89] "kube-apiserver-ha-238496-m03" [fe9eddc6-6bb4-4f78-891d-e5830247246f] Running
I0729 23:19:15.413347 29396 system_pods.go:89] "kube-controller-manager-ha-238496" [bb6bc2ad-54ec-42fa-8f18-e33cb50a8ce8] Running
I0729 23:19:15.413351 29396 system_pods.go:89] "kube-controller-manager-ha-238496-m02" [8836c211-ee9d-403a-8383-333c22f1b945] Running
I0729 23:19:15.413356 29396 system_pods.go:89] "kube-controller-manager-ha-238496-m03" [8e748a5e-733a-4be7-896f-0501c2d63ab9] Running
I0729 23:19:15.413360 29396 system_pods.go:89] "kube-proxy-84q2j" [4a6fb431-510a-4ecb-a8d3-e595512e0e52] Running
I0729 23:19:15.413363 29396 system_pods.go:89] "kube-proxy-m6vdn" [f3731d91-d919-4f7f-a7b9-2bf7ba93569b] Running
I0729 23:19:15.413370 29396 system_pods.go:89] "kube-proxy-nrvw6" [708cca57-5274-4ad9-871c-048f24b43a33] Running
I0729 23:19:15.413375 29396 system_pods.go:89] "kube-scheduler-ha-238496" [b4999631-2ffc-4684-ab41-7e065cbbe74b] Running
I0729 23:19:15.413383 29396 system_pods.go:89] "kube-scheduler-ha-238496-m02" [4eb7be71-6cad-4260-a4c0-6a97011e6ec5] Running
I0729 23:19:15.413387 29396 system_pods.go:89] "kube-scheduler-ha-238496-m03" [bc034dc2-6055-4edd-90f2-7c80b18c5842] Running
I0729 23:19:15.413394 29396 system_pods.go:89] "kube-vip-ha-238496" [f248f380-c48b-451a-82e7-0aeb1e0ba6eb] Running
I0729 23:19:15.413397 29396 system_pods.go:89] "kube-vip-ha-238496-m02" [39a50caf-f960-4d68-9235-d6dacace51c1] Running
I0729 23:19:15.413401 29396 system_pods.go:89] "kube-vip-ha-238496-m03" [317c63a1-f4c1-44b7-8f97-3a5c01f9a64e] Running
I0729 23:19:15.413404 29396 system_pods.go:89] "storage-provisioner" [2feba04d-7105-41cd-b308-747ed0079849] Running
I0729 23:19:15.413412 29396 system_pods.go:126] duration metric: took 209.91061ms to wait for k8s-apps to be running ...
I0729 23:19:15.413422 29396 system_svc.go:44] waiting for kubelet service to be running ....
I0729 23:19:15.413463 29396 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0729 23:19:15.431449 29396 system_svc.go:56] duration metric: took 18.020128ms WaitForService to wait for kubelet
I0729 23:19:15.431477 29396 kubeadm.go:582] duration metric: took 23.126590826s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0729 23:19:15.431495 29396 node_conditions.go:102] verifying NodePressure condition ...
I0729 23:19:15.599855 29396 request.go:629] Waited for 168.298521ms due to client-side throttling, not priority and fairness, request: GET:https://192.168.39.113:8443/api/v1/nodes
I0729 23:19:15.599919 29396 round_trippers.go:463] GET https://192.168.39.113:8443/api/v1/nodes
I0729 23:19:15.599925 29396 round_trippers.go:469] Request Headers:
I0729 23:19:15.599932 29396 round_trippers.go:473] Accept: application/json, */*
I0729 23:19:15.599939 29396 round_trippers.go:473] User-Agent: minikube-linux-amd64/v0.0.0 (linux/amd64) kubernetes/$Format
I0729 23:19:15.603551 29396 round_trippers.go:574] Response Status: 200 OK in 3 milliseconds
I0729 23:19:15.604409 29396 node_conditions.go:122] node storage ephemeral capacity is 17734596Ki
I0729 23:19:15.604432 29396 node_conditions.go:123] node cpu capacity is 2
I0729 23:19:15.604444 29396 node_conditions.go:122] node storage ephemeral capacity is 17734596Ki
I0729 23:19:15.604454 29396 node_conditions.go:123] node cpu capacity is 2
I0729 23:19:15.604459 29396 node_conditions.go:122] node storage ephemeral capacity is 17734596Ki
I0729 23:19:15.604464 29396 node_conditions.go:123] node cpu capacity is 2
I0729 23:19:15.604470 29396 node_conditions.go:105] duration metric: took 172.970183ms to run NodePressure ...
I0729 23:19:15.604485 29396 start.go:241] waiting for startup goroutines ...
I0729 23:19:15.604505 29396 start.go:255] writing updated cluster config ...
I0729 23:19:15.604809 29396 ssh_runner.go:195] Run: rm -f paused
I0729 23:19:15.655539 29396 start.go:600] kubectl: 1.30.3, cluster: 1.30.3 (minor skew: 0)
I0729 23:19:15.657726 29396 out.go:177] * Done! kubectl is now configured to use "ha-238496" cluster and "default" namespace by default
==> Docker <==
Jul 29 23:16:37 ha-238496 cri-dockerd[1092]: time="2024-07-29T23:16:37Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/582b2c7baceb501e49656618efcbf0f27619a18145c3ff398c3b789cc9bfdf95/resolv.conf as [nameserver 192.168.122.1]"
Jul 29 23:16:37 ha-238496 cri-dockerd[1092]: time="2024-07-29T23:16:37Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/a014bae0f7e88a3f1ec0a6a6132b4f98aa48560f60c128e668b46efd5c355bd7/resolv.conf as [nameserver 192.168.122.1]"
Jul 29 23:16:37 ha-238496 cri-dockerd[1092]: time="2024-07-29T23:16:37Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/a9397d0ccb3a1e588ab8cda3b68fb14860b17dc14f00ccc03418139af4c83cc8/resolv.conf as [nameserver 192.168.122.1]"
Jul 29 23:16:37 ha-238496 dockerd[1202]: time="2024-07-29T23:16:37.484476445Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul 29 23:16:37 ha-238496 dockerd[1202]: time="2024-07-29T23:16:37.485248709Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul 29 23:16:37 ha-238496 dockerd[1202]: time="2024-07-29T23:16:37.485357241Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul 29 23:16:37 ha-238496 dockerd[1202]: time="2024-07-29T23:16:37.485728790Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul 29 23:16:37 ha-238496 dockerd[1202]: time="2024-07-29T23:16:37.638588001Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul 29 23:16:37 ha-238496 dockerd[1202]: time="2024-07-29T23:16:37.639074699Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul 29 23:16:37 ha-238496 dockerd[1202]: time="2024-07-29T23:16:37.639295310Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul 29 23:16:37 ha-238496 dockerd[1202]: time="2024-07-29T23:16:37.639593280Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul 29 23:16:37 ha-238496 dockerd[1202]: time="2024-07-29T23:16:37.656417401Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul 29 23:16:37 ha-238496 dockerd[1202]: time="2024-07-29T23:16:37.656485114Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul 29 23:16:37 ha-238496 dockerd[1202]: time="2024-07-29T23:16:37.656499490Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul 29 23:16:37 ha-238496 dockerd[1202]: time="2024-07-29T23:16:37.656574401Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul 29 23:19:18 ha-238496 dockerd[1202]: time="2024-07-29T23:19:18.936226883Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul 29 23:19:18 ha-238496 dockerd[1202]: time="2024-07-29T23:19:18.936852960Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul 29 23:19:18 ha-238496 dockerd[1202]: time="2024-07-29T23:19:18.936964837Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul 29 23:19:18 ha-238496 dockerd[1202]: time="2024-07-29T23:19:18.937632906Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul 29 23:19:19 ha-238496 cri-dockerd[1092]: time="2024-07-29T23:19:19Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/aa1e1b89fad706a2cc592ca475ad4636a8b6fff855df129a27b09aaece54827a/resolv.conf as [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local options ndots:5]"
Jul 29 23:19:21 ha-238496 cri-dockerd[1092]: time="2024-07-29T23:19:21Z" level=info msg="Stop pulling image gcr.io/k8s-minikube/busybox:1.28: Status: Downloaded newer image for gcr.io/k8s-minikube/busybox:1.28"
Jul 29 23:19:21 ha-238496 dockerd[1202]: time="2024-07-29T23:19:21.398061329Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul 29 23:19:21 ha-238496 dockerd[1202]: time="2024-07-29T23:19:21.398641089Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul 29 23:19:21 ha-238496 dockerd[1202]: time="2024-07-29T23:19:21.398740310Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul 29 23:19:21 ha-238496 dockerd[1202]: time="2024-07-29T23:19:21.399226781Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
b8a213421914f gcr.io/k8s-minikube/busybox@sha256:9afb80db71730dbb303fe00765cbf34bddbdc6b66e49897fc2e1861967584b12 33 seconds ago Running busybox 0 aa1e1b89fad70 busybox-fc5497c4f-ftt4w
dd28444131756 cbb01a7bd410d 3 minutes ago Running coredns 0 a9397d0ccb3a1 coredns-7db6d8ff4d-tjplq
29db34bf17cce cbb01a7bd410d 3 minutes ago Running coredns 0 a014bae0f7e88 coredns-7db6d8ff4d-p8nps
a338297585908 6e38f40d628db 3 minutes ago Running storage-provisioner 0 582b2c7baceb5 storage-provisioner
906ec8f4a7599 kindest/kindnetd@sha256:da8ad203ec15a72c313015e5609db44bfad7c95d8ce63e87ff97c66363b5680a 3 minutes ago Running kindnet-cni 0 4f259ecefd5fb kindnet-55jmm
2bb8727b4690e 55bb025d2cfa5 3 minutes ago Running kube-proxy 0 cb40e8b38f063 kube-proxy-nrvw6
4fec030f63653 ghcr.io/kube-vip/kube-vip@sha256:360f0c5d02322075cc80edb9e4e0d2171e941e55072184f1f902203fafc81d0f 3 minutes ago Running kube-vip 0 9e857d29c684c kube-vip-ha-238496
2127946fcd8b4 76932a3b37d7e 3 minutes ago Running kube-controller-manager 0 73f194ce0a0ec kube-controller-manager-ha-238496
dc0f824c8c08d 3edc18e7b7672 3 minutes ago Running kube-scheduler 0 675aa145ad460 kube-scheduler-ha-238496
189714a08644c 3861cfcd7c04c 3 minutes ago Running etcd 0 d32ae036bb1ae etcd-ha-238496
4607f65fdc744 1f6d574d502f3 3 minutes ago Running kube-apiserver 0 a1b03a6b48501 kube-apiserver-ha-238496
==> coredns [29db34bf17cc] <==
CoreDNS-1.11.1
linux/amd64, go1.20.7, ae2bbc2
[INFO] 127.0.0.1:33135 - 55754 "HINFO IN 7267661357375732516.6129737729494210875. udp 57 false 512" NXDOMAIN qr,rd,ra 57 0.018889362s
[INFO] 10.244.2.3:51780 - 2 "PTR IN 10.0.96.10.in-addr.arpa. udp 41 false 512" NOERROR qr,aa,rd 116 0.000748186s
[INFO] 10.244.2.3:59083 - 4 "A IN kubernetes.io. udp 31 false 512" NOERROR qr,rd,ra 60 0.004127652s
[INFO] 10.244.2.3:36153 - 5 "PTR IN 148.40.75.147.in-addr.arpa. udp 44 false 512" NXDOMAIN qr,rd,ra 44 0.0195213s
[INFO] 10.244.1.2:33877 - 4 "A IN kubernetes.io. udp 31 false 512" NOERROR qr,aa,rd,ra 60 0.000197751s
[INFO] 10.244.2.3:52428 - 4 "AAAA IN kubernetes.default.default.svc.cluster.local. udp 62 false 512" NXDOMAIN qr,aa,rd 155 0.000313348s
[INFO] 10.244.2.3:54129 - 5 "AAAA IN kubernetes.default.svc.cluster.local. udp 54 false 512" NOERROR qr,aa,rd 147 0.000283487s
[INFO] 10.244.2.3:46649 - 7 "A IN kubernetes.default.default.svc.cluster.local. udp 62 false 512" NXDOMAIN qr,aa,rd 155 0.000209103s
[INFO] 10.244.0.4:32930 - 3 "AAAA IN kubernetes.default. udp 36 false 512" NXDOMAIN qr,rd,ra 36 0.002157659s
[INFO] 10.244.0.4:38959 - 4 "AAAA IN kubernetes.default.default.svc.cluster.local. udp 62 false 512" NXDOMAIN qr,aa,rd 155 0.000129551s
[INFO] 10.244.0.4:55194 - 5 "AAAA IN kubernetes.default.svc.cluster.local. udp 54 false 512" NOERROR qr,aa,rd 147 0.000128726s
[INFO] 10.244.0.4:34700 - 6 "A IN kubernetes.default. udp 36 false 512" NXDOMAIN qr,rd,ra 36 0.001779956s
[INFO] 10.244.1.2:34614 - 3 "AAAA IN kubernetes.default. udp 36 false 512" NXDOMAIN qr,rd,ra 36 0.001969757s
[INFO] 10.244.1.2:46578 - 4 "AAAA IN kubernetes.default.default.svc.cluster.local. udp 62 false 512" NXDOMAIN qr,aa,rd 155 0.000086621s
[INFO] 10.244.1.2:38697 - 5 "AAAA IN kubernetes.default.svc.cluster.local. udp 54 false 512" NOERROR qr,aa,rd 147 0.000067185s
[INFO] 10.244.1.2:40745 - 7 "A IN kubernetes.default.default.svc.cluster.local. udp 62 false 512" NXDOMAIN qr,aa,rd 155 0.000079935s
[INFO] 10.244.1.2:54339 - 9 "PTR IN 1.0.96.10.in-addr.arpa. udp 40 false 512" NOERROR qr,aa,rd 112 0.000152584s
[INFO] 10.244.2.3:35899 - 2 "PTR IN 10.0.96.10.in-addr.arpa. udp 41 false 512" NOERROR qr,aa,rd 116 0.000140543s
[INFO] 10.244.2.3:55438 - 3 "AAAA IN kubernetes.default.svc.cluster.local. udp 54 false 512" NOERROR qr,aa,rd 147 0.000179287s
[INFO] 10.244.0.4:54155 - 4 "A IN kubernetes.default.svc.cluster.local. udp 54 false 512" NOERROR qr,aa,rd 106 0.000194153s
[INFO] 10.244.1.2:38738 - 2 "PTR IN 10.0.96.10.in-addr.arpa. udp 41 false 512" NOERROR qr,aa,rd 116 0.000120454s
[INFO] 10.244.1.2:35786 - 3 "AAAA IN kubernetes.default.svc.cluster.local. udp 54 false 512" NOERROR qr,aa,rd 147 0.000177128s
[INFO] 10.244.1.2:48640 - 4 "A IN kubernetes.default.svc.cluster.local. udp 54 false 512" NOERROR qr,aa,rd 106 0.000103188s
==> coredns [dd2844413175] <==
[INFO] 10.244.0.4:50889 - 2 "PTR IN 10.0.96.10.in-addr.arpa. udp 41 false 512" NOERROR qr,aa,rd 116 0.000219623s
[INFO] 10.244.0.4:50147 - 3 "AAAA IN kubernetes.io. udp 31 false 512" NOERROR qr,aa,rd,ra 31 0.000114376s
[INFO] 10.244.0.4:48383 - 4 "A IN kubernetes.io. udp 31 false 512" NOERROR qr,rd,ra 60 0.000465063s
[INFO] 10.244.0.4:57201 - 5 "PTR IN 148.40.75.147.in-addr.arpa. udp 44 false 512" NXDOMAIN qr,rd,ra 44 0.001758859s
[INFO] 10.244.1.2:34940 - 2 "PTR IN 10.0.96.10.in-addr.arpa. udp 41 false 512" NOERROR qr,aa,rd 116 0.000148699s
[INFO] 10.244.1.2:52125 - 3 "AAAA IN kubernetes.io. udp 31 false 512" NOERROR qr,aa,rd,ra 31 0.000150404s
[INFO] 10.244.1.2:33924 - 5 "PTR IN 148.40.75.147.in-addr.arpa. udp 44 false 512" NXDOMAIN qr,rd,ra 44 0.001932601s
[INFO] 10.244.2.3:51276 - 2 "PTR IN 10.0.96.10.in-addr.arpa. udp 41 false 512" NOERROR qr,aa,rd 116 0.000139359s
[INFO] 10.244.2.3:50554 - 3 "AAAA IN kubernetes.default. udp 36 false 512" NXDOMAIN qr,rd,ra 36 0.01541771s
[INFO] 10.244.2.3:59531 - 6 "A IN kubernetes.default. udp 36 false 512" NXDOMAIN qr,rd,ra 36 0.003229091s
[INFO] 10.244.2.3:37331 - 8 "A IN kubernetes.default.svc.cluster.local. udp 54 false 512" NOERROR qr,aa,rd 106 0.000183708s
[INFO] 10.244.2.3:60486 - 9 "PTR IN 1.0.96.10.in-addr.arpa. udp 40 false 512" NOERROR qr,aa,rd 112 0.000184885s
[INFO] 10.244.0.4:56715 - 2 "PTR IN 10.0.96.10.in-addr.arpa. udp 41 false 512" NOERROR qr,aa,rd 116 0.000098738s
[INFO] 10.244.0.4:58597 - 7 "A IN kubernetes.default.default.svc.cluster.local. udp 62 false 512" NXDOMAIN qr,aa,rd 155 0.000114041s
[INFO] 10.244.0.4:44865 - 8 "A IN kubernetes.default.svc.cluster.local. udp 54 false 512" NOERROR qr,aa,rd 106 0.000052207s
[INFO] 10.244.0.4:45588 - 9 "PTR IN 1.0.96.10.in-addr.arpa. udp 40 false 512" NOERROR qr,aa,rd 112 0.00005185s
[INFO] 10.244.1.2:58043 - 2 "PTR IN 10.0.96.10.in-addr.arpa. udp 41 false 512" NOERROR qr,aa,rd 116 0.000118248s
[INFO] 10.244.1.2:42031 - 6 "A IN kubernetes.default. udp 36 false 512" NXDOMAIN qr,rd,ra 36 0.001467571s
[INFO] 10.244.1.2:41248 - 8 "A IN kubernetes.default.svc.cluster.local. udp 54 false 512" NOERROR qr,aa,rd 106 0.00011067s
[INFO] 10.244.2.3:38627 - 4 "A IN kubernetes.default.svc.cluster.local. udp 54 false 512" NOERROR qr,aa,rd 106 0.000149238s
[INFO] 10.244.2.3:41829 - 5 "PTR IN 1.0.96.10.in-addr.arpa. udp 40 false 512" NOERROR qr,aa,rd 112 0.000147486s
[INFO] 10.244.0.4:57864 - 2 "PTR IN 10.0.96.10.in-addr.arpa. udp 41 false 512" NOERROR qr,aa,rd 116 0.000122317s
[INFO] 10.244.0.4:41410 - 3 "AAAA IN kubernetes.default.svc.cluster.local. udp 54 false 512" NOERROR qr,aa,rd 147 0.000171073s
[INFO] 10.244.0.4:40270 - 5 "PTR IN 1.0.96.10.in-addr.arpa. udp 40 false 512" NOERROR qr,aa,rd 112 0.000071428s
[INFO] 10.244.1.2:53470 - 5 "PTR IN 1.0.96.10.in-addr.arpa. udp 40 false 512" NOERROR qr,aa,rd 112 0.000130306s
==> describe nodes <==
Name: ha-238496
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=ha-238496
kubernetes.io/os=linux
minikube.k8s.io/commit=b13baeaf4895dcc6a8c5d0ab64a27ff86dff4ae3
minikube.k8s.io/name=ha-238496
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_07_29T23_16_07_0700
minikube.k8s.io/version=v1.33.1
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Mon, 29 Jul 2024 23:16:05 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: ha-238496
AcquireTime: <unset>
RenewTime: Mon, 29 Jul 2024 23:19:49 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Mon, 29 Jul 2024 23:19:40 +0000 Mon, 29 Jul 2024 23:16:05 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Mon, 29 Jul 2024 23:19:40 +0000 Mon, 29 Jul 2024 23:16:05 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Mon, 29 Jul 2024 23:19:40 +0000 Mon, 29 Jul 2024 23:16:05 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Mon, 29 Jul 2024 23:19:40 +0000 Mon, 29 Jul 2024 23:16:36 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.39.113
Hostname: ha-238496
Capacity:
cpu: 2
ephemeral-storage: 17734596Ki
hugepages-2Mi: 0
memory: 2164184Ki
pods: 110
Allocatable:
cpu: 2
ephemeral-storage: 17734596Ki
hugepages-2Mi: 0
memory: 2164184Ki
pods: 110
System Info:
Machine ID: e97d1c0e8fa74f90a32b064d1e8b3b0d
System UUID: e97d1c0e-8fa7-4f90-a32b-064d1e8b3b0d
Boot ID: 0c7e5178-e51f-4168-a6f6-c6311d7885ba
Kernel Version: 5.10.207
OS Image: Buildroot 2023.02.9
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.1.0
Kubelet Version: v1.30.3
Kube-Proxy Version: v1.30.3
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (11 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox-fc5497c4f-ftt4w 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 38s
kube-system coredns-7db6d8ff4d-p8nps 100m (5%!)(MISSING) 0 (0%!)(MISSING) 70Mi (3%!)(MISSING) 170Mi (8%!)(MISSING) 3m35s
kube-system coredns-7db6d8ff4d-tjplq 100m (5%!)(MISSING) 0 (0%!)(MISSING) 70Mi (3%!)(MISSING) 170Mi (8%!)(MISSING) 3m35s
kube-system etcd-ha-238496 100m (5%!)(MISSING) 0 (0%!)(MISSING) 100Mi (4%!)(MISSING) 0 (0%!)(MISSING) 3m48s
kube-system kindnet-55jmm 100m (5%!)(MISSING) 100m (5%!)(MISSING) 50Mi (2%!)(MISSING) 50Mi (2%!)(MISSING) 3m35s
kube-system kube-apiserver-ha-238496 250m (12%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 3m48s
kube-system kube-controller-manager-ha-238496 200m (10%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 3m48s
kube-system kube-proxy-nrvw6 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 3m35s
kube-system kube-scheduler-ha-238496 100m (5%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 3m48s
kube-system kube-vip-ha-238496 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 3m48s
kube-system storage-provisioner 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 3m35s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 950m (47%!)(MISSING) 100m (5%!)(MISSING)
memory 290Mi (13%!)(MISSING) 390Mi (18%!)(MISSING)
ephemeral-storage 0 (0%!)(MISSING) 0 (0%!)(MISSING)
hugepages-2Mi 0 (0%!)(MISSING) 0 (0%!)(MISSING)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 3m34s kube-proxy
Normal Starting 3m48s kubelet Starting kubelet.
Normal NodeHasSufficientMemory 3m48s kubelet Node ha-238496 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 3m48s kubelet Node ha-238496 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 3m48s kubelet Node ha-238496 status is now: NodeHasSufficientPID
Normal NodeAllocatableEnforced 3m48s kubelet Updated Node Allocatable limit across pods
Normal RegisteredNode 3m36s node-controller Node ha-238496 event: Registered Node ha-238496 in Controller
Normal NodeReady 3m18s kubelet Node ha-238496 status is now: NodeReady
Normal RegisteredNode 2m9s node-controller Node ha-238496 event: Registered Node ha-238496 in Controller
Normal RegisteredNode 48s node-controller Node ha-238496 event: Registered Node ha-238496 in Controller
Name: ha-238496-m02
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=ha-238496-m02
kubernetes.io/os=linux
minikube.k8s.io/commit=b13baeaf4895dcc6a8c5d0ab64a27ff86dff4ae3
minikube.k8s.io/name=ha-238496
minikube.k8s.io/primary=false
minikube.k8s.io/updated_at=2024_07_29T23_17_29_0700
minikube.k8s.io/version=v1.33.1
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Mon, 29 Jul 2024 23:17:26 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: ha-238496-m02
AcquireTime: <unset>
RenewTime: Mon, 29 Jul 2024 23:19:49 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Mon, 29 Jul 2024 23:19:29 +0000 Mon, 29 Jul 2024 23:17:26 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Mon, 29 Jul 2024 23:19:29 +0000 Mon, 29 Jul 2024 23:17:26 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Mon, 29 Jul 2024 23:19:29 +0000 Mon, 29 Jul 2024 23:17:26 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Mon, 29 Jul 2024 23:19:29 +0000 Mon, 29 Jul 2024 23:17:52 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.39.226
Hostname: ha-238496-m02
Capacity:
cpu: 2
ephemeral-storage: 17734596Ki
hugepages-2Mi: 0
memory: 2164184Ki
pods: 110
Allocatable:
cpu: 2
ephemeral-storage: 17734596Ki
hugepages-2Mi: 0
memory: 2164184Ki
pods: 110
System Info:
Machine ID: 7fc7dd8b885b4325887ebc09b43b1482
System UUID: 7fc7dd8b-885b-4325-887e-bc09b43b1482
Boot ID: f7e261b6-597f-4c06-90dc-cfd2aeb63c93
Kernel Version: 5.10.207
OS Image: Buildroot 2023.02.9
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.1.0
Kubelet Version: v1.30.3
Kube-Proxy Version: v1.30.3
PodCIDR: 10.244.1.0/24
PodCIDRs: 10.244.1.0/24
Non-terminated Pods: (8 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox-fc5497c4f-scl6h 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 38s
kube-system etcd-ha-238496-m02 100m (5%!)(MISSING) 0 (0%!)(MISSING) 100Mi (4%!)(MISSING) 0 (0%!)(MISSING) 2m26s
kube-system kindnet-xvzff 100m (5%!)(MISSING) 100m (5%!)(MISSING) 50Mi (2%!)(MISSING) 50Mi (2%!)(MISSING) 2m28s
kube-system kube-apiserver-ha-238496-m02 250m (12%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 2m25s
kube-system kube-controller-manager-ha-238496-m02 200m (10%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 2m25s
kube-system kube-proxy-m6vdn 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 2m28s
kube-system kube-scheduler-ha-238496-m02 100m (5%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 2m24s
kube-system kube-vip-ha-238496-m02 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 2m23s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (37%!)(MISSING) 100m (5%!)(MISSING)
memory 150Mi (7%!)(MISSING) 50Mi (2%!)(MISSING)
ephemeral-storage 0 (0%!)(MISSING) 0 (0%!)(MISSING)
hugepages-2Mi 0 (0%!)(MISSING) 0 (0%!)(MISSING)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 2m22s kube-proxy
Normal NodeHasSufficientMemory 2m28s (x8 over 2m28s) kubelet Node ha-238496-m02 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 2m28s (x8 over 2m28s) kubelet Node ha-238496-m02 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 2m28s (x7 over 2m28s) kubelet Node ha-238496-m02 status is now: NodeHasSufficientPID
Normal NodeAllocatableEnforced 2m28s kubelet Updated Node Allocatable limit across pods
Normal RegisteredNode 2m26s node-controller Node ha-238496-m02 event: Registered Node ha-238496-m02 in Controller
Normal RegisteredNode 2m9s node-controller Node ha-238496-m02 event: Registered Node ha-238496-m02 in Controller
Normal RegisteredNode 48s node-controller Node ha-238496-m02 event: Registered Node ha-238496-m02 in Controller
Name: ha-238496-m03
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=ha-238496-m03
kubernetes.io/os=linux
minikube.k8s.io/commit=b13baeaf4895dcc6a8c5d0ab64a27ff86dff4ae3
minikube.k8s.io/name=ha-238496
minikube.k8s.io/primary=false
minikube.k8s.io/updated_at=2024_07_29T23_18_52_0700
minikube.k8s.io/version=v1.33.1
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Mon, 29 Jul 2024 23:18:46 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: ha-238496-m03
AcquireTime: <unset>
RenewTime: Mon, 29 Jul 2024 23:19:48 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Mon, 29 Jul 2024 23:19:48 +0000 Mon, 29 Jul 2024 23:18:46 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Mon, 29 Jul 2024 23:19:48 +0000 Mon, 29 Jul 2024 23:18:46 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Mon, 29 Jul 2024 23:19:48 +0000 Mon, 29 Jul 2024 23:18:46 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Mon, 29 Jul 2024 23:19:48 +0000 Mon, 29 Jul 2024 23:19:09 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.39.149
Hostname: ha-238496-m03
Capacity:
cpu: 2
ephemeral-storage: 17734596Ki
hugepages-2Mi: 0
memory: 2164184Ki
pods: 110
Allocatable:
cpu: 2
ephemeral-storage: 17734596Ki
hugepages-2Mi: 0
memory: 2164184Ki
pods: 110
System Info:
Machine ID: 614dd3033977494daebb32e9c448932b
System UUID: 614dd303-3977-494d-aebb-32e9c448932b
Boot ID: 7e41f2a0-0988-42a6-8f82-c1af035674ba
Kernel Version: 5.10.207
OS Image: Buildroot 2023.02.9
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.1.0
Kubelet Version: v1.30.3
Kube-Proxy Version: v1.30.3
PodCIDR: 10.244.2.0/24
PodCIDRs: 10.244.2.0/24
Non-terminated Pods: (8 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox-fc5497c4f-d42qb 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 38s
kube-system etcd-ha-238496-m03 100m (5%!)(MISSING) 0 (0%!)(MISSING) 100Mi (4%!)(MISSING) 0 (0%!)(MISSING) 65s
kube-system kindnet-kb2hw 100m (5%!)(MISSING) 100m (5%!)(MISSING) 50Mi (2%!)(MISSING) 50Mi (2%!)(MISSING) 67s
kube-system kube-apiserver-ha-238496-m03 250m (12%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 66s
kube-system kube-controller-manager-ha-238496-m03 200m (10%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 66s
kube-system kube-proxy-84q2j 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 67s
kube-system kube-scheduler-ha-238496-m03 100m (5%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 65s
kube-system kube-vip-ha-238496-m03 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 62s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (37%!)(MISSING) 100m (5%!)(MISSING)
memory 150Mi (7%!)(MISSING) 50Mi (2%!)(MISSING)
ephemeral-storage 0 (0%!)(MISSING) 0 (0%!)(MISSING)
hugepages-2Mi 0 (0%!)(MISSING) 0 (0%!)(MISSING)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 62s kube-proxy
Normal NodeAllocatableEnforced 68s kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 67s (x8 over 68s) kubelet Node ha-238496-m03 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 67s (x8 over 68s) kubelet Node ha-238496-m03 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 67s (x7 over 68s) kubelet Node ha-238496-m03 status is now: NodeHasSufficientPID
Normal RegisteredNode 66s node-controller Node ha-238496-m03 event: Registered Node ha-238496-m03 in Controller
Normal RegisteredNode 64s node-controller Node ha-238496-m03 event: Registered Node ha-238496-m03 in Controller
Normal RegisteredNode 48s node-controller Node ha-238496-m03 event: Registered Node ha-238496-m03 in Controller
==> dmesg <==
[ +4.617993] NFSD: Using /var/lib/nfs/v4recovery as the NFSv4 state recovery directory
[ +0.000007] NFSD: unable to find recovery directory /var/lib/nfs/v4recovery
[ +0.000001] NFSD: Unable to initialize client recovery tracking! (-2)
[ +8.922249] systemd-fstab-generator[510]: Ignoring "noauto" option for root device
[ +0.062788] kauditd_printk_skb: 1 callbacks suppressed
[ +0.060074] systemd-fstab-generator[522]: Ignoring "noauto" option for root device
[ +2.094393] systemd-fstab-generator[769]: Ignoring "noauto" option for root device
[ +0.305877] systemd-fstab-generator[804]: Ignoring "noauto" option for root device
[ +0.124394] systemd-fstab-generator[816]: Ignoring "noauto" option for root device
[ +0.135649] systemd-fstab-generator[830]: Ignoring "noauto" option for root device
[ +2.281421] kauditd_printk_skb: 205 callbacks suppressed
[ +0.210257] systemd-fstab-generator[1045]: Ignoring "noauto" option for root device
[ +0.120089] systemd-fstab-generator[1057]: Ignoring "noauto" option for root device
[ +0.128936] systemd-fstab-generator[1069]: Ignoring "noauto" option for root device
[ +0.145315] systemd-fstab-generator[1084]: Ignoring "noauto" option for root device
[ +3.616653] systemd-fstab-generator[1187]: Ignoring "noauto" option for root device
[ +2.741843] kauditd_printk_skb: 150 callbacks suppressed
[ +0.527955] systemd-fstab-generator[1442]: Ignoring "noauto" option for root device
[ +5.026301] systemd-fstab-generator[1627]: Ignoring "noauto" option for root device
[ +0.062925] kauditd_printk_skb: 54 callbacks suppressed
[Jul29 23:16] systemd-fstab-generator[2119]: Ignoring "noauto" option for root device
[ +0.104825] kauditd_printk_skb: 81 callbacks suppressed
[ +13.770767] kauditd_printk_skb: 12 callbacks suppressed
[ +5.274468] kauditd_printk_skb: 34 callbacks suppressed
[Jul29 23:17] kauditd_printk_skb: 28 callbacks suppressed
==> etcd [189714a08644] <==
{"level":"info","ts":"2024-07-29T23:18:47.227497Z","caller":"rafthttp/pipeline.go:72","msg":"started HTTP pipelining with remote peer","local-member-id":"8069059f79d446ff","remote-peer-id":"17cb303f1eeb5f01"}
{"level":"info","ts":"2024-07-29T23:18:47.228109Z","caller":"rafthttp/peer.go:137","msg":"started remote peer","remote-peer-id":"17cb303f1eeb5f01"}
{"level":"info","ts":"2024-07-29T23:18:47.229494Z","caller":"rafthttp/transport.go:317","msg":"added remote peer","local-member-id":"8069059f79d446ff","remote-peer-id":"17cb303f1eeb5f01","remote-peer-urls":["https://192.168.39.149:2380"]}
{"level":"info","ts":"2024-07-29T23:18:47.229328Z","caller":"rafthttp/stream.go:169","msg":"started stream writer with remote peer","local-member-id":"8069059f79d446ff","remote-peer-id":"17cb303f1eeb5f01"}
{"level":"info","ts":"2024-07-29T23:18:47.229398Z","caller":"rafthttp/stream.go:169","msg":"started stream writer with remote peer","local-member-id":"8069059f79d446ff","remote-peer-id":"17cb303f1eeb5f01"}
{"level":"info","ts":"2024-07-29T23:18:47.229409Z","caller":"rafthttp/stream.go:395","msg":"started stream reader with remote peer","stream-reader-type":"stream MsgApp v2","local-member-id":"8069059f79d446ff","remote-peer-id":"17cb303f1eeb5f01"}
{"level":"info","ts":"2024-07-29T23:18:47.229443Z","caller":"rafthttp/stream.go:395","msg":"started stream reader with remote peer","stream-reader-type":"stream Message","local-member-id":"8069059f79d446ff","remote-peer-id":"17cb303f1eeb5f01"}
{"level":"warn","ts":"2024-07-29T23:18:47.326321Z","caller":"etcdhttp/peer.go:150","msg":"failed to promote a member","member-id":"17cb303f1eeb5f01","error":"etcdserver: can only promote a learner member which is in sync with leader"}
{"level":"warn","ts":"2024-07-29T23:18:48.316118Z","caller":"etcdhttp/peer.go:150","msg":"failed to promote a member","member-id":"17cb303f1eeb5f01","error":"etcdserver: can only promote a learner member which is in sync with leader"}
{"level":"warn","ts":"2024-07-29T23:18:49.316606Z","caller":"etcdhttp/peer.go:150","msg":"failed to promote a member","member-id":"17cb303f1eeb5f01","error":"etcdserver: can only promote a learner member which is in sync with leader"}
{"level":"info","ts":"2024-07-29T23:18:49.476459Z","caller":"rafthttp/peer_status.go:53","msg":"peer became active","peer-id":"17cb303f1eeb5f01"}
{"level":"info","ts":"2024-07-29T23:18:49.497913Z","caller":"rafthttp/stream.go:412","msg":"established TCP streaming connection with remote peer","stream-reader-type":"stream Message","local-member-id":"8069059f79d446ff","remote-peer-id":"17cb303f1eeb5f01"}
{"level":"info","ts":"2024-07-29T23:18:49.499017Z","caller":"rafthttp/stream.go:412","msg":"established TCP streaming connection with remote peer","stream-reader-type":"stream MsgApp v2","local-member-id":"8069059f79d446ff","remote-peer-id":"17cb303f1eeb5f01"}
{"level":"info","ts":"2024-07-29T23:18:49.51338Z","caller":"rafthttp/stream.go:249","msg":"set message encoder","from":"8069059f79d446ff","to":"17cb303f1eeb5f01","stream-type":"stream MsgApp v2"}
{"level":"info","ts":"2024-07-29T23:18:49.513524Z","caller":"rafthttp/stream.go:274","msg":"established TCP streaming connection with remote peer","stream-writer-type":"stream MsgApp v2","local-member-id":"8069059f79d446ff","remote-peer-id":"17cb303f1eeb5f01"}
{"level":"info","ts":"2024-07-29T23:18:49.51521Z","caller":"rafthttp/stream.go:249","msg":"set message encoder","from":"8069059f79d446ff","to":"17cb303f1eeb5f01","stream-type":"stream Message"}
{"level":"info","ts":"2024-07-29T23:18:49.515325Z","caller":"rafthttp/stream.go:274","msg":"established TCP streaming connection with remote peer","stream-writer-type":"stream Message","local-member-id":"8069059f79d446ff","remote-peer-id":"17cb303f1eeb5f01"}
{"level":"warn","ts":"2024-07-29T23:18:50.315769Z","caller":"etcdhttp/peer.go:150","msg":"failed to promote a member","member-id":"17cb303f1eeb5f01","error":"etcdserver: can only promote a learner member which is in sync with leader"}
{"level":"info","ts":"2024-07-29T23:18:51.318391Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"8069059f79d446ff switched to configuration voters=(1714517130804420353 4026652155228842115 9252933091911288575)"}
{"level":"info","ts":"2024-07-29T23:18:51.31855Z","caller":"membership/cluster.go:535","msg":"promote member","cluster-id":"3af003d6f0036250","local-member-id":"8069059f79d446ff"}
{"level":"info","ts":"2024-07-29T23:18:51.318607Z","caller":"etcdserver/server.go:1946","msg":"applied a configuration change through raft","local-member-id":"8069059f79d446ff","raft-conf-change":"ConfChangeAddNode","raft-conf-change-node-id":"17cb303f1eeb5f01"}
{"level":"warn","ts":"2024-07-29T23:19:09.155027Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"108.709386ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/services/endpoints/kube-system/k8s.io-minikube-hostpath\" ","response":"range_response_count:1 size:1109"}
{"level":"info","ts":"2024-07-29T23:19:09.155317Z","caller":"traceutil/trace.go:171","msg":"trace[761163028] range","detail":"{range_begin:/registry/services/endpoints/kube-system/k8s.io-minikube-hostpath; range_end:; response_count:1; response_revision:1021; }","duration":"109.088738ms","start":"2024-07-29T23:19:09.046193Z","end":"2024-07-29T23:19:09.155282Z","steps":["trace[761163028] 'agreement among raft nodes before linearized reading' (duration: 47.665659ms)","trace[761163028] 'range keys from in-memory index tree' (duration: 60.974798ms)"],"step_count":2}
{"level":"warn","ts":"2024-07-29T23:19:09.192095Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"114.48704ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/minions/ha-238496-m03\" ","response":"range_response_count:1 size:4375"}
{"level":"info","ts":"2024-07-29T23:19:09.192344Z","caller":"traceutil/trace.go:171","msg":"trace[352014540] range","detail":"{range_begin:/registry/minions/ha-238496-m03; range_end:; response_count:1; response_revision:1021; }","duration":"114.742006ms","start":"2024-07-29T23:19:09.077585Z","end":"2024-07-29T23:19:09.192327Z","steps":["trace[352014540] 'agreement among raft nodes before linearized reading' (duration: 114.137886ms)"],"step_count":1}
==> kernel <==
23:19:54 up 4 min, 0 users, load average: 0.37, 0.28, 0.12
Linux ha-238496 5.10.207 #1 SMP Tue Jul 23 04:25:44 UTC 2024 x86_64 GNU/Linux
PRETTY_NAME="Buildroot 2023.02.9"
==> kindnet [906ec8f4a759] <==
I0729 23:19:05.130611 1 main.go:299] handling current node
I0729 23:19:15.127058 1 main.go:295] Handling node with IPs: map[192.168.39.226:{}]
I0729 23:19:15.127370 1 main.go:322] Node ha-238496-m02 has CIDR [10.244.1.0/24]
I0729 23:19:15.127673 1 main.go:295] Handling node with IPs: map[192.168.39.149:{}]
I0729 23:19:15.127801 1 main.go:322] Node ha-238496-m03 has CIDR [10.244.2.0/24]
I0729 23:19:15.127989 1 main.go:295] Handling node with IPs: map[192.168.39.113:{}]
I0729 23:19:15.128033 1 main.go:299] handling current node
I0729 23:19:25.123599 1 main.go:295] Handling node with IPs: map[192.168.39.113:{}]
I0729 23:19:25.123658 1 main.go:299] handling current node
I0729 23:19:25.123678 1 main.go:295] Handling node with IPs: map[192.168.39.226:{}]
I0729 23:19:25.123684 1 main.go:322] Node ha-238496-m02 has CIDR [10.244.1.0/24]
I0729 23:19:25.123960 1 main.go:295] Handling node with IPs: map[192.168.39.149:{}]
I0729 23:19:25.123986 1 main.go:322] Node ha-238496-m03 has CIDR [10.244.2.0/24]
I0729 23:19:35.127512 1 main.go:295] Handling node with IPs: map[192.168.39.113:{}]
I0729 23:19:35.127587 1 main.go:299] handling current node
I0729 23:19:35.127606 1 main.go:295] Handling node with IPs: map[192.168.39.226:{}]
I0729 23:19:35.127612 1 main.go:322] Node ha-238496-m02 has CIDR [10.244.1.0/24]
I0729 23:19:35.128099 1 main.go:295] Handling node with IPs: map[192.168.39.149:{}]
I0729 23:19:35.128132 1 main.go:322] Node ha-238496-m03 has CIDR [10.244.2.0/24]
I0729 23:19:45.123262 1 main.go:295] Handling node with IPs: map[192.168.39.113:{}]
I0729 23:19:45.123309 1 main.go:299] handling current node
I0729 23:19:45.123325 1 main.go:295] Handling node with IPs: map[192.168.39.226:{}]
I0729 23:19:45.123331 1 main.go:322] Node ha-238496-m02 has CIDR [10.244.1.0/24]
I0729 23:19:45.123590 1 main.go:295] Handling node with IPs: map[192.168.39.149:{}]
I0729 23:19:45.123602 1 main.go:322] Node ha-238496-m03 has CIDR [10.244.2.0/24]
==> kube-apiserver [4607f65fdc74] <==
I0729 23:16:03.392746 1 controller.go:615] quota admission added evaluator for: leases.coordination.k8s.io
I0729 23:16:04.170969 1 storage_scheduling.go:95] created PriorityClass system-node-critical with value 2000001000
I0729 23:16:04.177627 1 storage_scheduling.go:95] created PriorityClass system-cluster-critical with value 2000000000
I0729 23:16:04.177698 1 storage_scheduling.go:111] all system priority classes are created successfully or already exist.
I0729 23:16:04.865424 1 controller.go:615] quota admission added evaluator for: roles.rbac.authorization.k8s.io
I0729 23:16:04.906737 1 controller.go:615] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
I0729 23:16:04.995294 1 alloc.go:330] "allocated clusterIPs" service="default/kubernetes" clusterIPs={"IPv4":"10.96.0.1"}
W0729 23:16:05.001863 1 lease.go:265] Resetting endpoints for master service "kubernetes" to [192.168.39.113]
I0729 23:16:05.003053 1 controller.go:615] quota admission added evaluator for: endpoints
I0729 23:16:05.007289 1 controller.go:615] quota admission added evaluator for: endpointslices.discovery.k8s.io
I0729 23:16:05.187097 1 controller.go:615] quota admission added evaluator for: serviceaccounts
I0729 23:16:06.220058 1 controller.go:615] quota admission added evaluator for: deployments.apps
I0729 23:16:06.244978 1 alloc.go:330] "allocated clusterIPs" service="kube-system/kube-dns" clusterIPs={"IPv4":"10.96.0.10"}
I0729 23:16:06.262209 1 controller.go:615] quota admission added evaluator for: daemonsets.apps
I0729 23:16:19.197134 1 controller.go:615] quota admission added evaluator for: controllerrevisions.apps
I0729 23:16:19.306858 1 controller.go:615] quota admission added evaluator for: replicasets.apps
E0729 23:19:51.666832 1 conn.go:339] Error on socket receive: read tcp 192.168.39.254:8443->192.168.39.1:46784: use of closed network connection
E0729 23:19:51.900211 1 conn.go:339] Error on socket receive: read tcp 192.168.39.254:8443->192.168.39.1:46800: use of closed network connection
E0729 23:19:52.089837 1 conn.go:339] Error on socket receive: read tcp 192.168.39.254:8443->192.168.39.1:46812: use of closed network connection
E0729 23:19:52.401910 1 conn.go:339] Error on socket receive: read tcp 192.168.39.254:8443->192.168.39.1:46848: use of closed network connection
E0729 23:19:52.595609 1 conn.go:339] Error on socket receive: read tcp 192.168.39.254:8443->192.168.39.1:46858: use of closed network connection
E0729 23:19:52.792814 1 conn.go:339] Error on socket receive: read tcp 192.168.39.254:8443->192.168.39.1:46886: use of closed network connection
E0729 23:19:53.082395 1 conn.go:339] Error on socket receive: read tcp 192.168.39.254:8443->192.168.39.1:46916: use of closed network connection
E0729 23:19:53.260351 1 conn.go:339] Error on socket receive: read tcp 192.168.39.254:8443->192.168.39.1:46932: use of closed network connection
E0729 23:19:53.462817 1 conn.go:339] Error on socket receive: read tcp 192.168.39.254:8443->192.168.39.1:46954: use of closed network connection
==> kube-controller-manager [2127946fcd8b] <==
I0729 23:19:16.735098 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="161.7027ms"
I0729 23:19:16.948440 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="212.939882ms"
I0729 23:19:16.995409 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="46.910529ms"
I0729 23:19:17.038535 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="43.06149ms"
I0729 23:19:17.039551 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="937.395µs"
I0729 23:19:17.067733 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="13.626198ms"
I0729 23:19:17.069562 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="76.586µs"
I0729 23:19:18.260665 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="69.796µs"
I0729 23:19:18.282572 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="76.075µs"
I0729 23:19:18.288809 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="69.98µs"
I0729 23:19:18.314246 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="726.325µs"
I0729 23:19:18.324551 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="80.065µs"
I0729 23:19:18.329486 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="74.04µs"
I0729 23:19:18.518217 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="106.659µs"
I0729 23:19:18.544739 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="48.531µs"
I0729 23:19:19.827540 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="34.691285ms"
I0729 23:19:19.827958 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="350.714µs"
I0729 23:19:20.641120 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="43.588126ms"
I0729 23:19:20.641419 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="160.847µs"
I0729 23:19:22.391760 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="49.170401ms"
I0729 23:19:22.391880 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="65.585µs"
I0729 23:19:50.847022 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="72.998µs"
I0729 23:19:51.883422 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="86.014µs"
I0729 23:19:51.902395 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="51.001µs"
I0729 23:19:51.913198 1 replica_set.go:676] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/busybox-fc5497c4f" duration="63.514µs"
==> kube-proxy [2bb8727b4690] <==
I0729 23:16:20.042835 1 server_linux.go:69] "Using iptables proxy"
I0729 23:16:20.054809 1 server.go:1062] "Successfully retrieved node IP(s)" IPs=["192.168.39.113"]
I0729 23:16:20.102319 1 server_linux.go:143] "No iptables support for family" ipFamily="IPv6"
I0729 23:16:20.102375 1 server.go:661] "kube-proxy running in single-stack mode" ipFamily="IPv4"
I0729 23:16:20.102397 1 server_linux.go:165] "Using iptables Proxier"
I0729 23:16:20.105809 1 proxier.go:243] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses"
I0729 23:16:20.106413 1 server.go:872] "Version info" version="v1.30.3"
I0729 23:16:20.106450 1 server.go:874] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0729 23:16:20.108364 1 config.go:192] "Starting service config controller"
I0729 23:16:20.108406 1 shared_informer.go:313] Waiting for caches to sync for service config
I0729 23:16:20.108456 1 config.go:101] "Starting endpoint slice config controller"
I0729 23:16:20.108462 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0729 23:16:20.109671 1 config.go:319] "Starting node config controller"
I0729 23:16:20.109718 1 shared_informer.go:313] Waiting for caches to sync for node config
I0729 23:16:20.209255 1 shared_informer.go:320] Caches are synced for endpoint slice config
I0729 23:16:20.209265 1 shared_informer.go:320] Caches are synced for service config
I0729 23:16:20.209811 1 shared_informer.go:320] Caches are synced for node config
==> kube-scheduler [dc0f824c8c08] <==
E0729 23:16:04.323274 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
W0729 23:16:04.326788 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0729 23:16:04.326980 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
W0729 23:16:04.383451 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0729 23:16:04.383717 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
W0729 23:16:04.534628 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0729 23:16:04.535476 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
W0729 23:16:04.556376 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0729 23:16:04.556427 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
W0729 23:16:04.573697 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0729 23:16:04.573750 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
W0729 23:16:04.653266 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E0729 23:16:04.653315 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
W0729 23:16:04.740804 1 reflector.go:547] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0729 23:16:04.740888 1 reflector.go:150] runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
I0729 23:16:06.457680 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
E0729 23:18:47.120262 1 framework.go:1286] "Plugin Failed" err="Operation cannot be fulfilled on pods/binding \"kindnet-kb2hw\": pod kindnet-kb2hw is already assigned to node \"ha-238496-m03\"" plugin="DefaultBinder" pod="kube-system/kindnet-kb2hw" node="ha-238496-m03"
E0729 23:18:47.122790 1 schedule_one.go:338] "scheduler cache ForgetPod failed" err="pod ef875a41-530f-48ba-b034-d08a8a7acbbc(kube-system/kindnet-kb2hw) wasn't assumed so cannot be forgotten" pod="kube-system/kindnet-kb2hw"
E0729 23:18:47.122887 1 schedule_one.go:1046] "Error scheduling pod; retrying" err="running Bind plugin \"DefaultBinder\": Operation cannot be fulfilled on pods/binding \"kindnet-kb2hw\": pod kindnet-kb2hw is already assigned to node \"ha-238496-m03\"" pod="kube-system/kindnet-kb2hw"
I0729 23:18:47.123210 1 schedule_one.go:1059] "Pod has been assigned to node. Abort adding it back to queue." pod="kube-system/kindnet-kb2hw" node="ha-238496-m03"
I0729 23:19:16.509862 1 cache.go:503] "Pod was added to a different node than it was assumed" podKey="776d65d2-03d8-4edb-9781-6d7d4967e364" pod="default/busybox-fc5497c4f-8ql68" assumedNode="ha-238496-m03" currentNode="ha-238496-m02"
E0729 23:19:16.512667 1 framework.go:1286] "Plugin Failed" err="Operation cannot be fulfilled on pods/binding \"busybox-fc5497c4f-8ql68\": pod busybox-fc5497c4f-8ql68 is already assigned to node \"ha-238496-m03\"" plugin="DefaultBinder" pod="default/busybox-fc5497c4f-8ql68" node="ha-238496-m02"
E0729 23:19:16.512731 1 schedule_one.go:338] "scheduler cache ForgetPod failed" err="pod 776d65d2-03d8-4edb-9781-6d7d4967e364(default/busybox-fc5497c4f-8ql68) was assumed on ha-238496-m02 but assigned to ha-238496-m03" pod="default/busybox-fc5497c4f-8ql68"
E0729 23:19:16.512750 1 schedule_one.go:1046] "Error scheduling pod; retrying" err="running Bind plugin \"DefaultBinder\": Operation cannot be fulfilled on pods/binding \"busybox-fc5497c4f-8ql68\": pod busybox-fc5497c4f-8ql68 is already assigned to node \"ha-238496-m03\"" pod="default/busybox-fc5497c4f-8ql68"
I0729 23:19:16.512778 1 schedule_one.go:1059] "Pod has been assigned to node. Abort adding it back to queue." pod="default/busybox-fc5497c4f-8ql68" node="ha-238496-m03"
==> kubelet <==
Jul 29 23:19:16 ha-238496 kubelet[2126]: I0729 23:19:16.604936 2126 topology_manager.go:215] "Topology Admit Handler" podUID="f100377c-7ba9-4012-a20e-ac31710cdc43" podNamespace="default" podName="busybox-fc5497c4f-2rl8g"
Jul 29 23:19:16 ha-238496 kubelet[2126]: E0729 23:19:16.665071 2126 pod_workers.go:1298] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-sdvlm], unattached volumes=[], failed to process volumes=[]: context canceled" pod="default/busybox-fc5497c4f-2rl8g" podUID="f100377c-7ba9-4012-a20e-ac31710cdc43"
Jul 29 23:19:16 ha-238496 kubelet[2126]: I0729 23:19:16.696959 2126 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdvlm\" (UniqueName: \"kubernetes.io/projected/f100377c-7ba9-4012-a20e-ac31710cdc43-kube-api-access-sdvlm\") pod \"busybox-fc5497c4f-2rl8g\" (UID: \"f100377c-7ba9-4012-a20e-ac31710cdc43\") " pod="default/busybox-fc5497c4f-2rl8g"
Jul 29 23:19:16 ha-238496 kubelet[2126]: I0729 23:19:16.714770 2126 status_manager.go:877] "Failed to update status for pod" pod="default/busybox-fc5497c4f-2rl8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f100377c-7ba9-4012-a20e-ac31710cdc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2024-07-29T23:19:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2024-07-29T23:19:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2024-07-29T23:19:16Z\\\",\\\"message\\\":\\\"containers with unready status: [busybox]\\\",\\\"reason\\\":\\\"Cont
ainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2024-07-29T23:19:16Z\\\",\\\"message\\\":\\\"containers with unready status: [busybox]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"gcr.io/k8s-minikube/busybox:1.28\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"busybox\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}}}],\\\"hostIP\\\":\\\"192.168.39.113\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.39.113\\\"}],\\\"startTime\\\":\\\"2024-07-29T23:19:16Z\\\"}}\" for pod \"default\"/\"busybox-fc5497c4f-2rl8g\": pods \"busybox-fc5497c4f-2rl8g\" not found"
Jul 29 23:19:16 ha-238496 kubelet[2126]: I0729 23:19:16.732196 2126 topology_manager.go:215] "Topology Admit Handler" podUID="22830bb4-9b86-42f8-8354-48e4b8d2f29b" podNamespace="default" podName="busybox-fc5497c4f-df7wx"
Jul 29 23:19:16 ha-238496 kubelet[2126]: I0729 23:19:16.743186 2126 topology_manager.go:215] "Topology Admit Handler" podUID="8ad271f0-7e72-47c4-86ce-f549f1e2fc64" podNamespace="default" podName="busybox-fc5497c4f-6d7mp"
Jul 29 23:19:16 ha-238496 kubelet[2126]: I0729 23:19:16.798376 2126 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr2c8\" (UniqueName: \"kubernetes.io/projected/22830bb4-9b86-42f8-8354-48e4b8d2f29b-kube-api-access-kr2c8\") pod \"busybox-fc5497c4f-df7wx\" (UID: \"22830bb4-9b86-42f8-8354-48e4b8d2f29b\") " pod="default/busybox-fc5497c4f-df7wx"
Jul 29 23:19:16 ha-238496 kubelet[2126]: I0729 23:19:16.798673 2126 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4l6x\" (UniqueName: \"kubernetes.io/projected/8ad271f0-7e72-47c4-86ce-f549f1e2fc64-kube-api-access-h4l6x\") pod \"busybox-fc5497c4f-6d7mp\" (UID: \"8ad271f0-7e72-47c4-86ce-f549f1e2fc64\") " pod="default/busybox-fc5497c4f-6d7mp"
Jul 29 23:19:16 ha-238496 kubelet[2126]: E0729 23:19:16.809833 2126 pod_workers.go:1298] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-kr2c8], unattached volumes=[], failed to process volumes=[]: context canceled" pod="default/busybox-fc5497c4f-df7wx" podUID="22830bb4-9b86-42f8-8354-48e4b8d2f29b"
Jul 29 23:19:16 ha-238496 kubelet[2126]: E0729 23:19:16.810972 2126 pod_workers.go:1298] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-h4l6x], unattached volumes=[], failed to process volumes=[]: context canceled" pod="default/busybox-fc5497c4f-6d7mp" podUID="8ad271f0-7e72-47c4-86ce-f549f1e2fc64"
Jul 29 23:19:16 ha-238496 kubelet[2126]: E0729 23:19:16.813121 2126 projected.go:200] Error preparing data for projected volume kube-api-access-sdvlm for pod default/busybox-fc5497c4f-2rl8g: failed to fetch token: pod "busybox-fc5497c4f-2rl8g" not found
Jul 29 23:19:16 ha-238496 kubelet[2126]: E0729 23:19:16.813755 2126 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f100377c-7ba9-4012-a20e-ac31710cdc43-kube-api-access-sdvlm podName:f100377c-7ba9-4012-a20e-ac31710cdc43 nodeName:}" failed. No retries permitted until 2024-07-29 23:19:17.313457402 +0000 UTC m=+191.294774399 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-sdvlm" (UniqueName: "kubernetes.io/projected/f100377c-7ba9-4012-a20e-ac31710cdc43-kube-api-access-sdvlm") pod "busybox-fc5497c4f-2rl8g" (UID: "f100377c-7ba9-4012-a20e-ac31710cdc43") : failed to fetch token: pod "busybox-fc5497c4f-2rl8g" not found
Jul 29 23:19:17 ha-238496 kubelet[2126]: I0729 23:19:17.303331 2126 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr2c8\" (UniqueName: \"kubernetes.io/projected/22830bb4-9b86-42f8-8354-48e4b8d2f29b-kube-api-access-kr2c8\") pod \"22830bb4-9b86-42f8-8354-48e4b8d2f29b\" (UID: \"22830bb4-9b86-42f8-8354-48e4b8d2f29b\") "
Jul 29 23:19:17 ha-238496 kubelet[2126]: I0729 23:19:17.303434 2126 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4l6x\" (UniqueName: \"kubernetes.io/projected/8ad271f0-7e72-47c4-86ce-f549f1e2fc64-kube-api-access-h4l6x\") pod \"8ad271f0-7e72-47c4-86ce-f549f1e2fc64\" (UID: \"8ad271f0-7e72-47c4-86ce-f549f1e2fc64\") "
Jul 29 23:19:17 ha-238496 kubelet[2126]: I0729 23:19:17.303616 2126 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-sdvlm\" (UniqueName: \"kubernetes.io/projected/f100377c-7ba9-4012-a20e-ac31710cdc43-kube-api-access-sdvlm\") on node \"ha-238496\" DevicePath \"\""
Jul 29 23:19:17 ha-238496 kubelet[2126]: I0729 23:19:17.308363 2126 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad271f0-7e72-47c4-86ce-f549f1e2fc64-kube-api-access-h4l6x" (OuterVolumeSpecName: "kube-api-access-h4l6x") pod "8ad271f0-7e72-47c4-86ce-f549f1e2fc64" (UID: "8ad271f0-7e72-47c4-86ce-f549f1e2fc64"). InnerVolumeSpecName "kube-api-access-h4l6x". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jul 29 23:19:17 ha-238496 kubelet[2126]: I0729 23:19:17.313036 2126 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22830bb4-9b86-42f8-8354-48e4b8d2f29b-kube-api-access-kr2c8" (OuterVolumeSpecName: "kube-api-access-kr2c8") pod "22830bb4-9b86-42f8-8354-48e4b8d2f29b" (UID: "22830bb4-9b86-42f8-8354-48e4b8d2f29b"). InnerVolumeSpecName "kube-api-access-kr2c8". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jul 29 23:19:17 ha-238496 kubelet[2126]: I0729 23:19:17.404689 2126 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-kr2c8\" (UniqueName: \"kubernetes.io/projected/22830bb4-9b86-42f8-8354-48e4b8d2f29b-kube-api-access-kr2c8\") on node \"ha-238496\" DevicePath \"\""
Jul 29 23:19:17 ha-238496 kubelet[2126]: I0729 23:19:17.404774 2126 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-h4l6x\" (UniqueName: \"kubernetes.io/projected/8ad271f0-7e72-47c4-86ce-f549f1e2fc64-kube-api-access-h4l6x\") on node \"ha-238496\" DevicePath \"\""
Jul 29 23:19:18 ha-238496 kubelet[2126]: I0729 23:19:18.193836 2126 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f100377c-7ba9-4012-a20e-ac31710cdc43" path="/var/lib/kubelet/pods/f100377c-7ba9-4012-a20e-ac31710cdc43/volumes"
Jul 29 23:19:18 ha-238496 kubelet[2126]: I0729 23:19:18.513992 2126 topology_manager.go:215] "Topology Admit Handler" podUID="1af09aea-0782-4e55-a8d0-349bcfa014a2" podNamespace="default" podName="busybox-fc5497c4f-ftt4w"
Jul 29 23:19:18 ha-238496 kubelet[2126]: I0729 23:19:18.613129 2126 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bssld\" (UniqueName: \"kubernetes.io/projected/1af09aea-0782-4e55-a8d0-349bcfa014a2-kube-api-access-bssld\") pod \"busybox-fc5497c4f-ftt4w\" (UID: \"1af09aea-0782-4e55-a8d0-349bcfa014a2\") " pod="default/busybox-fc5497c4f-ftt4w"
Jul 29 23:19:20 ha-238496 kubelet[2126]: I0729 23:19:20.184289 2126 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22830bb4-9b86-42f8-8354-48e4b8d2f29b" path="/var/lib/kubelet/pods/22830bb4-9b86-42f8-8354-48e4b8d2f29b/volumes"
Jul 29 23:19:20 ha-238496 kubelet[2126]: I0729 23:19:20.184878 2126 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ad271f0-7e72-47c4-86ce-f549f1e2fc64" path="/var/lib/kubelet/pods/8ad271f0-7e72-47c4-86ce-f549f1e2fc64/volumes"
Jul 29 23:19:22 ha-238496 kubelet[2126]: I0729 23:19:22.342487 2126 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="default/busybox-fc5497c4f-ftt4w" podStartSLOduration=4.260317486 podStartE2EDuration="6.34243599s" podCreationTimestamp="2024-07-29 23:19:16 +0000 UTC" firstStartedPulling="2024-07-29 23:19:19.188446274 +0000 UTC m=+193.169763285" lastFinishedPulling="2024-07-29 23:19:21.27056479 +0000 UTC m=+195.251881789" observedRunningTime="2024-07-29 23:19:22.342022908 +0000 UTC m=+196.323339926" watchObservedRunningTime="2024-07-29 23:19:22.34243599 +0000 UTC m=+196.323753012"
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p ha-238496 -n ha-238496
helpers_test.go:261: (dbg) Run: kubectl --context ha-238496 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:285: <<< TestMultiControlPlane/serial/DeployApp FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestMultiControlPlane/serial/DeployApp (39.18s)