=== RUN TestScheduledStopUnix
scheduled_stop_test.go:128: (dbg) Run: out/minikube-linux-arm64 start -p scheduled-stop-721021 --memory=3072 --driver=docker --container-runtime=containerd
E1017 19:31:49.539825 295207 cert_rotation.go:172] "Loading client cert failed" err="open /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/functional-611928/client.crt: no such file or directory" logger="tls-transport-cache.UnhandledError" key="key"
scheduled_stop_test.go:128: (dbg) Done: out/minikube-linux-arm64 start -p scheduled-stop-721021 --memory=3072 --driver=docker --container-runtime=containerd: (36.687309154s)
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-721021 --schedule 5m
scheduled_stop_test.go:191: (dbg) Run: out/minikube-linux-arm64 status --format={{.TimeToStop}} -p scheduled-stop-721021 -n scheduled-stop-721021
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-721021 --schedule 15s
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:98: process 444542 running but should have been killed on reschedule of stop
panic.go:636: *** TestScheduledStopUnix FAILED at 2025-10-17 19:32:18.6729617 +0000 UTC m=+2141.541315876
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:223: ======> post-mortem[TestScheduledStopUnix]: network settings <======
helpers_test.go:230: HOST ENV snapshots: PROXY env: HTTP_PROXY="<empty>" HTTPS_PROXY="<empty>" NO_PROXY="<empty>"
helpers_test.go:238: ======> post-mortem[TestScheduledStopUnix]: docker inspect <======
helpers_test.go:239: (dbg) Run: docker inspect scheduled-stop-721021
helpers_test.go:243: (dbg) docker inspect scheduled-stop-721021:
-- stdout --
[
{
"Id": "0740d168a83103eaed8950edf42298db90bb6b7efdd7b50e5817c39ead248dc0",
"Created": "2025-10-17T19:31:46.886236504Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 442550,
"ExitCode": 0,
"Error": "",
"StartedAt": "2025-10-17T19:31:46.954491549Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:551264e61976f283a9fbfb2241e8ff3a6dda7ce0fb240891319c40d01d82fdd7",
"ResolvConfPath": "/var/lib/docker/containers/0740d168a83103eaed8950edf42298db90bb6b7efdd7b50e5817c39ead248dc0/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/0740d168a83103eaed8950edf42298db90bb6b7efdd7b50e5817c39ead248dc0/hostname",
"HostsPath": "/var/lib/docker/containers/0740d168a83103eaed8950edf42298db90bb6b7efdd7b50e5817c39ead248dc0/hosts",
"LogPath": "/var/lib/docker/containers/0740d168a83103eaed8950edf42298db90bb6b7efdd7b50e5817c39ead248dc0/0740d168a83103eaed8950edf42298db90bb6b7efdd7b50e5817c39ead248dc0-json.log",
"Name": "/scheduled-stop-721021",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"scheduled-stop-721021:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {}
},
"NetworkMode": "scheduled-stop-721021",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 3221225472,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 6442450944,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"ID": "0740d168a83103eaed8950edf42298db90bb6b7efdd7b50e5817c39ead248dc0",
"LowerDir": "/var/lib/docker/overlay2/0077f54c9b1feaac87de1221e7bc4aee0eb66bc5acce309f1271119857c6a922-init/diff:/var/lib/docker/overlay2/f9034bf2c594fd4cb4a3940590c585e097abafd8979abf7d3b575170bb4a0dae/diff",
"MergedDir": "/var/lib/docker/overlay2/0077f54c9b1feaac87de1221e7bc4aee0eb66bc5acce309f1271119857c6a922/merged",
"UpperDir": "/var/lib/docker/overlay2/0077f54c9b1feaac87de1221e7bc4aee0eb66bc5acce309f1271119857c6a922/diff",
"WorkDir": "/var/lib/docker/overlay2/0077f54c9b1feaac87de1221e7bc4aee0eb66bc5acce309f1271119857c6a922/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "scheduled-stop-721021",
"Source": "/var/lib/docker/volumes/scheduled-stop-721021/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "scheduled-stop-721021",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760609789-21757@sha256:9824b20f4774128fcb298ad0e6cac7649729886cfba9d444b2305c743a5044c6",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "scheduled-stop-721021",
"name.minikube.sigs.k8s.io": "scheduled-stop-721021",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "9c1d4272bdfd4430c48cec2fa3995f802308d0667b9d50b95dfca3cbd351a55c",
"SandboxKey": "/var/run/docker/netns/9c1d4272bdfd",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33339"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33340"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33343"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33341"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33342"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"scheduled-stop-721021": {
"IPAMConfig": {
"IPv4Address": "192.168.76.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "42:76:9a:03:e7:c8",
"DriverOpts": null,
"GwPriority": 0,
"NetworkID": "68d2b9bd0d71375a95aafeadef98559b99d0cddd600c94accb5a2fd04ea91d8b",
"EndpointID": "0e0adb603b68c58cd7b4dc0453b8dd44a853bb30d65c39f8aceba7b411b9c6b6",
"Gateway": "192.168.76.1",
"IPAddress": "192.168.76.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"scheduled-stop-721021",
"0740d168a831"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:247: (dbg) Run: out/minikube-linux-arm64 status --format={{.Host}} -p scheduled-stop-721021 -n scheduled-stop-721021
helpers_test.go:252: <<< TestScheduledStopUnix FAILED: start of post-mortem logs <<<
helpers_test.go:253: ======> post-mortem[TestScheduledStopUnix]: minikube logs <======
helpers_test.go:255: (dbg) Run: out/minikube-linux-arm64 -p scheduled-stop-721021 logs -n 25
helpers_test.go:255: (dbg) Done: out/minikube-linux-arm64 -p scheduled-stop-721021 logs -n 25: (1.143468456s)
helpers_test.go:260: TestScheduledStopUnix logs:
-- stdout --
==> Audit <==
┌─────────┬─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┬───────────────────────┬─────────┬─────────┬─────────────────────┬─────────────────────┐
│ COMMAND │ ARGS │ PROFILE │ USER │ VERSION │ START TIME │ END TIME │
├─────────┼─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┼───────────────────────┼─────────┼─────────┼─────────────────────┼─────────────────────┤
│ stop │ -p multinode-327259 │ multinode-327259 │ jenkins │ v1.37.0 │ 17 Oct 25 19:26 UTC │ 17 Oct 25 19:26 UTC │
│ start │ -p multinode-327259 --wait=true -v=5 --alsologtostderr │ multinode-327259 │ jenkins │ v1.37.0 │ 17 Oct 25 19:26 UTC │ 17 Oct 25 19:27 UTC │
│ node │ list -p multinode-327259 │ multinode-327259 │ jenkins │ v1.37.0 │ 17 Oct 25 19:27 UTC │ │
│ node │ multinode-327259 node delete m03 │ multinode-327259 │ jenkins │ v1.37.0 │ 17 Oct 25 19:27 UTC │ 17 Oct 25 19:27 UTC │
│ stop │ multinode-327259 stop │ multinode-327259 │ jenkins │ v1.37.0 │ 17 Oct 25 19:27 UTC │ 17 Oct 25 19:27 UTC │
│ start │ -p multinode-327259 --wait=true -v=5 --alsologtostderr --driver=docker --container-runtime=containerd │ multinode-327259 │ jenkins │ v1.37.0 │ 17 Oct 25 19:27 UTC │ 17 Oct 25 19:28 UTC │
│ node │ list -p multinode-327259 │ multinode-327259 │ jenkins │ v1.37.0 │ 17 Oct 25 19:28 UTC │ │
│ start │ -p multinode-327259-m02 --driver=docker --container-runtime=containerd │ multinode-327259-m02 │ jenkins │ v1.37.0 │ 17 Oct 25 19:28 UTC │ │
│ start │ -p multinode-327259-m03 --driver=docker --container-runtime=containerd │ multinode-327259-m03 │ jenkins │ v1.37.0 │ 17 Oct 25 19:28 UTC │ 17 Oct 25 19:29 UTC │
│ node │ add -p multinode-327259 │ multinode-327259 │ jenkins │ v1.37.0 │ 17 Oct 25 19:29 UTC │ │
│ delete │ -p multinode-327259-m03 │ multinode-327259-m03 │ jenkins │ v1.37.0 │ 17 Oct 25 19:29 UTC │ 17 Oct 25 19:29 UTC │
│ delete │ -p multinode-327259 │ multinode-327259 │ jenkins │ v1.37.0 │ 17 Oct 25 19:29 UTC │ 17 Oct 25 19:29 UTC │
│ start │ -p test-preload-035366 --memory=3072 --alsologtostderr --wait=true --preload=false --driver=docker --container-runtime=containerd --kubernetes-version=v1.32.0 │ test-preload-035366 │ jenkins │ v1.37.0 │ 17 Oct 25 19:29 UTC │ 17 Oct 25 19:30 UTC │
│ image │ test-preload-035366 image pull gcr.io/k8s-minikube/busybox │ test-preload-035366 │ jenkins │ v1.37.0 │ 17 Oct 25 19:30 UTC │ 17 Oct 25 19:30 UTC │
│ stop │ -p test-preload-035366 │ test-preload-035366 │ jenkins │ v1.37.0 │ 17 Oct 25 19:30 UTC │ 17 Oct 25 19:30 UTC │
│ start │ -p test-preload-035366 --memory=3072 --alsologtostderr -v=1 --wait=true --driver=docker --container-runtime=containerd │ test-preload-035366 │ jenkins │ v1.37.0 │ 17 Oct 25 19:30 UTC │ 17 Oct 25 19:31 UTC │
│ image │ test-preload-035366 image list │ test-preload-035366 │ jenkins │ v1.37.0 │ 17 Oct 25 19:31 UTC │ 17 Oct 25 19:31 UTC │
│ delete │ -p test-preload-035366 │ test-preload-035366 │ jenkins │ v1.37.0 │ 17 Oct 25 19:31 UTC │ 17 Oct 25 19:31 UTC │
│ start │ -p scheduled-stop-721021 --memory=3072 --driver=docker --container-runtime=containerd │ scheduled-stop-721021 │ jenkins │ v1.37.0 │ 17 Oct 25 19:31 UTC │ 17 Oct 25 19:32 UTC │
│ stop │ -p scheduled-stop-721021 --schedule 5m │ scheduled-stop-721021 │ jenkins │ v1.37.0 │ 17 Oct 25 19:32 UTC │ │
│ stop │ -p scheduled-stop-721021 --schedule 5m │ scheduled-stop-721021 │ jenkins │ v1.37.0 │ 17 Oct 25 19:32 UTC │ │
│ stop │ -p scheduled-stop-721021 --schedule 5m │ scheduled-stop-721021 │ jenkins │ v1.37.0 │ 17 Oct 25 19:32 UTC │ │
│ stop │ -p scheduled-stop-721021 --schedule 15s │ scheduled-stop-721021 │ jenkins │ v1.37.0 │ 17 Oct 25 19:32 UTC │ │
│ stop │ -p scheduled-stop-721021 --schedule 15s │ scheduled-stop-721021 │ jenkins │ v1.37.0 │ 17 Oct 25 19:32 UTC │ │
│ stop │ -p scheduled-stop-721021 --schedule 15s │ scheduled-stop-721021 │ jenkins │ v1.37.0 │ 17 Oct 25 19:32 UTC │ │
└─────────┴─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┴───────────────────────┴─────────┴─────────┴─────────────────────┴─────────────────────┘
==> Last Start <==
Log file created at: 2025/10/17 19:31:41
Running on machine: ip-172-31-31-251
Binary: Built with gc go1.24.6 for linux/arm64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I1017 19:31:41.481369 442158 out.go:360] Setting OutFile to fd 1 ...
I1017 19:31:41.481492 442158 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1017 19:31:41.481496 442158 out.go:374] Setting ErrFile to fd 2...
I1017 19:31:41.481500 442158 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1017 19:31:41.481758 442158 root.go:338] Updating PATH: /home/jenkins/minikube-integration/21753-293342/.minikube/bin
I1017 19:31:41.482161 442158 out.go:368] Setting JSON to false
I1017 19:31:41.482976 442158 start.go:131] hostinfo: {"hostname":"ip-172-31-31-251","uptime":8047,"bootTime":1760721454,"procs":153,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1084-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"982e3628-3742-4b3e-bb63-ac1b07660ec7"}
I1017 19:31:41.483033 442158 start.go:141] virtualization:
I1017 19:31:41.489132 442158 out.go:179] * [scheduled-stop-721021] minikube v1.37.0 on Ubuntu 20.04 (arm64)
I1017 19:31:41.492616 442158 out.go:179] - MINIKUBE_LOCATION=21753
I1017 19:31:41.492670 442158 notify.go:220] Checking for updates...
I1017 19:31:41.499153 442158 out.go:179] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I1017 19:31:41.502468 442158 out.go:179] - KUBECONFIG=/home/jenkins/minikube-integration/21753-293342/kubeconfig
I1017 19:31:41.505555 442158 out.go:179] - MINIKUBE_HOME=/home/jenkins/minikube-integration/21753-293342/.minikube
I1017 19:31:41.508751 442158 out.go:179] - MINIKUBE_BIN=out/minikube-linux-arm64
I1017 19:31:41.511914 442158 out.go:179] - MINIKUBE_FORCE_SYSTEMD=
I1017 19:31:41.515359 442158 driver.go:421] Setting default libvirt URI to qemu:///system
I1017 19:31:41.544580 442158 docker.go:123] docker version: linux-28.1.1:Docker Engine - Community
I1017 19:31:41.544745 442158 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1017 19:31:41.608763 442158 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:23 OomKillDisable:true NGoroutines:42 SystemTime:2025-10-17 19:31:41.599070679 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1084-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214839296 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:28.1.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:05044ec0a9a75232cad458027ca83437aae3f4da Expected:} RuncCommit:{ID:v1.2.5-0-g59923ef Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx P
ath:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.23.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.35.1]] Warnings:<nil>}}
I1017 19:31:41.608886 442158 docker.go:318] overlay module found
I1017 19:31:41.612230 442158 out.go:179] * Using the docker driver based on user configuration
I1017 19:31:41.615254 442158 start.go:305] selected driver: docker
I1017 19:31:41.615271 442158 start.go:925] validating driver "docker" against <nil>
I1017 19:31:41.615283 442158 start.go:936] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I1017 19:31:41.616045 442158 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1017 19:31:41.679582 442158 info.go:266] docker info: {ID:EOU5:DNGX:XN6V:L2FZ:UXRM:5TWK:EVUR:KC2F:GT7Z:Y4O4:GB77:5PD3 Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:23 OomKillDisable:true NGoroutines:42 SystemTime:2025-10-17 19:31:41.669457173 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1084-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214839296 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-31-251 Labels:[] ExperimentalBuild:false ServerVersion:28.1.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:05044ec0a9a75232cad458027ca83437aae3f4da Expected:} RuncCommit:{ID:v1.2.5-0-g59923ef Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx P
ath:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.23.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.35.1]] Warnings:<nil>}}
I1017 19:31:41.679725 442158 start_flags.go:327] no existing cluster config was found, will generate one from the flags
I1017 19:31:41.679979 442158 start_flags.go:974] Wait components to verify : map[apiserver:true system_pods:true]
I1017 19:31:41.683067 442158 out.go:179] * Using Docker driver with root privileges
I1017 19:31:41.685969 442158 cni.go:84] Creating CNI manager for ""
I1017 19:31:41.686032 442158 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I1017 19:31:41.686039 442158 start_flags.go:336] Found "CNI" CNI - setting NetworkPlugin=cni
I1017 19:31:41.686118 442158 start.go:349] cluster config:
{Name:scheduled-stop-721021 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760609789-21757@sha256:9824b20f4774128fcb298ad0e6cac7649729886cfba9d444b2305c743a5044c6 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-721021 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local
ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I1017 19:31:41.691224 442158 out.go:179] * Starting "scheduled-stop-721021" primary control-plane node in "scheduled-stop-721021" cluster
I1017 19:31:41.694205 442158 cache.go:123] Beginning downloading kic base image for docker with containerd
I1017 19:31:41.697336 442158 out.go:179] * Pulling base image v0.0.48-1760609789-21757 ...
I1017 19:31:41.700217 442158 preload.go:183] Checking if preload exists for k8s version v1.34.1 and runtime containerd
I1017 19:31:41.700267 442158 preload.go:198] Found local preload: /home/jenkins/minikube-integration/21753-293342/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-containerd-overlay2-arm64.tar.lz4
I1017 19:31:41.700294 442158 cache.go:58] Caching tarball of preloaded images
I1017 19:31:41.700298 442158 image.go:81] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760609789-21757@sha256:9824b20f4774128fcb298ad0e6cac7649729886cfba9d444b2305c743a5044c6 in local docker daemon
I1017 19:31:41.700378 442158 preload.go:233] Found /home/jenkins/minikube-integration/21753-293342/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-containerd-overlay2-arm64.tar.lz4 in cache, skipping download
I1017 19:31:41.700386 442158 cache.go:61] Finished verifying existence of preloaded tar for v1.34.1 on containerd
I1017 19:31:41.700714 442158 profile.go:143] Saving config to /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/config.json ...
I1017 19:31:41.700732 442158 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/config.json: {Name:mkf2ce33ff84a4b7e13503793fc67c4553458844 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1017 19:31:41.718588 442158 image.go:100] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760609789-21757@sha256:9824b20f4774128fcb298ad0e6cac7649729886cfba9d444b2305c743a5044c6 in local docker daemon, skipping pull
I1017 19:31:41.718604 442158 cache.go:147] gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760609789-21757@sha256:9824b20f4774128fcb298ad0e6cac7649729886cfba9d444b2305c743a5044c6 exists in daemon, skipping load
I1017 19:31:41.718624 442158 cache.go:232] Successfully downloaded all kic artifacts
I1017 19:31:41.718648 442158 start.go:360] acquireMachinesLock for scheduled-stop-721021: {Name:mk6eefae9232cfd2f1e6dd70f554ddf535a83fdb Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I1017 19:31:41.718767 442158 start.go:364] duration metric: took 97.946µs to acquireMachinesLock for "scheduled-stop-721021"
I1017 19:31:41.718799 442158 start.go:93] Provisioning new machine with config: &{Name:scheduled-stop-721021 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760609789-21757@sha256:9824b20f4774128fcb298ad0e6cac7649729886cfba9d444b2305c743a5044c6 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-721021 Namespace:default APIServerHAVIP: A
PIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: Static
IP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
I1017 19:31:41.718857 442158 start.go:125] createHost starting for "" (driver="docker")
I1017 19:31:41.722291 442158 out.go:252] * Creating docker container (CPUs=2, Memory=3072MB) ...
I1017 19:31:41.722519 442158 start.go:159] libmachine.API.Create for "scheduled-stop-721021" (driver="docker")
I1017 19:31:41.722556 442158 client.go:168] LocalClient.Create starting
I1017 19:31:41.722639 442158 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21753-293342/.minikube/certs/ca.pem
I1017 19:31:41.722675 442158 main.go:141] libmachine: Decoding PEM data...
I1017 19:31:41.722690 442158 main.go:141] libmachine: Parsing certificate...
I1017 19:31:41.722743 442158 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21753-293342/.minikube/certs/cert.pem
I1017 19:31:41.722761 442158 main.go:141] libmachine: Decoding PEM data...
I1017 19:31:41.722770 442158 main.go:141] libmachine: Parsing certificate...
I1017 19:31:41.723140 442158 cli_runner.go:164] Run: docker network inspect scheduled-stop-721021 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W1017 19:31:41.739075 442158 cli_runner.go:211] docker network inspect scheduled-stop-721021 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I1017 19:31:41.739197 442158 network_create.go:284] running [docker network inspect scheduled-stop-721021] to gather additional debugging logs...
I1017 19:31:41.739222 442158 cli_runner.go:164] Run: docker network inspect scheduled-stop-721021
W1017 19:31:41.753226 442158 cli_runner.go:211] docker network inspect scheduled-stop-721021 returned with exit code 1
I1017 19:31:41.753245 442158 network_create.go:287] error running [docker network inspect scheduled-stop-721021]: docker network inspect scheduled-stop-721021: exit status 1
stdout:
[]
stderr:
Error response from daemon: network scheduled-stop-721021 not found
I1017 19:31:41.753257 442158 network_create.go:289] output of [docker network inspect scheduled-stop-721021]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network scheduled-stop-721021 not found
** /stderr **
I1017 19:31:41.753362 442158 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I1017 19:31:41.769743 442158 network.go:211] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-c602ac4a53ce IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:d6:3a:bd:94:45:79} reservation:<nil>}
I1017 19:31:41.770021 442158 network.go:211] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-5cb4b5ef9747 IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:ca:5e:47:95:f9:7a} reservation:<nil>}
I1017 19:31:41.770217 442158 network.go:211] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-7a045f7499b4 IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:ae:8a:46:f0:39:62} reservation:<nil>}
I1017 19:31:41.770553 442158 network.go:206] using free private subnet 192.168.76.0/24: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x40019f1050}
I1017 19:31:41.770566 442158 network_create.go:124] attempt to create docker network scheduled-stop-721021 192.168.76.0/24 with gateway 192.168.76.1 and MTU of 1500 ...
I1017 19:31:41.770625 442158 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.76.0/24 --gateway=192.168.76.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=scheduled-stop-721021 scheduled-stop-721021
I1017 19:31:41.830281 442158 network_create.go:108] docker network scheduled-stop-721021 192.168.76.0/24 created
I1017 19:31:41.830304 442158 kic.go:121] calculated static IP "192.168.76.2" for the "scheduled-stop-721021" container
I1017 19:31:41.830378 442158 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I1017 19:31:41.846234 442158 cli_runner.go:164] Run: docker volume create scheduled-stop-721021 --label name.minikube.sigs.k8s.io=scheduled-stop-721021 --label created_by.minikube.sigs.k8s.io=true
I1017 19:31:41.869928 442158 oci.go:103] Successfully created a docker volume scheduled-stop-721021
I1017 19:31:41.870015 442158 cli_runner.go:164] Run: docker run --rm --name scheduled-stop-721021-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-721021 --entrypoint /usr/bin/test -v scheduled-stop-721021:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760609789-21757@sha256:9824b20f4774128fcb298ad0e6cac7649729886cfba9d444b2305c743a5044c6 -d /var/lib
I1017 19:31:42.420279 442158 oci.go:107] Successfully prepared a docker volume scheduled-stop-721021
I1017 19:31:42.420326 442158 preload.go:183] Checking if preload exists for k8s version v1.34.1 and runtime containerd
I1017 19:31:42.420345 442158 kic.go:194] Starting extracting preloaded images to volume ...
I1017 19:31:42.420420 442158 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/21753-293342/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-721021:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760609789-21757@sha256:9824b20f4774128fcb298ad0e6cac7649729886cfba9d444b2305c743a5044c6 -I lz4 -xf /preloaded.tar -C /extractDir
I1017 19:31:46.804508 442158 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/21753-293342/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-721021:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760609789-21757@sha256:9824b20f4774128fcb298ad0e6cac7649729886cfba9d444b2305c743a5044c6 -I lz4 -xf /preloaded.tar -C /extractDir: (4.384037865s)
I1017 19:31:46.804529 442158 kic.go:203] duration metric: took 4.384181661s to extract preloaded images to volume ...
W1017 19:31:46.804675 442158 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I1017 19:31:46.804793 442158 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I1017 19:31:46.871328 442158 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname scheduled-stop-721021 --name scheduled-stop-721021 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-721021 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=scheduled-stop-721021 --network scheduled-stop-721021 --ip 192.168.76.2 --volume scheduled-stop-721021:/var --security-opt apparmor=unconfined --memory=3072mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760609789-21757@sha256:9824b20f4774128fcb298ad0e6cac7649729886cfba9d444b2305c743a5044c6
I1017 19:31:47.218536 442158 cli_runner.go:164] Run: docker container inspect scheduled-stop-721021 --format={{.State.Running}}
I1017 19:31:47.240071 442158 cli_runner.go:164] Run: docker container inspect scheduled-stop-721021 --format={{.State.Status}}
I1017 19:31:47.265360 442158 cli_runner.go:164] Run: docker exec scheduled-stop-721021 stat /var/lib/dpkg/alternatives/iptables
I1017 19:31:47.316996 442158 oci.go:144] the created container "scheduled-stop-721021" has a running status.
I1017 19:31:47.317025 442158 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/21753-293342/.minikube/machines/scheduled-stop-721021/id_rsa...
I1017 19:31:47.860900 442158 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/21753-293342/.minikube/machines/scheduled-stop-721021/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I1017 19:31:47.880245 442158 cli_runner.go:164] Run: docker container inspect scheduled-stop-721021 --format={{.State.Status}}
I1017 19:31:47.897069 442158 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I1017 19:31:47.897079 442158 kic_runner.go:114] Args: [docker exec --privileged scheduled-stop-721021 chown docker:docker /home/docker/.ssh/authorized_keys]
I1017 19:31:47.936660 442158 cli_runner.go:164] Run: docker container inspect scheduled-stop-721021 --format={{.State.Status}}
I1017 19:31:47.953533 442158 machine.go:93] provisionDockerMachine start ...
I1017 19:31:47.953652 442158 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-721021
I1017 19:31:47.970875 442158 main.go:141] libmachine: Using SSH client type: native
I1017 19:31:47.971308 442158 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef110] 0x3f18d0 <nil> [] 0s} 127.0.0.1 33339 <nil> <nil>}
I1017 19:31:47.971316 442158 main.go:141] libmachine: About to run SSH command:
hostname
I1017 19:31:47.971942 442158 main.go:141] libmachine: Error dialing TCP: ssh: handshake failed: read tcp 127.0.0.1:48670->127.0.0.1:33339: read: connection reset by peer
I1017 19:31:51.123137 442158 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-721021
I1017 19:31:51.123152 442158 ubuntu.go:182] provisioning hostname "scheduled-stop-721021"
I1017 19:31:51.123216 442158 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-721021
I1017 19:31:51.142136 442158 main.go:141] libmachine: Using SSH client type: native
I1017 19:31:51.142441 442158 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef110] 0x3f18d0 <nil> [] 0s} 127.0.0.1 33339 <nil> <nil>}
I1017 19:31:51.142450 442158 main.go:141] libmachine: About to run SSH command:
sudo hostname scheduled-stop-721021 && echo "scheduled-stop-721021" | sudo tee /etc/hostname
I1017 19:31:51.300519 442158 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-721021
I1017 19:31:51.300591 442158 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-721021
I1017 19:31:51.319684 442158 main.go:141] libmachine: Using SSH client type: native
I1017 19:31:51.319997 442158 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef110] 0x3f18d0 <nil> [] 0s} 127.0.0.1 33339 <nil> <nil>}
I1017 19:31:51.320011 442158 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sscheduled-stop-721021' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 scheduled-stop-721021/g' /etc/hosts;
else
echo '127.0.1.1 scheduled-stop-721021' | sudo tee -a /etc/hosts;
fi
fi
I1017 19:31:51.467354 442158 main.go:141] libmachine: SSH cmd err, output: <nil>:
I1017 19:31:51.467371 442158 ubuntu.go:188] set auth options {CertDir:/home/jenkins/minikube-integration/21753-293342/.minikube CaCertPath:/home/jenkins/minikube-integration/21753-293342/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/21753-293342/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/21753-293342/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/21753-293342/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/21753-293342/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/21753-293342/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/21753-293342/.minikube}
I1017 19:31:51.467386 442158 ubuntu.go:190] setting up certificates
I1017 19:31:51.467394 442158 provision.go:84] configureAuth start
I1017 19:31:51.467452 442158 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-721021
I1017 19:31:51.486522 442158 provision.go:143] copyHostCerts
I1017 19:31:51.486593 442158 exec_runner.go:144] found /home/jenkins/minikube-integration/21753-293342/.minikube/ca.pem, removing ...
I1017 19:31:51.486606 442158 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21753-293342/.minikube/ca.pem
I1017 19:31:51.486680 442158 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21753-293342/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/21753-293342/.minikube/ca.pem (1082 bytes)
I1017 19:31:51.486776 442158 exec_runner.go:144] found /home/jenkins/minikube-integration/21753-293342/.minikube/cert.pem, removing ...
I1017 19:31:51.486779 442158 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21753-293342/.minikube/cert.pem
I1017 19:31:51.486805 442158 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21753-293342/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/21753-293342/.minikube/cert.pem (1123 bytes)
I1017 19:31:51.486879 442158 exec_runner.go:144] found /home/jenkins/minikube-integration/21753-293342/.minikube/key.pem, removing ...
I1017 19:31:51.486883 442158 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21753-293342/.minikube/key.pem
I1017 19:31:51.486910 442158 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21753-293342/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/21753-293342/.minikube/key.pem (1675 bytes)
I1017 19:31:51.486964 442158 provision.go:117] generating server cert: /home/jenkins/minikube-integration/21753-293342/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/21753-293342/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/21753-293342/.minikube/certs/ca-key.pem org=jenkins.scheduled-stop-721021 san=[127.0.0.1 192.168.76.2 localhost minikube scheduled-stop-721021]
I1017 19:31:52.418222 442158 provision.go:177] copyRemoteCerts
I1017 19:31:52.418280 442158 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I1017 19:31:52.418328 442158 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-721021
I1017 19:31:52.438251 442158 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33339 SSHKeyPath:/home/jenkins/minikube-integration/21753-293342/.minikube/machines/scheduled-stop-721021/id_rsa Username:docker}
I1017 19:31:52.542930 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
I1017 19:31:52.560569 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/machines/server.pem --> /etc/docker/server.pem (1229 bytes)
I1017 19:31:52.579718 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I1017 19:31:52.596399 442158 provision.go:87] duration metric: took 1.128981532s to configureAuth
I1017 19:31:52.596415 442158 ubuntu.go:206] setting minikube options for container-runtime
I1017 19:31:52.596605 442158 config.go:182] Loaded profile config "scheduled-stop-721021": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.34.1
I1017 19:31:52.596610 442158 machine.go:96] duration metric: took 4.643068125s to provisionDockerMachine
I1017 19:31:52.596616 442158 client.go:171] duration metric: took 10.874055845s to LocalClient.Create
I1017 19:31:52.596637 442158 start.go:167] duration metric: took 10.874119132s to libmachine.API.Create "scheduled-stop-721021"
I1017 19:31:52.596643 442158 start.go:293] postStartSetup for "scheduled-stop-721021" (driver="docker")
I1017 19:31:52.596651 442158 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I1017 19:31:52.596702 442158 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I1017 19:31:52.596746 442158 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-721021
I1017 19:31:52.613722 442158 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33339 SSHKeyPath:/home/jenkins/minikube-integration/21753-293342/.minikube/machines/scheduled-stop-721021/id_rsa Username:docker}
I1017 19:31:52.714969 442158 ssh_runner.go:195] Run: cat /etc/os-release
I1017 19:31:52.718301 442158 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I1017 19:31:52.718332 442158 info.go:137] Remote host: Debian GNU/Linux 12 (bookworm)
I1017 19:31:52.718342 442158 filesync.go:126] Scanning /home/jenkins/minikube-integration/21753-293342/.minikube/addons for local assets ...
I1017 19:31:52.718397 442158 filesync.go:126] Scanning /home/jenkins/minikube-integration/21753-293342/.minikube/files for local assets ...
I1017 19:31:52.718484 442158 filesync.go:149] local asset: /home/jenkins/minikube-integration/21753-293342/.minikube/files/etc/ssl/certs/2952072.pem -> 2952072.pem in /etc/ssl/certs
I1017 19:31:52.718591 442158 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I1017 19:31:52.726131 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/files/etc/ssl/certs/2952072.pem --> /etc/ssl/certs/2952072.pem (1708 bytes)
I1017 19:31:52.743599 442158 start.go:296] duration metric: took 146.942099ms for postStartSetup
I1017 19:31:52.743961 442158 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-721021
I1017 19:31:52.761425 442158 profile.go:143] Saving config to /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/config.json ...
I1017 19:31:52.761695 442158 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I1017 19:31:52.761732 442158 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-721021
I1017 19:31:52.779268 442158 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33339 SSHKeyPath:/home/jenkins/minikube-integration/21753-293342/.minikube/machines/scheduled-stop-721021/id_rsa Username:docker}
I1017 19:31:52.879949 442158 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I1017 19:31:52.884538 442158 start.go:128] duration metric: took 11.165666939s to createHost
I1017 19:31:52.884551 442158 start.go:83] releasing machines lock for "scheduled-stop-721021", held for 11.165777185s
I1017 19:31:52.884618 442158 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-721021
I1017 19:31:52.901570 442158 ssh_runner.go:195] Run: cat /version.json
I1017 19:31:52.901608 442158 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-721021
I1017 19:31:52.901851 442158 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I1017 19:31:52.901900 442158 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-721021
I1017 19:31:52.922386 442158 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33339 SSHKeyPath:/home/jenkins/minikube-integration/21753-293342/.minikube/machines/scheduled-stop-721021/id_rsa Username:docker}
I1017 19:31:52.923149 442158 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33339 SSHKeyPath:/home/jenkins/minikube-integration/21753-293342/.minikube/machines/scheduled-stop-721021/id_rsa Username:docker}
I1017 19:31:53.022847 442158 ssh_runner.go:195] Run: systemctl --version
I1017 19:31:53.115435 442158 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
W1017 19:31:53.120120 442158 cni.go:209] loopback cni configuration skipped: "/etc/cni/net.d/*loopback.conf*" not found
I1017 19:31:53.120183 442158 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I1017 19:31:53.148666 442158 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/10-crio-bridge.conflist.disabled] bridge cni config(s)
I1017 19:31:53.148679 442158 start.go:495] detecting cgroup driver to use...
I1017 19:31:53.148711 442158 detect.go:187] detected "cgroupfs" cgroup driver on host os
I1017 19:31:53.148772 442158 ssh_runner.go:195] Run: sudo systemctl stop -f crio
I1017 19:31:53.164261 442158 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I1017 19:31:53.177237 442158 docker.go:218] disabling cri-docker service (if available) ...
I1017 19:31:53.177307 442158 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
I1017 19:31:53.195986 442158 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
I1017 19:31:53.213684 442158 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
I1017 19:31:53.330161 442158 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
I1017 19:31:53.450630 442158 docker.go:234] disabling docker service ...
I1017 19:31:53.450706 442158 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
I1017 19:31:53.474003 442158 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
I1017 19:31:53.488287 442158 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
I1017 19:31:53.605787 442158 ssh_runner.go:195] Run: sudo systemctl mask docker.service
I1017 19:31:53.722513 442158 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
I1017 19:31:53.737245 442158 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I1017 19:31:53.752377 442158 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10.1"|' /etc/containerd/config.toml"
I1017 19:31:53.761797 442158 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I1017 19:31:53.771048 442158 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I1017 19:31:53.771128 442158 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I1017 19:31:53.781458 442158 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I1017 19:31:53.790751 442158 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I1017 19:31:53.799529 442158 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I1017 19:31:53.812420 442158 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I1017 19:31:53.821139 442158 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I1017 19:31:53.830331 442158 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I1017 19:31:53.839169 442158 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I1017 19:31:53.848190 442158 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I1017 19:31:53.856212 442158 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I1017 19:31:53.863760 442158 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1017 19:31:53.973835 442158 ssh_runner.go:195] Run: sudo systemctl restart containerd
I1017 19:31:54.128075 442158 start.go:542] Will wait 60s for socket path /run/containerd/containerd.sock
I1017 19:31:54.128137 442158 ssh_runner.go:195] Run: stat /run/containerd/containerd.sock
I1017 19:31:54.132130 442158 start.go:563] Will wait 60s for crictl version
I1017 19:31:54.132191 442158 ssh_runner.go:195] Run: which crictl
I1017 19:31:54.135818 442158 ssh_runner.go:195] Run: sudo /usr/local/bin/crictl version
I1017 19:31:54.169207 442158 start.go:579] Version: 0.1.0
RuntimeName: containerd
RuntimeVersion: v1.7.28
RuntimeApiVersion: v1
I1017 19:31:54.169274 442158 ssh_runner.go:195] Run: containerd --version
I1017 19:31:54.191047 442158 ssh_runner.go:195] Run: containerd --version
I1017 19:31:54.217993 442158 out.go:179] * Preparing Kubernetes v1.34.1 on containerd 1.7.28 ...
I1017 19:31:54.220928 442158 cli_runner.go:164] Run: docker network inspect scheduled-stop-721021 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I1017 19:31:54.237156 442158 ssh_runner.go:195] Run: grep 192.168.76.1 host.minikube.internal$ /etc/hosts
I1017 19:31:54.240909 442158 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.76.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I1017 19:31:54.250928 442158 kubeadm.go:883] updating cluster {Name:scheduled-stop-721021 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760609789-21757@sha256:9824b20f4774128fcb298ad0e6cac7649729886cfba9d444b2305c743a5044c6 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-721021 Namespace:default APIServerHAVIP: APIServerName:miniku
beCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSH
AuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I1017 19:31:54.251068 442158 preload.go:183] Checking if preload exists for k8s version v1.34.1 and runtime containerd
I1017 19:31:54.251149 442158 ssh_runner.go:195] Run: sudo crictl images --output json
I1017 19:31:54.275909 442158 containerd.go:627] all images are preloaded for containerd runtime.
I1017 19:31:54.275921 442158 containerd.go:534] Images already preloaded, skipping extraction
I1017 19:31:54.275979 442158 ssh_runner.go:195] Run: sudo crictl images --output json
I1017 19:31:54.299814 442158 containerd.go:627] all images are preloaded for containerd runtime.
I1017 19:31:54.299827 442158 cache_images.go:85] Images are preloaded, skipping loading
I1017 19:31:54.299834 442158 kubeadm.go:934] updating node { 192.168.76.2 8443 v1.34.1 containerd true true} ...
I1017 19:31:54.299935 442158 kubeadm.go:946] kubelet [Unit]
Wants=containerd.service
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.34.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=scheduled-stop-721021 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.76.2
[Install]
config:
{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-721021 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I1017 19:31:54.299998 442158 ssh_runner.go:195] Run: sudo crictl info
I1017 19:31:54.325313 442158 cni.go:84] Creating CNI manager for ""
I1017 19:31:54.325323 442158 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I1017 19:31:54.325342 442158 kubeadm.go:85] Using pod CIDR: 10.244.0.0/16
I1017 19:31:54.325365 442158 kubeadm.go:190] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.76.2 APIServerPort:8443 KubernetesVersion:v1.34.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:scheduled-stop-721021 NodeName:scheduled-stop-721021 DNSDomain:cluster.local CRISocket:/run/containerd/containerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.76.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.76.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt St
aticPodPath:/etc/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///run/containerd/containerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I1017 19:31:54.325480 442158 kubeadm.go:196] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta4
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.76.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///run/containerd/containerd.sock
name: "scheduled-stop-721021"
kubeletExtraArgs:
- name: "node-ip"
value: "192.168.76.2"
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta4
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.76.2"]
extraArgs:
- name: "enable-admission-plugins"
value: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
- name: "allocate-node-cidrs"
value: "true"
- name: "leader-elect"
value: "false"
scheduler:
extraArgs:
- name: "leader-elect"
value: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
kubernetesVersion: v1.34.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///run/containerd/containerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I1017 19:31:54.325548 442158 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.34.1
I1017 19:31:54.333062 442158 binaries.go:44] Found k8s binaries, skipping transfer
I1017 19:31:54.333130 442158 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I1017 19:31:54.340856 442158 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (325 bytes)
I1017 19:31:54.354024 442158 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I1017 19:31:54.367495 442158 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2234 bytes)
I1017 19:31:54.380571 442158 ssh_runner.go:195] Run: grep 192.168.76.2 control-plane.minikube.internal$ /etc/hosts
I1017 19:31:54.384250 442158 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.76.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I1017 19:31:54.394176 442158 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1017 19:31:54.508691 442158 ssh_runner.go:195] Run: sudo systemctl start kubelet
I1017 19:31:54.524742 442158 certs.go:69] Setting up /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021 for IP: 192.168.76.2
I1017 19:31:54.524766 442158 certs.go:195] generating shared ca certs ...
I1017 19:31:54.524781 442158 certs.go:227] acquiring lock for ca certs: {Name:mk38f7c861e302c587a0e6a579ad0896fc72eea1 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1017 19:31:54.524949 442158 certs.go:236] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/21753-293342/.minikube/ca.key
I1017 19:31:54.524999 442158 certs.go:236] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/21753-293342/.minikube/proxy-client-ca.key
I1017 19:31:54.525005 442158 certs.go:257] generating profile certs ...
I1017 19:31:54.525069 442158 certs.go:364] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/client.key
I1017 19:31:54.525089 442158 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/client.crt with IP's: []
I1017 19:31:55.514789 442158 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/client.crt ...
I1017 19:31:55.514805 442158 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/client.crt: {Name:mk485e4a75fb226dae85e89875ff14fb7b8cc918 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1017 19:31:55.515014 442158 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/client.key ...
I1017 19:31:55.515027 442158 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/client.key: {Name:mk603245e857a24f80eda3efe410a29aa46b8faa Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1017 19:31:55.515150 442158 certs.go:364] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/apiserver.key.813953e3
I1017 19:31:55.515164 442158 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/apiserver.crt.813953e3 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.76.2]
I1017 19:31:55.885938 442158 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/apiserver.crt.813953e3 ...
I1017 19:31:55.885953 442158 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/apiserver.crt.813953e3: {Name:mk9936a70c4c3e71ad78991fe87906383777c246 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1017 19:31:55.886156 442158 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/apiserver.key.813953e3 ...
I1017 19:31:55.886164 442158 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/apiserver.key.813953e3: {Name:mk2743c4147d6f307d8150662c09e02c418816d2 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1017 19:31:55.886246 442158 certs.go:382] copying /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/apiserver.crt.813953e3 -> /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/apiserver.crt
I1017 19:31:55.886336 442158 certs.go:386] copying /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/apiserver.key.813953e3 -> /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/apiserver.key
I1017 19:31:55.886390 442158 certs.go:364] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/proxy-client.key
I1017 19:31:55.886403 442158 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/proxy-client.crt with IP's: []
I1017 19:31:56.369377 442158 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/proxy-client.crt ...
I1017 19:31:56.369392 442158 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/proxy-client.crt: {Name:mk0d018c0261bd80e93325b78911896ab720fad6 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1017 19:31:56.369578 442158 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/proxy-client.key ...
I1017 19:31:56.369585 442158 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/proxy-client.key: {Name:mk51464a33b8ec326c813f5fcc750a7f49981f13 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1017 19:31:56.369767 442158 certs.go:484] found cert: /home/jenkins/minikube-integration/21753-293342/.minikube/certs/295207.pem (1338 bytes)
W1017 19:31:56.369803 442158 certs.go:480] ignoring /home/jenkins/minikube-integration/21753-293342/.minikube/certs/295207_empty.pem, impossibly tiny 0 bytes
I1017 19:31:56.369810 442158 certs.go:484] found cert: /home/jenkins/minikube-integration/21753-293342/.minikube/certs/ca-key.pem (1675 bytes)
I1017 19:31:56.369840 442158 certs.go:484] found cert: /home/jenkins/minikube-integration/21753-293342/.minikube/certs/ca.pem (1082 bytes)
I1017 19:31:56.369870 442158 certs.go:484] found cert: /home/jenkins/minikube-integration/21753-293342/.minikube/certs/cert.pem (1123 bytes)
I1017 19:31:56.369894 442158 certs.go:484] found cert: /home/jenkins/minikube-integration/21753-293342/.minikube/certs/key.pem (1675 bytes)
I1017 19:31:56.369935 442158 certs.go:484] found cert: /home/jenkins/minikube-integration/21753-293342/.minikube/files/etc/ssl/certs/2952072.pem (1708 bytes)
I1017 19:31:56.370520 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I1017 19:31:56.389330 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I1017 19:31:56.408035 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I1017 19:31:56.426846 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I1017 19:31:56.445829 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1432 bytes)
I1017 19:31:56.463805 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I1017 19:31:56.480728 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I1017 19:31:56.499072 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/profiles/scheduled-stop-721021/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
I1017 19:31:56.517052 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I1017 19:31:56.534656 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/certs/295207.pem --> /usr/share/ca-certificates/295207.pem (1338 bytes)
I1017 19:31:56.551799 442158 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21753-293342/.minikube/files/etc/ssl/certs/2952072.pem --> /usr/share/ca-certificates/2952072.pem (1708 bytes)
I1017 19:31:56.569452 442158 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I1017 19:31:56.581912 442158 ssh_runner.go:195] Run: openssl version
I1017 19:31:56.588067 442158 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/295207.pem && ln -fs /usr/share/ca-certificates/295207.pem /etc/ssl/certs/295207.pem"
I1017 19:31:56.596286 442158 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/295207.pem
I1017 19:31:56.599989 442158 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 Oct 17 19:04 /usr/share/ca-certificates/295207.pem
I1017 19:31:56.600043 442158 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/295207.pem
I1017 19:31:56.640892 442158 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/295207.pem /etc/ssl/certs/51391683.0"
I1017 19:31:56.649168 442158 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/2952072.pem && ln -fs /usr/share/ca-certificates/2952072.pem /etc/ssl/certs/2952072.pem"
I1017 19:31:56.657289 442158 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/2952072.pem
I1017 19:31:56.660782 442158 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 Oct 17 19:04 /usr/share/ca-certificates/2952072.pem
I1017 19:31:56.660836 442158 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/2952072.pem
I1017 19:31:56.704049 442158 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/2952072.pem /etc/ssl/certs/3ec20f2e.0"
I1017 19:31:56.713367 442158 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I1017 19:31:56.721865 442158 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I1017 19:31:56.725695 442158 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Oct 17 18:57 /usr/share/ca-certificates/minikubeCA.pem
I1017 19:31:56.725752 442158 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I1017 19:31:56.766567 442158 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I1017 19:31:56.774705 442158 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I1017 19:31:56.778497 442158 certs.go:400] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I1017 19:31:56.778543 442158 kubeadm.go:400] StartCluster: {Name:scheduled-stop-721021 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760609789-21757@sha256:9824b20f4774128fcb298ad0e6cac7649729886cfba9d444b2305c743a5044c6 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-721021 Namespace:default APIServerHAVIP: APIServerName:minikubeC
A APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAut
hSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I1017 19:31:56.778611 442158 cri.go:54] listing CRI containers in root /run/containerd/runc/k8s.io: {State:paused Name: Namespaces:[kube-system]}
I1017 19:31:56.778681 442158 ssh_runner.go:195] Run: sudo -s eval "crictl ps -a --quiet --label io.kubernetes.pod.namespace=kube-system"
I1017 19:31:56.803920 442158 cri.go:89] found id: ""
I1017 19:31:56.803990 442158 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I1017 19:31:56.811704 442158 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I1017 19:31:56.819746 442158 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I1017 19:31:56.819804 442158 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I1017 19:31:56.827709 442158 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I1017 19:31:56.827718 442158 kubeadm.go:157] found existing configuration files:
I1017 19:31:56.827769 442158 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I1017 19:31:56.835466 442158 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I1017 19:31:56.835526 442158 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I1017 19:31:56.843325 442158 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I1017 19:31:56.851205 442158 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I1017 19:31:56.851269 442158 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I1017 19:31:56.858544 442158 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I1017 19:31:56.866156 442158 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I1017 19:31:56.866217 442158 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I1017 19:31:56.874246 442158 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I1017 19:31:56.881917 442158 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I1017 19:31:56.881977 442158 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I1017 19:31:56.889207 442158 ssh_runner.go:286] Start: sudo /bin/bash -c "env PATH="/var/lib/minikube/binaries/v1.34.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I1017 19:31:56.946671 442158 kubeadm.go:318] [init] Using Kubernetes version: v1.34.1
I1017 19:31:56.946756 442158 kubeadm.go:318] [preflight] Running pre-flight checks
I1017 19:31:56.976338 442158 kubeadm.go:318] [preflight] The system verification failed. Printing the output from the verification:
I1017 19:31:56.976405 442158 kubeadm.go:318] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1084-aws[0m
I1017 19:31:56.976441 442158 kubeadm.go:318] [0;37mOS[0m: [0;32mLinux[0m
I1017 19:31:56.976487 442158 kubeadm.go:318] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I1017 19:31:56.976536 442158 kubeadm.go:318] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I1017 19:31:56.976584 442158 kubeadm.go:318] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I1017 19:31:56.976633 442158 kubeadm.go:318] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I1017 19:31:56.976682 442158 kubeadm.go:318] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I1017 19:31:56.976755 442158 kubeadm.go:318] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I1017 19:31:56.976801 442158 kubeadm.go:318] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I1017 19:31:56.976850 442158 kubeadm.go:318] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I1017 19:31:56.976900 442158 kubeadm.go:318] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I1017 19:31:57.059067 442158 kubeadm.go:318] [preflight] Pulling images required for setting up a Kubernetes cluster
I1017 19:31:57.059192 442158 kubeadm.go:318] [preflight] This might take a minute or two, depending on the speed of your internet connection
I1017 19:31:57.059285 442158 kubeadm.go:318] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I1017 19:31:57.071464 442158 kubeadm.go:318] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I1017 19:31:57.075322 442158 out.go:252] - Generating certificates and keys ...
I1017 19:31:57.075423 442158 kubeadm.go:318] [certs] Using existing ca certificate authority
I1017 19:31:57.075514 442158 kubeadm.go:318] [certs] Using existing apiserver certificate and key on disk
I1017 19:31:58.546123 442158 kubeadm.go:318] [certs] Generating "apiserver-kubelet-client" certificate and key
I1017 19:31:58.951173 442158 kubeadm.go:318] [certs] Generating "front-proxy-ca" certificate and key
I1017 19:32:00.206787 442158 kubeadm.go:318] [certs] Generating "front-proxy-client" certificate and key
I1017 19:32:00.742874 442158 kubeadm.go:318] [certs] Generating "etcd/ca" certificate and key
I1017 19:32:01.409966 442158 kubeadm.go:318] [certs] Generating "etcd/server" certificate and key
I1017 19:32:01.410496 442158 kubeadm.go:318] [certs] etcd/server serving cert is signed for DNS names [localhost scheduled-stop-721021] and IPs [192.168.76.2 127.0.0.1 ::1]
I1017 19:32:02.140401 442158 kubeadm.go:318] [certs] Generating "etcd/peer" certificate and key
I1017 19:32:02.140728 442158 kubeadm.go:318] [certs] etcd/peer serving cert is signed for DNS names [localhost scheduled-stop-721021] and IPs [192.168.76.2 127.0.0.1 ::1]
I1017 19:32:02.242326 442158 kubeadm.go:318] [certs] Generating "etcd/healthcheck-client" certificate and key
I1017 19:32:03.446106 442158 kubeadm.go:318] [certs] Generating "apiserver-etcd-client" certificate and key
I1017 19:32:03.709967 442158 kubeadm.go:318] [certs] Generating "sa" key and public key
I1017 19:32:03.710103 442158 kubeadm.go:318] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I1017 19:32:04.083319 442158 kubeadm.go:318] [kubeconfig] Writing "admin.conf" kubeconfig file
I1017 19:32:05.394916 442158 kubeadm.go:318] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I1017 19:32:05.485206 442158 kubeadm.go:318] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I1017 19:32:05.956149 442158 kubeadm.go:318] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I1017 19:32:06.179160 442158 kubeadm.go:318] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I1017 19:32:06.179743 442158 kubeadm.go:318] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I1017 19:32:06.182447 442158 kubeadm.go:318] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I1017 19:32:06.185959 442158 out.go:252] - Booting up control plane ...
I1017 19:32:06.186061 442158 kubeadm.go:318] [control-plane] Creating static Pod manifest for "kube-apiserver"
I1017 19:32:06.186137 442158 kubeadm.go:318] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I1017 19:32:06.186202 442158 kubeadm.go:318] [control-plane] Creating static Pod manifest for "kube-scheduler"
I1017 19:32:06.204593 442158 kubeadm.go:318] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I1017 19:32:06.204890 442158 kubeadm.go:318] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/instance-config.yaml"
I1017 19:32:06.212719 442158 kubeadm.go:318] [patches] Applied patch of type "application/strategic-merge-patch+json" to target "kubeletconfiguration"
I1017 19:32:06.213111 442158 kubeadm.go:318] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I1017 19:32:06.213363 442158 kubeadm.go:318] [kubelet-start] Starting the kubelet
I1017 19:32:06.370880 442158 kubeadm.go:318] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I1017 19:32:06.370991 442158 kubeadm.go:318] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I1017 19:32:08.367522 442158 kubeadm.go:318] [kubelet-check] The kubelet is healthy after 2.0010838s
I1017 19:32:08.367630 442158 kubeadm.go:318] [control-plane-check] Waiting for healthy control plane components. This can take up to 4m0s
I1017 19:32:08.367714 442158 kubeadm.go:318] [control-plane-check] Checking kube-apiserver at https://192.168.76.2:8443/livez
I1017 19:32:08.368052 442158 kubeadm.go:318] [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz
I1017 19:32:08.368145 442158 kubeadm.go:318] [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez
I1017 19:32:11.541957 442158 kubeadm.go:318] [control-plane-check] kube-controller-manager is healthy after 3.173412518s
I1017 19:32:13.857051 442158 kubeadm.go:318] [control-plane-check] kube-scheduler is healthy after 5.488691125s
I1017 19:32:14.870574 442158 kubeadm.go:318] [control-plane-check] kube-apiserver is healthy after 6.50241565s
I1017 19:32:14.890512 442158 kubeadm.go:318] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I1017 19:32:14.913453 442158 kubeadm.go:318] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I1017 19:32:14.929872 442158 kubeadm.go:318] [upload-certs] Skipping phase. Please see --upload-certs
I1017 19:32:14.930069 442158 kubeadm.go:318] [mark-control-plane] Marking the node scheduled-stop-721021 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I1017 19:32:14.948603 442158 kubeadm.go:318] [bootstrap-token] Using token: a8ob4l.6wqycmui5pgezkcw
I1017 19:32:14.951647 442158 out.go:252] - Configuring RBAC rules ...
I1017 19:32:14.951797 442158 kubeadm.go:318] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I1017 19:32:14.959078 442158 kubeadm.go:318] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I1017 19:32:14.968539 442158 kubeadm.go:318] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I1017 19:32:14.977146 442158 kubeadm.go:318] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I1017 19:32:14.982487 442158 kubeadm.go:318] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I1017 19:32:14.988229 442158 kubeadm.go:318] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I1017 19:32:15.277566 442158 kubeadm.go:318] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I1017 19:32:15.703930 442158 kubeadm.go:318] [addons] Applied essential addon: CoreDNS
I1017 19:32:16.279881 442158 kubeadm.go:318] [addons] Applied essential addon: kube-proxy
I1017 19:32:16.281039 442158 kubeadm.go:318]
I1017 19:32:16.281107 442158 kubeadm.go:318] Your Kubernetes control-plane has initialized successfully!
I1017 19:32:16.281112 442158 kubeadm.go:318]
I1017 19:32:16.281191 442158 kubeadm.go:318] To start using your cluster, you need to run the following as a regular user:
I1017 19:32:16.281195 442158 kubeadm.go:318]
I1017 19:32:16.281221 442158 kubeadm.go:318] mkdir -p $HOME/.kube
I1017 19:32:16.281281 442158 kubeadm.go:318] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I1017 19:32:16.281333 442158 kubeadm.go:318] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I1017 19:32:16.281337 442158 kubeadm.go:318]
I1017 19:32:16.281392 442158 kubeadm.go:318] Alternatively, if you are the root user, you can run:
I1017 19:32:16.281396 442158 kubeadm.go:318]
I1017 19:32:16.281445 442158 kubeadm.go:318] export KUBECONFIG=/etc/kubernetes/admin.conf
I1017 19:32:16.281452 442158 kubeadm.go:318]
I1017 19:32:16.281506 442158 kubeadm.go:318] You should now deploy a pod network to the cluster.
I1017 19:32:16.281584 442158 kubeadm.go:318] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I1017 19:32:16.281654 442158 kubeadm.go:318] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I1017 19:32:16.281657 442158 kubeadm.go:318]
I1017 19:32:16.281744 442158 kubeadm.go:318] You can now join any number of control-plane nodes by copying certificate authorities
I1017 19:32:16.281823 442158 kubeadm.go:318] and service account keys on each node and then running the following as root:
I1017 19:32:16.281827 442158 kubeadm.go:318]
I1017 19:32:16.281937 442158 kubeadm.go:318] kubeadm join control-plane.minikube.internal:8443 --token a8ob4l.6wqycmui5pgezkcw \
I1017 19:32:16.282045 442158 kubeadm.go:318] --discovery-token-ca-cert-hash sha256:eee4af4104cb622b3eef346e3910ffa72395facf14ceade3ed2ea1e0b7c7339d \
I1017 19:32:16.282065 442158 kubeadm.go:318] --control-plane
I1017 19:32:16.282069 442158 kubeadm.go:318]
I1017 19:32:16.282157 442158 kubeadm.go:318] Then you can join any number of worker nodes by running the following on each as root:
I1017 19:32:16.282160 442158 kubeadm.go:318]
I1017 19:32:16.282245 442158 kubeadm.go:318] kubeadm join control-plane.minikube.internal:8443 --token a8ob4l.6wqycmui5pgezkcw \
I1017 19:32:16.282350 442158 kubeadm.go:318] --discovery-token-ca-cert-hash sha256:eee4af4104cb622b3eef346e3910ffa72395facf14ceade3ed2ea1e0b7c7339d
I1017 19:32:16.285491 442158 kubeadm.go:318] [WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2
I1017 19:32:16.285707 442158 kubeadm.go:318] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1084-aws\n", err: exit status 1
I1017 19:32:16.285809 442158 kubeadm.go:318] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I1017 19:32:16.285824 442158 cni.go:84] Creating CNI manager for ""
I1017 19:32:16.285830 442158 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I1017 19:32:16.289100 442158 out.go:179] * Configuring CNI (Container Networking Interface) ...
I1017 19:32:16.292141 442158 ssh_runner.go:195] Run: stat /opt/cni/bin/portmap
I1017 19:32:16.299845 442158 cni.go:182] applying CNI manifest using /var/lib/minikube/binaries/v1.34.1/kubectl ...
I1017 19:32:16.299855 442158 ssh_runner.go:362] scp memory --> /var/tmp/minikube/cni.yaml (2601 bytes)
I1017 19:32:16.313302 442158 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.1/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml
I1017 19:32:16.603022 442158 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I1017 19:32:16.603172 442158 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I1017 19:32:16.603251 442158 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes scheduled-stop-721021 minikube.k8s.io/updated_at=2025_10_17T19_32_16_0700 minikube.k8s.io/version=v1.37.0 minikube.k8s.io/commit=73a80cc9bc99174c010556d98400e9fa16adda9d minikube.k8s.io/name=scheduled-stop-721021 minikube.k8s.io/primary=true
I1017 19:32:16.777311 442158 ops.go:34] apiserver oom_adj: -16
I1017 19:32:16.777337 442158 kubeadm.go:1113] duration metric: took 174.23611ms to wait for elevateKubeSystemPrivileges
I1017 19:32:16.777349 442158 kubeadm.go:402] duration metric: took 19.998811234s to StartCluster
I1017 19:32:16.777366 442158 settings.go:142] acquiring lock: {Name:mk22a110852cb75bdd72d5afb7ba21aef3607ebd Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1017 19:32:16.777427 442158 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/21753-293342/kubeconfig
I1017 19:32:16.778100 442158 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21753-293342/kubeconfig: {Name:mk81d724f6ee552ab600164b5de855c00bf45d8b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1017 19:32:16.778297 442158 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
I1017 19:32:16.778380 442158 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I1017 19:32:16.778619 442158 config.go:182] Loaded profile config "scheduled-stop-721021": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.34.1
I1017 19:32:16.778619 442158 addons.go:511] enable addons start: toEnable=map[ambassador:false amd-gpu-device-plugin:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubetail:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-rancher:false volcano:false volumesnapshots:false yakd:false]
I1017 19:32:16.778756 442158 addons.go:69] Setting storage-provisioner=true in profile "scheduled-stop-721021"
I1017 19:32:16.778771 442158 addons.go:238] Setting addon storage-provisioner=true in "scheduled-stop-721021"
I1017 19:32:16.778777 442158 addons.go:69] Setting default-storageclass=true in profile "scheduled-stop-721021"
I1017 19:32:16.778790 442158 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "scheduled-stop-721021"
I1017 19:32:16.778802 442158 host.go:66] Checking if "scheduled-stop-721021" exists ...
I1017 19:32:16.779244 442158 cli_runner.go:164] Run: docker container inspect scheduled-stop-721021 --format={{.State.Status}}
I1017 19:32:16.779322 442158 cli_runner.go:164] Run: docker container inspect scheduled-stop-721021 --format={{.State.Status}}
I1017 19:32:16.782265 442158 out.go:179] * Verifying Kubernetes components...
I1017 19:32:16.786086 442158 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1017 19:32:16.820197 442158 addons.go:238] Setting addon default-storageclass=true in "scheduled-stop-721021"
I1017 19:32:16.820227 442158 host.go:66] Checking if "scheduled-stop-721021" exists ...
I1017 19:32:16.823051 442158 cli_runner.go:164] Run: docker container inspect scheduled-stop-721021 --format={{.State.Status}}
I1017 19:32:16.830716 442158 out.go:179] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I1017 19:32:16.836171 442158 addons.go:435] installing /etc/kubernetes/addons/storage-provisioner.yaml
I1017 19:32:16.836183 442158 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I1017 19:32:16.836268 442158 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-721021
I1017 19:32:16.852762 442158 addons.go:435] installing /etc/kubernetes/addons/storageclass.yaml
I1017 19:32:16.852775 442158 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I1017 19:32:16.852835 442158 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-721021
I1017 19:32:16.878903 442158 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33339 SSHKeyPath:/home/jenkins/minikube-integration/21753-293342/.minikube/machines/scheduled-stop-721021/id_rsa Username:docker}
I1017 19:32:16.889446 442158 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33339 SSHKeyPath:/home/jenkins/minikube-integration/21753-293342/.minikube/machines/scheduled-stop-721021/id_rsa Username:docker}
I1017 19:32:17.153817 442158 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.76.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I1017 19:32:17.153903 442158 ssh_runner.go:195] Run: sudo systemctl start kubelet
I1017 19:32:17.165201 442158 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.34.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I1017 19:32:17.229203 442158 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.34.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I1017 19:32:17.529044 442158 api_server.go:52] waiting for apiserver process to appear ...
I1017 19:32:17.529099 442158 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I1017 19:32:17.529170 442158 start.go:976] {"host.minikube.internal": 192.168.76.1} host record injected into CoreDNS's ConfigMap
I1017 19:32:17.814443 442158 api_server.go:72] duration metric: took 1.036112366s to wait for apiserver process to appear ...
I1017 19:32:17.814455 442158 api_server.go:88] waiting for apiserver healthz status ...
I1017 19:32:17.814470 442158 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
I1017 19:32:17.837109 442158 api_server.go:279] https://192.168.76.2:8443/healthz returned 200:
ok
I1017 19:32:17.838387 442158 api_server.go:141] control plane version: v1.34.1
I1017 19:32:17.838402 442158 api_server.go:131] duration metric: took 23.941659ms to wait for apiserver health ...
I1017 19:32:17.838409 442158 system_pods.go:43] waiting for kube-system pods to appear ...
I1017 19:32:17.841535 442158 out.go:179] * Enabled addons: storage-provisioner, default-storageclass
I1017 19:32:17.843024 442158 system_pods.go:59] 5 kube-system pods found
I1017 19:32:17.843045 442158 system_pods.go:61] "etcd-scheduled-stop-721021" [f6912f9b-b6df-4314-b571-611bc0ec2d7d] Running / Ready:ContainersNotReady (containers with unready status: [etcd]) / ContainersReady:ContainersNotReady (containers with unready status: [etcd])
I1017 19:32:17.843057 442158 system_pods.go:61] "kube-apiserver-scheduled-stop-721021" [60eeec4a-d8a4-49e4-9a63-afe716a5d5c2] Running / Ready:ContainersNotReady (containers with unready status: [kube-apiserver]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-apiserver])
I1017 19:32:17.843065 442158 system_pods.go:61] "kube-controller-manager-scheduled-stop-721021" [7352d2b1-11f1-4309-bd98-c5678f7c770f] Running / Ready:ContainersNotReady (containers with unready status: [kube-controller-manager]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-controller-manager])
I1017 19:32:17.843075 442158 system_pods.go:61] "kube-scheduler-scheduled-stop-721021" [db423157-9c38-4353-be27-3125ae8585ef] Running / Ready:ContainersNotReady (containers with unready status: [kube-scheduler]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-scheduler])
I1017 19:32:17.843087 442158 system_pods.go:61] "storage-provisioner" [ede2419e-4ffc-49cc-9775-8817fad057ef] Pending: PodScheduled:Unschedulable (0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. no new claims to deallocate, preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.)
I1017 19:32:17.843092 442158 system_pods.go:74] duration metric: took 4.679107ms to wait for pod list to return data ...
I1017 19:32:17.843130 442158 kubeadm.go:586] duration metric: took 1.064786097s to wait for: map[apiserver:true system_pods:true]
I1017 19:32:17.843143 442158 node_conditions.go:102] verifying NodePressure condition ...
I1017 19:32:17.846996 442158 addons.go:514] duration metric: took 1.06837189s for enable addons: enabled=[storage-provisioner default-storageclass]
I1017 19:32:17.847344 442158 node_conditions.go:122] node storage ephemeral capacity is 203034800Ki
I1017 19:32:17.847360 442158 node_conditions.go:123] node cpu capacity is 2
I1017 19:32:17.847371 442158 node_conditions.go:105] duration metric: took 4.224447ms to run NodePressure ...
I1017 19:32:17.847382 442158 start.go:241] waiting for startup goroutines ...
I1017 19:32:18.032695 442158 kapi.go:214] "coredns" deployment in "kube-system" namespace and "scheduled-stop-721021" context rescaled to 1 replicas
I1017 19:32:18.032729 442158 start.go:246] waiting for cluster config update ...
I1017 19:32:18.032741 442158 start.go:255] writing updated cluster config ...
I1017 19:32:18.033048 442158 ssh_runner.go:195] Run: rm -f paused
I1017 19:32:18.096956 442158 start.go:624] kubectl: 1.33.2, cluster: 1.34.1 (minor skew: 1)
I1017 19:32:18.100334 442158 out.go:179] * Done! kubectl is now configured to use "scheduled-stop-721021" cluster and "default" namespace by default
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD NAMESPACE
5354ad5c007c4 a1894772a478e 10 seconds ago Running etcd 0 8351682527531 etcd-scheduled-stop-721021 kube-system
492199d8e7b9c 43911e833d64d 10 seconds ago Running kube-apiserver 0 0cbae525c76c1 kube-apiserver-scheduled-stop-721021 kube-system
add436cf75fd2 b5f57ec6b9867 10 seconds ago Running kube-scheduler 0 9fbee9d343450 kube-scheduler-scheduled-stop-721021 kube-system
7f65dc7e156e5 7eb2c6ff0c5a7 10 seconds ago Running kube-controller-manager 0 89f95609b8eb4 kube-controller-manager-scheduled-stop-721021 kube-system
==> containerd <==
Oct 17 19:31:54 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:31:54.127036382Z" level=info msg="containerd successfully booted in 0.085286s"
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.546811784Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:etcd-scheduled-stop-721021,Uid:74087c1f265e54593e818a1bdf8a438e,Namespace:kube-system,Attempt:0,}"
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.550405208Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-scheduled-stop-721021,Uid:7e8f0033fca107fc141f37281e43bf2c,Namespace:kube-system,Attempt:0,}"
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.554233531Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-scheduled-stop-721021,Uid:8afd441a577fb9029dd8a39d1dbb3251,Namespace:kube-system,Attempt:0,}"
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.558239128Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-scheduled-stop-721021,Uid:261ae8e6a197ea01ba4a6cb85ea7fe96,Namespace:kube-system,Attempt:0,}"
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.755661836Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-scheduled-stop-721021,Uid:261ae8e6a197ea01ba4a6cb85ea7fe96,Namespace:kube-system,Attempt:0,} returns sandbox id \"9fbee9d343450830d5713d13c05a89cde060d9086276e57f737495e2d6b2a4d3\""
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.756135458Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-scheduled-stop-721021,Uid:8afd441a577fb9029dd8a39d1dbb3251,Namespace:kube-system,Attempt:0,} returns sandbox id \"89f95609b8eb443f0901e3a372da82d7f82eb270267c3202f63f7bd5091c8fcc\""
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.765093372Z" level=info msg="CreateContainer within sandbox \"9fbee9d343450830d5713d13c05a89cde060d9086276e57f737495e2d6b2a4d3\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}"
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.768473690Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-scheduled-stop-721021,Uid:7e8f0033fca107fc141f37281e43bf2c,Namespace:kube-system,Attempt:0,} returns sandbox id \"0cbae525c76c173ef2998af3a8ce3a269a11200f75aa54529af4e9fbca1fff6f\""
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.770872432Z" level=info msg="CreateContainer within sandbox \"89f95609b8eb443f0901e3a372da82d7f82eb270267c3202f63f7bd5091c8fcc\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}"
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.777502242Z" level=info msg="CreateContainer within sandbox \"0cbae525c76c173ef2998af3a8ce3a269a11200f75aa54529af4e9fbca1fff6f\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}"
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.799692864Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:etcd-scheduled-stop-721021,Uid:74087c1f265e54593e818a1bdf8a438e,Namespace:kube-system,Attempt:0,} returns sandbox id \"83516825275319eed1f6a93822cb304251cb57c085564f0f585842323146f129\""
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.807199165Z" level=info msg="CreateContainer within sandbox \"83516825275319eed1f6a93822cb304251cb57c085564f0f585842323146f129\" for container &ContainerMetadata{Name:etcd,Attempt:0,}"
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.816562508Z" level=info msg="CreateContainer within sandbox \"89f95609b8eb443f0901e3a372da82d7f82eb270267c3202f63f7bd5091c8fcc\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"7f65dc7e156e525ad795efba0af642bbf610cd38aee26487b558961f34772c64\""
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.817672109Z" level=info msg="StartContainer for \"7f65dc7e156e525ad795efba0af642bbf610cd38aee26487b558961f34772c64\""
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.828310077Z" level=info msg="CreateContainer within sandbox \"9fbee9d343450830d5713d13c05a89cde060d9086276e57f737495e2d6b2a4d3\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"add436cf75fd2c075f1b25505c80e85243d845bbea63e0222b78a4b4daee1b3b\""
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.829935681Z" level=info msg="StartContainer for \"add436cf75fd2c075f1b25505c80e85243d845bbea63e0222b78a4b4daee1b3b\""
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.830630976Z" level=info msg="CreateContainer within sandbox \"0cbae525c76c173ef2998af3a8ce3a269a11200f75aa54529af4e9fbca1fff6f\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"492199d8e7b9cb2db17623a583845e6a26ba5667b976507cddc428a361d769d3\""
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.831209100Z" level=info msg="StartContainer for \"492199d8e7b9cb2db17623a583845e6a26ba5667b976507cddc428a361d769d3\""
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.855774209Z" level=info msg="CreateContainer within sandbox \"83516825275319eed1f6a93822cb304251cb57c085564f0f585842323146f129\" for &ContainerMetadata{Name:etcd,Attempt:0,} returns container id \"5354ad5c007c4bc54ccabb9154422278141289fea933f48802a77ebbbcfa09cc\""
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.859039366Z" level=info msg="StartContainer for \"5354ad5c007c4bc54ccabb9154422278141289fea933f48802a77ebbbcfa09cc\""
Oct 17 19:32:08 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:08.922270577Z" level=info msg="StartContainer for \"7f65dc7e156e525ad795efba0af642bbf610cd38aee26487b558961f34772c64\" returns successfully"
Oct 17 19:32:09 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:09.003349086Z" level=info msg="StartContainer for \"5354ad5c007c4bc54ccabb9154422278141289fea933f48802a77ebbbcfa09cc\" returns successfully"
Oct 17 19:32:09 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:09.010224693Z" level=info msg="StartContainer for \"add436cf75fd2c075f1b25505c80e85243d845bbea63e0222b78a4b4daee1b3b\" returns successfully"
Oct 17 19:32:09 scheduled-stop-721021 containerd[764]: time="2025-10-17T19:32:09.127412431Z" level=info msg="StartContainer for \"492199d8e7b9cb2db17623a583845e6a26ba5667b976507cddc428a361d769d3\" returns successfully"
==> describe nodes <==
Name: scheduled-stop-721021
Roles: control-plane
Labels: beta.kubernetes.io/arch=arm64
beta.kubernetes.io/os=linux
kubernetes.io/arch=arm64
kubernetes.io/hostname=scheduled-stop-721021
kubernetes.io/os=linux
minikube.k8s.io/commit=73a80cc9bc99174c010556d98400e9fa16adda9d
minikube.k8s.io/name=scheduled-stop-721021
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2025_10_17T19_32_16_0700
minikube.k8s.io/version=v1.37.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
Annotations: volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Fri, 17 Oct 2025 19:32:13 +0000
Taints: node.kubernetes.io/not-ready:NoSchedule
Unschedulable: false
Lease:
HolderIdentity: scheduled-stop-721021
AcquireTime: <unset>
RenewTime: Fri, 17 Oct 2025 19:32:15 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Fri, 17 Oct 2025 19:32:15 +0000 Fri, 17 Oct 2025 19:32:09 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Fri, 17 Oct 2025 19:32:15 +0000 Fri, 17 Oct 2025 19:32:09 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Fri, 17 Oct 2025 19:32:15 +0000 Fri, 17 Oct 2025 19:32:09 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready False Fri, 17 Oct 2025 19:32:15 +0000 Fri, 17 Oct 2025 19:32:09 +0000 KubeletNotReady container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized
Addresses:
InternalIP: 192.168.76.2
Hostname: scheduled-stop-721021
Capacity:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022304Ki
pods: 110
Allocatable:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022304Ki
pods: 110
System Info:
Machine ID: c52191f5187031740f634bad68f0c727
System UUID: 1cd06f7e-d4a4-4a50-a963-703816927db7
Boot ID: 571a5863-a2dd-484d-8a17-263cc3da9adf
Kernel Version: 5.15.0-1084-aws
OS Image: Debian GNU/Linux 12 (bookworm)
Operating System: linux
Architecture: arm64
Container Runtime Version: containerd://1.7.28
Kubelet Version: v1.34.1
Kube-Proxy Version:
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (4 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
kube-system etcd-scheduled-stop-721021 100m (5%) 0 (0%) 100Mi (1%) 0 (0%) 4s
kube-system kube-apiserver-scheduled-stop-721021 250m (12%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-controller-manager-scheduled-stop-721021 200m (10%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-scheduler-scheduled-stop-721021 100m (5%) 0 (0%) 0 (0%) 0 (0%) 4s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 650m (32%) 0 (0%)
memory 100Mi (1%) 0 (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
hugepages-32Mi 0 (0%) 0 (0%)
hugepages-64Ki 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 4s kubelet Starting kubelet.
Warning CgroupV1 4s kubelet cgroup v1 support is in maintenance mode, please migrate to cgroup v2
Normal NodeAllocatableEnforced 4s kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 4s kubelet Node scheduled-stop-721021 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 4s kubelet Node scheduled-stop-721021 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 4s kubelet Node scheduled-stop-721021 status is now: NodeHasSufficientPID
==> dmesg <==
[Oct17 18:08] kauditd_printk_skb: 8 callbacks suppressed
[Oct17 18:56] kauditd_printk_skb: 8 callbacks suppressed
[Oct17 19:11] hrtimer: interrupt took 20156783 ns
==> etcd [5354ad5c007c4bc54ccabb9154422278141289fea933f48802a77ebbbcfa09cc] <==
{"level":"warn","ts":"2025-10-17T19:32:11.430018Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59374","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.445539Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59404","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.489084Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59414","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.491474Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59430","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.504129Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59442","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.526293Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59454","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.536500Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59470","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.560422Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59494","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.572957Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59514","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.617998Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59534","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.620073Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59544","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.639766Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59556","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.654096Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59568","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.678336Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59590","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.694047Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59600","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.711320Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59622","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.729122Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59654","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.744064Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59664","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.766702Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59670","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.779996Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59688","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.839721Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59714","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.864726Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59728","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.888051Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59738","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.921043Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59746","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-17T19:32:11.970119Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:59760","server-name":"","error":"EOF"}
==> kernel <==
19:32:20 up 2:14, 0 user, load average: 1.69, 1.85, 2.20
Linux scheduled-stop-721021 5.15.0-1084-aws #91~20.04.1-Ubuntu SMP Fri May 2 07:00:04 UTC 2025 aarch64 GNU/Linux
PRETTY_NAME="Debian GNU/Linux 12 (bookworm)"
==> kube-apiserver [492199d8e7b9cb2db17623a583845e6a26ba5667b976507cddc428a361d769d3] <==
I1017 19:32:13.069963 1 shared_informer.go:356] "Caches are synced" controller="node_authorizer"
E1017 19:32:13.072482 1 controller.go:145] "Failed to ensure lease exists, will retry" err="namespaces \"kube-system\" not found" interval="200ms"
I1017 19:32:13.076119 1 shared_informer.go:356] "Caches are synced" controller="*generic.policySource[*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicy,*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicyBinding,k8s.io/apiserver/pkg/admission/plugin/policy/validating.Validator]"
I1017 19:32:13.076212 1 policy_source.go:240] refreshing policies
E1017 19:32:13.094777 1 controller.go:148] "Unhandled Error" err="while syncing ConfigMap \"kube-system/kube-apiserver-legacy-service-account-token-tracking\", err: namespaces \"kube-system\" not found" logger="UnhandledError"
I1017 19:32:13.138227 1 controller.go:667] quota admission added evaluator for: namespaces
E1017 19:32:13.165117 1 controller.go:148] "Unhandled Error" err="while syncing ConfigMap \"kube-system/kube-apiserver-legacy-service-account-token-tracking\", err: namespaces \"kube-system\" not found" logger="UnhandledError"
I1017 19:32:13.183565 1 cidrallocator.go:301] created ClusterIP allocator for Service CIDR 10.96.0.0/12
I1017 19:32:13.183949 1 default_servicecidr_controller.go:228] Setting default ServiceCIDR condition Ready to True
I1017 19:32:13.231843 1 cidrallocator.go:277] updated ClusterIP allocator for Service CIDR 10.96.0.0/12
I1017 19:32:13.232186 1 default_servicecidr_controller.go:137] Shutting down kubernetes-service-cidr-controller
I1017 19:32:13.286621 1 controller.go:667] quota admission added evaluator for: leases.coordination.k8s.io
I1017 19:32:13.743180 1 storage_scheduling.go:95] created PriorityClass system-node-critical with value 2000001000
I1017 19:32:13.762318 1 storage_scheduling.go:95] created PriorityClass system-cluster-critical with value 2000000000
I1017 19:32:13.762569 1 storage_scheduling.go:111] all system priority classes are created successfully or already exist.
I1017 19:32:14.555882 1 controller.go:667] quota admission added evaluator for: roles.rbac.authorization.k8s.io
I1017 19:32:14.603726 1 controller.go:667] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
I1017 19:32:14.748478 1 alloc.go:328] "allocated clusterIPs" service="default/kubernetes" clusterIPs={"IPv4":"10.96.0.1"}
W1017 19:32:14.756369 1 lease.go:265] Resetting endpoints for master service "kubernetes" to [192.168.76.2]
I1017 19:32:14.757921 1 controller.go:667] quota admission added evaluator for: endpoints
I1017 19:32:14.765521 1 controller.go:667] quota admission added evaluator for: endpointslices.discovery.k8s.io
I1017 19:32:14.900291 1 controller.go:667] quota admission added evaluator for: serviceaccounts
I1017 19:32:15.686298 1 controller.go:667] quota admission added evaluator for: deployments.apps
I1017 19:32:15.702435 1 alloc.go:328] "allocated clusterIPs" service="kube-system/kube-dns" clusterIPs={"IPv4":"10.96.0.10"}
I1017 19:32:15.712551 1 controller.go:667] quota admission added evaluator for: daemonsets.apps
==> kube-controller-manager [7f65dc7e156e525ad795efba0af642bbf610cd38aee26487b558961f34772c64] <==
I1017 19:32:19.934234 1 shared_informer.go:356] "Caches are synced" controller="legacy-service-account-token-cleaner"
I1017 19:32:19.947259 1 shared_informer.go:356] "Caches are synced" controller="bootstrap_signer"
I1017 19:32:19.947601 1 shared_informer.go:356] "Caches are synced" controller="endpoint_slice"
I1017 19:32:19.947642 1 shared_informer.go:356] "Caches are synced" controller="taint"
I1017 19:32:19.947695 1 node_lifecycle_controller.go:1221] "Initializing eviction metric for zone" logger="node-lifecycle-controller" zone=""
I1017 19:32:19.947749 1 node_lifecycle_controller.go:873] "Missing timestamp for Node. Assuming now as a timestamp" logger="node-lifecycle-controller" node="scheduled-stop-721021"
I1017 19:32:19.947781 1 shared_informer.go:356] "Caches are synced" controller="resource_claim"
I1017 19:32:19.947859 1 shared_informer.go:356] "Caches are synced" controller="crt configmap"
I1017 19:32:19.947994 1 shared_informer.go:356] "Caches are synced" controller="ReplicaSet"
I1017 19:32:19.948009 1 shared_informer.go:356] "Caches are synced" controller="endpoint_slice_mirroring"
I1017 19:32:19.949115 1 node_lifecycle_controller.go:1025] "Controller detected that all Nodes are not-Ready. Entering master disruption mode" logger="node-lifecycle-controller"
I1017 19:32:19.949529 1 shared_informer.go:356] "Caches are synced" controller="garbage collector"
I1017 19:32:19.949550 1 shared_informer.go:356] "Caches are synced" controller="validatingadmissionpolicy-status"
I1017 19:32:19.949578 1 shared_informer.go:356] "Caches are synced" controller="certificate-csrsigning-kubelet-serving"
I1017 19:32:19.949614 1 shared_informer.go:356] "Caches are synced" controller="certificate-csrsigning-kubelet-client"
I1017 19:32:19.949643 1 shared_informer.go:356] "Caches are synced" controller="certificate-csrsigning-kube-apiserver-client"
I1017 19:32:19.949666 1 shared_informer.go:356] "Caches are synced" controller="cronjob"
I1017 19:32:19.949688 1 shared_informer.go:356] "Caches are synced" controller="certificate-csrsigning-legacy-unknown"
I1017 19:32:19.949713 1 shared_informer.go:356] "Caches are synced" controller="service account"
I1017 19:32:19.953971 1 shared_informer.go:356] "Caches are synced" controller="ephemeral"
I1017 19:32:19.955957 1 shared_informer.go:356] "Caches are synced" controller="namespace"
I1017 19:32:19.956457 1 shared_informer.go:356] "Caches are synced" controller="garbage collector"
I1017 19:32:19.956472 1 garbagecollector.go:154] "Garbage collector: all resource monitors have synced" logger="garbage-collector-controller"
I1017 19:32:19.956477 1 garbagecollector.go:157] "Proceeding to collect garbage" logger="garbage-collector-controller"
I1017 19:32:19.970556 1 shared_informer.go:356] "Caches are synced" controller="persistent volume"
==> kube-scheduler [add436cf75fd2c075f1b25505c80e85243d845bbea63e0222b78a4b4daee1b3b] <==
I1017 19:32:13.838263 1 server.go:177] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I1017 19:32:13.841204 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
I1017 19:32:13.842674 1 shared_informer.go:349] "Waiting for caches to sync" controller="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
I1017 19:32:13.843995 1 secure_serving.go:211] Serving securely on 127.0.0.1:10259
I1017 19:32:13.844190 1 tlsconfig.go:243] "Starting DynamicServingCertificateController"
E1017 19:32:13.854831 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError" reflector="runtime/asm_arm64.s:1223" type="*v1.ConfigMap"
E1017 19:32:13.855292 1 reflector.go:205] "Failed to watch" err="failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.StorageClass"
E1017 19:32:13.855542 1 reflector.go:205] "Failed to watch" err="failed to list *v1.VolumeAttachment: volumeattachments.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"volumeattachments\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.VolumeAttachment"
E1017 19:32:13.855802 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service"
E1017 19:32:13.863095 1 reflector.go:205] "Failed to watch" err="failed to list *v1.DeviceClass: deviceclasses.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"deviceclasses\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.DeviceClass"
E1017 19:32:13.863529 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PodDisruptionBudget"
E1017 19:32:13.863723 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ResourceClaim: resourceclaims.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"resourceclaims\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ResourceClaim"
E1017 19:32:13.863741 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicationController"
E1017 19:32:13.863793 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node"
E1017 19:32:13.864073 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ResourceSlice: resourceslices.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"resourceslices\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ResourceSlice"
E1017 19:32:13.864151 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicaSet"
E1017 19:32:13.864200 1 reflector.go:205] "Failed to watch" err="failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.StatefulSet"
E1017 19:32:13.864243 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIStorageCapacity"
E1017 19:32:13.864314 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod"
E1017 19:32:13.864438 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolumeClaim"
E1017 19:32:13.864598 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace"
E1017 19:32:13.864675 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver"
E1017 19:32:13.864711 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSINode"
E1017 19:32:13.864749 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolume"
I1017 19:32:15.243588 1 shared_informer.go:356] "Caches are synced" controller="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
==> kubelet <==
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.023161 1522 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8afd441a577fb9029dd8a39d1dbb3251-kubeconfig\") pod \"kube-controller-manager-scheduled-stop-721021\" (UID: \"8afd441a577fb9029dd8a39d1dbb3251\") " pod="kube-system/kube-controller-manager-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.023193 1522 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/261ae8e6a197ea01ba4a6cb85ea7fe96-kubeconfig\") pod \"kube-scheduler-scheduled-stop-721021\" (UID: \"261ae8e6a197ea01ba4a6cb85ea7fe96\") " pod="kube-system/kube-scheduler-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.023212 1522 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7e8f0033fca107fc141f37281e43bf2c-etc-ca-certificates\") pod \"kube-apiserver-scheduled-stop-721021\" (UID: \"7e8f0033fca107fc141f37281e43bf2c\") " pod="kube-system/kube-apiserver-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.023232 1522 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7e8f0033fca107fc141f37281e43bf2c-usr-local-share-ca-certificates\") pod \"kube-apiserver-scheduled-stop-721021\" (UID: \"7e8f0033fca107fc141f37281e43bf2c\") " pod="kube-system/kube-apiserver-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.023258 1522 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7e8f0033fca107fc141f37281e43bf2c-k8s-certs\") pod \"kube-apiserver-scheduled-stop-721021\" (UID: \"7e8f0033fca107fc141f37281e43bf2c\") " pod="kube-system/kube-apiserver-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.023280 1522 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7e8f0033fca107fc141f37281e43bf2c-usr-share-ca-certificates\") pod \"kube-apiserver-scheduled-stop-721021\" (UID: \"7e8f0033fca107fc141f37281e43bf2c\") " pod="kube-system/kube-apiserver-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.023299 1522 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/8afd441a577fb9029dd8a39d1dbb3251-flexvolume-dir\") pod \"kube-controller-manager-scheduled-stop-721021\" (UID: \"8afd441a577fb9029dd8a39d1dbb3251\") " pod="kube-system/kube-controller-manager-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.023318 1522 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8afd441a577fb9029dd8a39d1dbb3251-usr-local-share-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-721021\" (UID: \"8afd441a577fb9029dd8a39d1dbb3251\") " pod="kube-system/kube-controller-manager-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.023342 1522 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8afd441a577fb9029dd8a39d1dbb3251-usr-share-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-721021\" (UID: \"8afd441a577fb9029dd8a39d1dbb3251\") " pod="kube-system/kube-controller-manager-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.023372 1522 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-data\" (UniqueName: \"kubernetes.io/host-path/74087c1f265e54593e818a1bdf8a438e-etcd-data\") pod \"etcd-scheduled-stop-721021\" (UID: \"74087c1f265e54593e818a1bdf8a438e\") " pod="kube-system/etcd-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.023397 1522 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8afd441a577fb9029dd8a39d1dbb3251-ca-certs\") pod \"kube-controller-manager-scheduled-stop-721021\" (UID: \"8afd441a577fb9029dd8a39d1dbb3251\") " pod="kube-system/kube-controller-manager-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.600476 1522 apiserver.go:52] "Watching apiserver"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.621966 1522 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.733456 1522 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.733968 1522 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.734104 1522 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/etcd-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: E1017 19:32:16.752121 1522 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-scheduler-scheduled-stop-721021\" already exists" pod="kube-system/kube-scheduler-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: E1017 19:32:16.752609 1522 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"etcd-scheduled-stop-721021\" already exists" pod="kube-system/etcd-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: E1017 19:32:16.753940 1522 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-apiserver-scheduled-stop-721021\" already exists" pod="kube-system/kube-apiserver-scheduled-stop-721021"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.765207 1522 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-scheduled-stop-721021" podStartSLOduration=1.76518829 podStartE2EDuration="1.76518829s" podCreationTimestamp="2025-10-17 19:32:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-17 19:32:16.747735181 +0000 UTC m=+1.233758332" watchObservedRunningTime="2025-10-17 19:32:16.76518829 +0000 UTC m=+1.251211433"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.786621 1522 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-scheduled-stop-721021" podStartSLOduration=1.786407601 podStartE2EDuration="1.786407601s" podCreationTimestamp="2025-10-17 19:32:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-17 19:32:16.76616276 +0000 UTC m=+1.252185911" watchObservedRunningTime="2025-10-17 19:32:16.786407601 +0000 UTC m=+1.272430752"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.794625 1522 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/etcd-scheduled-stop-721021" podStartSLOduration=1.794603003 podStartE2EDuration="1.794603003s" podCreationTimestamp="2025-10-17 19:32:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-17 19:32:16.794374668 +0000 UTC m=+1.280397851" watchObservedRunningTime="2025-10-17 19:32:16.794603003 +0000 UTC m=+1.280626334"
Oct 17 19:32:16 scheduled-stop-721021 kubelet[1522]: I1017 19:32:16.825726 1522 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-scheduled-stop-721021" podStartSLOduration=1.825706204 podStartE2EDuration="1.825706204s" podCreationTimestamp="2025-10-17 19:32:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-17 19:32:16.818193141 +0000 UTC m=+1.304216292" watchObservedRunningTime="2025-10-17 19:32:16.825706204 +0000 UTC m=+1.311729396"
Oct 17 19:32:19 scheduled-stop-721021 kubelet[1522]: I1017 19:32:19.947080 1522 kuberuntime_manager.go:1828] "Updating runtime config through cri with podcidr" CIDR="10.244.0.0/24"
Oct 17 19:32:19 scheduled-stop-721021 kubelet[1522]: I1017 19:32:19.951615 1522 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="10.244.0.0/24"
-- /stdout --
helpers_test.go:262: (dbg) Run: out/minikube-linux-arm64 status --format={{.APIServer}} -p scheduled-stop-721021 -n scheduled-stop-721021
helpers_test.go:269: (dbg) Run: kubectl --context scheduled-stop-721021 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:280: non-running pods: kindnet-dx9vp kube-proxy-498h5 storage-provisioner
helpers_test.go:282: ======> post-mortem[TestScheduledStopUnix]: describe non-running pods <======
helpers_test.go:285: (dbg) Run: kubectl --context scheduled-stop-721021 describe pod kindnet-dx9vp kube-proxy-498h5 storage-provisioner
helpers_test.go:285: (dbg) Non-zero exit: kubectl --context scheduled-stop-721021 describe pod kindnet-dx9vp kube-proxy-498h5 storage-provisioner: exit status 1 (115.451822ms)
** stderr **
Error from server (NotFound): pods "kindnet-dx9vp" not found
Error from server (NotFound): pods "kube-proxy-498h5" not found
Error from server (NotFound): pods "storage-provisioner" not found
** /stderr **
helpers_test.go:287: kubectl --context scheduled-stop-721021 describe pod kindnet-dx9vp kube-proxy-498h5 storage-provisioner: exit status 1
helpers_test.go:175: Cleaning up "scheduled-stop-721021" profile ...
helpers_test.go:178: (dbg) Run: out/minikube-linux-arm64 delete -p scheduled-stop-721021
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p scheduled-stop-721021: (2.030650816s)
--- FAIL: TestScheduledStopUnix (41.56s)