=== RUN TestScheduledStopUnix
scheduled_stop_test.go:128: (dbg) Run: out/minikube-linux-arm64 start -p scheduled-stop-727712 --memory=3072 --driver=docker --container-runtime=containerd
scheduled_stop_test.go:128: (dbg) Done: out/minikube-linux-arm64 start -p scheduled-stop-727712 --memory=3072 --driver=docker --container-runtime=containerd: (31.352237281s)
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-727712 --schedule 5m
scheduled_stop_test.go:191: (dbg) Run: out/minikube-linux-arm64 status --format={{.TimeToStop}} -p scheduled-stop-727712 -n scheduled-stop-727712
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-727712 --schedule 15s
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:98: process 420762 running but should have been killed on reschedule of stop
panic.go:636: *** TestScheduledStopUnix FAILED at 2025-10-27 22:52:08.577502959 +0000 UTC m=+2245.522959878
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:223: ======> post-mortem[TestScheduledStopUnix]: network settings <======
helpers_test.go:230: HOST ENV snapshots: PROXY env: HTTP_PROXY="<empty>" HTTPS_PROXY="<empty>" NO_PROXY="<empty>"
helpers_test.go:238: ======> post-mortem[TestScheduledStopUnix]: docker inspect <======
helpers_test.go:239: (dbg) Run: docker inspect scheduled-stop-727712
helpers_test.go:243: (dbg) docker inspect scheduled-stop-727712:
-- stdout --
[
{
"Id": "26f4adc48ea9e1d0d2328b79342904f7d44dc4199732de8eb586b348570410db",
"Created": "2025-10-27T22:51:42.162840997Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 418781,
"ExitCode": 0,
"Error": "",
"StartedAt": "2025-10-27T22:51:42.250549814Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:1b8004df0b408966a254b2ecd4551aa85aaac4627e7e9cb1cefc14dfe51ec273",
"ResolvConfPath": "/var/lib/docker/containers/26f4adc48ea9e1d0d2328b79342904f7d44dc4199732de8eb586b348570410db/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/26f4adc48ea9e1d0d2328b79342904f7d44dc4199732de8eb586b348570410db/hostname",
"HostsPath": "/var/lib/docker/containers/26f4adc48ea9e1d0d2328b79342904f7d44dc4199732de8eb586b348570410db/hosts",
"LogPath": "/var/lib/docker/containers/26f4adc48ea9e1d0d2328b79342904f7d44dc4199732de8eb586b348570410db/26f4adc48ea9e1d0d2328b79342904f7d44dc4199732de8eb586b348570410db-json.log",
"Name": "/scheduled-stop-727712",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"scheduled-stop-727712:/var",
"/lib/modules:/lib/modules:ro"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {}
},
"NetworkMode": "scheduled-stop-727712",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 3221225472,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 6442450944,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"ID": "26f4adc48ea9e1d0d2328b79342904f7d44dc4199732de8eb586b348570410db",
"LowerDir": "/var/lib/docker/overlay2/b8348dfa9ed4e8656da24a2aaac9c6325141ec8ade63a22644060cbc120292fd-init/diff:/var/lib/docker/overlay2/71868b2d6b922761474a3006f56cee03abbe2c6fed1e66f903ecd8890c7d8e07/diff",
"MergedDir": "/var/lib/docker/overlay2/b8348dfa9ed4e8656da24a2aaac9c6325141ec8ade63a22644060cbc120292fd/merged",
"UpperDir": "/var/lib/docker/overlay2/b8348dfa9ed4e8656da24a2aaac9c6325141ec8ade63a22644060cbc120292fd/diff",
"WorkDir": "/var/lib/docker/overlay2/b8348dfa9ed4e8656da24a2aaac9c6325141ec8ade63a22644060cbc120292fd/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "volume",
"Name": "scheduled-stop-727712",
"Source": "/var/lib/docker/volumes/scheduled-stop-727712/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
},
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
}
],
"Config": {
"Hostname": "scheduled-stop-727712",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "scheduled-stop-727712",
"name.minikube.sigs.k8s.io": "scheduled-stop-727712",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "3c0d2aa65e191f3c6347049c4066147e7d16fb80f518a217ee64d23a3a2f664a",
"SandboxKey": "/var/run/docker/netns/3c0d2aa65e19",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33335"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33336"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33339"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33337"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33338"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"scheduled-stop-727712": {
"IPAMConfig": {
"IPv4Address": "192.168.76.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "82:22:b6:35:a3:5c",
"DriverOpts": null,
"GwPriority": 0,
"NetworkID": "c75eda9a6d12aa9214752d938b427744a6950652a79bd3d7d0a7f2b1f810837f",
"EndpointID": "842f200da634f460d7e43bbb76ae226b4c9979432837700a0fb10a4961dea540",
"Gateway": "192.168.76.1",
"IPAddress": "192.168.76.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"scheduled-stop-727712",
"26f4adc48ea9"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:247: (dbg) Run: out/minikube-linux-arm64 status --format={{.Host}} -p scheduled-stop-727712 -n scheduled-stop-727712
helpers_test.go:252: <<< TestScheduledStopUnix FAILED: start of post-mortem logs <<<
helpers_test.go:253: ======> post-mortem[TestScheduledStopUnix]: minikube logs <======
helpers_test.go:255: (dbg) Run: out/minikube-linux-arm64 -p scheduled-stop-727712 logs -n 25
helpers_test.go:255: (dbg) Done: out/minikube-linux-arm64 -p scheduled-stop-727712 logs -n 25: (1.178102477s)
helpers_test.go:260: TestScheduledStopUnix logs:
-- stdout --
==> Audit <==
┌─────────┬─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┬───────────────────────┬─────────┬─────────┬─────────────────────┬─────────────────────┐
│ COMMAND │ ARGS │ PROFILE │ USER │ VERSION │ START TIME │ END TIME │
├─────────┼─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┼───────────────────────┼─────────┼─────────┼─────────────────────┼─────────────────────┤
│ stop │ -p multinode-889386 │ multinode-889386 │ jenkins │ v1.37.0 │ 27 Oct 25 22:46 UTC │ 27 Oct 25 22:46 UTC │
│ start │ -p multinode-889386 --wait=true -v=5 --alsologtostderr │ multinode-889386 │ jenkins │ v1.37.0 │ 27 Oct 25 22:46 UTC │ 27 Oct 25 22:47 UTC │
│ node │ list -p multinode-889386 │ multinode-889386 │ jenkins │ v1.37.0 │ 27 Oct 25 22:47 UTC │ │
│ node │ multinode-889386 node delete m03 │ multinode-889386 │ jenkins │ v1.37.0 │ 27 Oct 25 22:47 UTC │ 27 Oct 25 22:47 UTC │
│ stop │ multinode-889386 stop │ multinode-889386 │ jenkins │ v1.37.0 │ 27 Oct 25 22:47 UTC │ 27 Oct 25 22:47 UTC │
│ start │ -p multinode-889386 --wait=true -v=5 --alsologtostderr --driver=docker --container-runtime=containerd │ multinode-889386 │ jenkins │ v1.37.0 │ 27 Oct 25 22:48 UTC │ 27 Oct 25 22:48 UTC │
│ node │ list -p multinode-889386 │ multinode-889386 │ jenkins │ v1.37.0 │ 27 Oct 25 22:48 UTC │ │
│ start │ -p multinode-889386-m02 --driver=docker --container-runtime=containerd │ multinode-889386-m02 │ jenkins │ v1.37.0 │ 27 Oct 25 22:48 UTC │ │
│ start │ -p multinode-889386-m03 --driver=docker --container-runtime=containerd │ multinode-889386-m03 │ jenkins │ v1.37.0 │ 27 Oct 25 22:48 UTC │ 27 Oct 25 22:49 UTC │
│ node │ add -p multinode-889386 │ multinode-889386 │ jenkins │ v1.37.0 │ 27 Oct 25 22:49 UTC │ │
│ delete │ -p multinode-889386-m03 │ multinode-889386-m03 │ jenkins │ v1.37.0 │ 27 Oct 25 22:49 UTC │ 27 Oct 25 22:49 UTC │
│ delete │ -p multinode-889386 │ multinode-889386 │ jenkins │ v1.37.0 │ 27 Oct 25 22:49 UTC │ 27 Oct 25 22:49 UTC │
│ start │ -p test-preload-289848 --memory=3072 --alsologtostderr --wait=true --preload=false --driver=docker --container-runtime=containerd --kubernetes-version=v1.32.0 │ test-preload-289848 │ jenkins │ v1.37.0 │ 27 Oct 25 22:49 UTC │ 27 Oct 25 22:50 UTC │
│ image │ test-preload-289848 image pull gcr.io/k8s-minikube/busybox │ test-preload-289848 │ jenkins │ v1.37.0 │ 27 Oct 25 22:50 UTC │ 27 Oct 25 22:50 UTC │
│ stop │ -p test-preload-289848 │ test-preload-289848 │ jenkins │ v1.37.0 │ 27 Oct 25 22:50 UTC │ 27 Oct 25 22:50 UTC │
│ start │ -p test-preload-289848 --memory=3072 --alsologtostderr -v=1 --wait=true --driver=docker --container-runtime=containerd │ test-preload-289848 │ jenkins │ v1.37.0 │ 27 Oct 25 22:50 UTC │ 27 Oct 25 22:51 UTC │
│ image │ test-preload-289848 image list │ test-preload-289848 │ jenkins │ v1.37.0 │ 27 Oct 25 22:51 UTC │ 27 Oct 25 22:51 UTC │
│ delete │ -p test-preload-289848 │ test-preload-289848 │ jenkins │ v1.37.0 │ 27 Oct 25 22:51 UTC │ 27 Oct 25 22:51 UTC │
│ start │ -p scheduled-stop-727712 --memory=3072 --driver=docker --container-runtime=containerd │ scheduled-stop-727712 │ jenkins │ v1.37.0 │ 27 Oct 25 22:51 UTC │ 27 Oct 25 22:52 UTC │
│ stop │ -p scheduled-stop-727712 --schedule 5m │ scheduled-stop-727712 │ jenkins │ v1.37.0 │ 27 Oct 25 22:52 UTC │ │
│ stop │ -p scheduled-stop-727712 --schedule 5m │ scheduled-stop-727712 │ jenkins │ v1.37.0 │ 27 Oct 25 22:52 UTC │ │
│ stop │ -p scheduled-stop-727712 --schedule 5m │ scheduled-stop-727712 │ jenkins │ v1.37.0 │ 27 Oct 25 22:52 UTC │ │
│ stop │ -p scheduled-stop-727712 --schedule 15s │ scheduled-stop-727712 │ jenkins │ v1.37.0 │ 27 Oct 25 22:52 UTC │ │
│ stop │ -p scheduled-stop-727712 --schedule 15s │ scheduled-stop-727712 │ jenkins │ v1.37.0 │ 27 Oct 25 22:52 UTC │ │
│ stop │ -p scheduled-stop-727712 --schedule 15s │ scheduled-stop-727712 │ jenkins │ v1.37.0 │ 27 Oct 25 22:52 UTC │ │
└─────────┴─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┴───────────────────────┴─────────┴─────────┴─────────────────────┴─────────────────────┘
==> Last Start <==
Log file created at: 2025/10/27 22:51:36
Running on machine: ip-172-31-24-2
Binary: Built with gc go1.24.6 for linux/arm64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I1027 22:51:36.745133 418392 out.go:360] Setting OutFile to fd 1 ...
I1027 22:51:36.745444 418392 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1027 22:51:36.745448 418392 out.go:374] Setting ErrFile to fd 2...
I1027 22:51:36.745452 418392 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1027 22:51:36.745750 418392 root.go:340] Updating PATH: /home/jenkins/minikube-integration/21790-269600/.minikube/bin
I1027 22:51:36.746171 418392 out.go:368] Setting JSON to false
I1027 22:51:36.747016 418392 start.go:133] hostinfo: {"hostname":"ip-172-31-24-2","uptime":9247,"bootTime":1761596250,"procs":156,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1084-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"6d436adf-771e-4269-b9a3-c25fd4fca4f5"}
I1027 22:51:36.747092 418392 start.go:143] virtualization:
I1027 22:51:36.750884 418392 out.go:179] * [scheduled-stop-727712] minikube v1.37.0 on Ubuntu 20.04 (arm64)
I1027 22:51:36.755320 418392 out.go:179] - MINIKUBE_LOCATION=21790
I1027 22:51:36.755396 418392 notify.go:221] Checking for updates...
I1027 22:51:36.762056 418392 out.go:179] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I1027 22:51:36.765171 418392 out.go:179] - KUBECONFIG=/home/jenkins/minikube-integration/21790-269600/kubeconfig
I1027 22:51:36.768529 418392 out.go:179] - MINIKUBE_HOME=/home/jenkins/minikube-integration/21790-269600/.minikube
I1027 22:51:36.771589 418392 out.go:179] - MINIKUBE_BIN=out/minikube-linux-arm64
I1027 22:51:36.774795 418392 out.go:179] - MINIKUBE_FORCE_SYSTEMD=
I1027 22:51:36.777915 418392 driver.go:422] Setting default libvirt URI to qemu:///system
I1027 22:51:36.809668 418392 docker.go:123] docker version: linux-28.1.1:Docker Engine - Community
I1027 22:51:36.809788 418392 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1027 22:51:36.867316 418392 info.go:266] docker info: {ID:J4M5:W6MX:GOX4:4LAQ:VI7E:VJNF:J3OP:OPBH:GF7G:PPY4:WQWD:7N4L Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:23 OomKillDisable:true NGoroutines:42 SystemTime:2025-10-27 22:51:36.858224211 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1084-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214831104 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-24-2 Labels:[] ExperimentalBuild:false ServerVersion:28.1.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:05044ec0a9a75232cad458027ca83437aae3f4da Expected:} RuncCommit:{ID:v1.2.5-0-g59923ef Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Pat
h:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.23.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.35.1]] Warnings:<nil>}}
I1027 22:51:36.867407 418392 docker.go:318] overlay module found
I1027 22:51:36.872614 418392 out.go:179] * Using the docker driver based on user configuration
I1027 22:51:36.875582 418392 start.go:307] selected driver: docker
I1027 22:51:36.875593 418392 start.go:928] validating driver "docker" against <nil>
I1027 22:51:36.875613 418392 start.go:939] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I1027 22:51:36.876328 418392 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1027 22:51:36.930638 418392 info.go:266] docker info: {ID:J4M5:W6MX:GOX4:4LAQ:VI7E:VJNF:J3OP:OPBH:GF7G:PPY4:WQWD:7N4L Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:23 OomKillDisable:true NGoroutines:42 SystemTime:2025-10-27 22:51:36.921817383 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1084-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214831104 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-24-2 Labels:[] ExperimentalBuild:false ServerVersion:28.1.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:05044ec0a9a75232cad458027ca83437aae3f4da Expected:} RuncCommit:{ID:v1.2.5-0-g59923ef Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Pat
h:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.23.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.35.1]] Warnings:<nil>}}
I1027 22:51:36.930786 418392 start_flags.go:326] no existing cluster config was found, will generate one from the flags
I1027 22:51:36.931031 418392 start_flags.go:973] Wait components to verify : map[apiserver:true system_pods:true]
I1027 22:51:36.934018 418392 out.go:179] * Using Docker driver with root privileges
I1027 22:51:36.936905 418392 cni.go:84] Creating CNI manager for ""
I1027 22:51:36.936973 418392 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I1027 22:51:36.936987 418392 start_flags.go:335] Found "CNI" CNI - setting NetworkPlugin=cni
I1027 22:51:36.937076 418392 start.go:351] cluster config:
{Name:scheduled-stop-727712 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-727712 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local
ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I1027 22:51:36.940227 418392 out.go:179] * Starting "scheduled-stop-727712" primary control-plane node in "scheduled-stop-727712" cluster
I1027 22:51:36.943090 418392 cache.go:124] Beginning downloading kic base image for docker with containerd
I1027 22:51:36.946082 418392 out.go:179] * Pulling base image v0.0.48-1760939008-21773 ...
I1027 22:51:36.949005 418392 preload.go:183] Checking if preload exists for k8s version v1.34.1 and runtime containerd
I1027 22:51:36.949079 418392 preload.go:198] Found local preload: /home/jenkins/minikube-integration/21790-269600/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-containerd-overlay2-arm64.tar.lz4
I1027 22:51:36.949088 418392 cache.go:59] Caching tarball of preloaded images
I1027 22:51:36.949094 418392 image.go:81] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 in local docker daemon
I1027 22:51:36.949183 418392 preload.go:233] Found /home/jenkins/minikube-integration/21790-269600/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-containerd-overlay2-arm64.tar.lz4 in cache, skipping download
I1027 22:51:36.949193 418392 cache.go:62] Finished verifying existence of preloaded tar for v1.34.1 on containerd
I1027 22:51:36.949526 418392 profile.go:143] Saving config to /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/config.json ...
I1027 22:51:36.949544 418392 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/config.json: {Name:mk8a2f8f715fa6845c4839e4a3e5384abd0495c7 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1027 22:51:36.968993 418392 image.go:100] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 in local docker daemon, skipping pull
I1027 22:51:36.969005 418392 cache.go:148] gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 exists in daemon, skipping load
I1027 22:51:36.969024 418392 cache.go:233] Successfully downloaded all kic artifacts
I1027 22:51:36.969046 418392 start.go:360] acquireMachinesLock for scheduled-stop-727712: {Name:mke1bc6c810737b084be64d6ad0575eea26ff494 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I1027 22:51:36.969157 418392 start.go:364] duration metric: took 97.306µs to acquireMachinesLock for "scheduled-stop-727712"
I1027 22:51:36.969181 418392 start.go:93] Provisioning new machine with config: &{Name:scheduled-stop-727712 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-727712 Namespace:default APIServerHAVIP: A
PIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: Static
IP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
I1027 22:51:36.969245 418392 start.go:125] createHost starting for "" (driver="docker")
I1027 22:51:36.974514 418392 out.go:252] * Creating docker container (CPUs=2, Memory=3072MB) ...
I1027 22:51:36.974754 418392 start.go:159] libmachine.API.Create for "scheduled-stop-727712" (driver="docker")
I1027 22:51:36.974788 418392 client.go:173] LocalClient.Create starting
I1027 22:51:36.974888 418392 main.go:143] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21790-269600/.minikube/certs/ca.pem
I1027 22:51:36.974927 418392 main.go:143] libmachine: Decoding PEM data...
I1027 22:51:36.974939 418392 main.go:143] libmachine: Parsing certificate...
I1027 22:51:36.975009 418392 main.go:143] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21790-269600/.minikube/certs/cert.pem
I1027 22:51:36.975029 418392 main.go:143] libmachine: Decoding PEM data...
I1027 22:51:36.975038 418392 main.go:143] libmachine: Parsing certificate...
I1027 22:51:36.975407 418392 cli_runner.go:164] Run: docker network inspect scheduled-stop-727712 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W1027 22:51:36.990876 418392 cli_runner.go:211] docker network inspect scheduled-stop-727712 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I1027 22:51:36.990980 418392 network_create.go:284] running [docker network inspect scheduled-stop-727712] to gather additional debugging logs...
I1027 22:51:36.990997 418392 cli_runner.go:164] Run: docker network inspect scheduled-stop-727712
W1027 22:51:37.009096 418392 cli_runner.go:211] docker network inspect scheduled-stop-727712 returned with exit code 1
I1027 22:51:37.009123 418392 network_create.go:287] error running [docker network inspect scheduled-stop-727712]: docker network inspect scheduled-stop-727712: exit status 1
stdout:
[]
stderr:
Error response from daemon: network scheduled-stop-727712 not found
I1027 22:51:37.009144 418392 network_create.go:289] output of [docker network inspect scheduled-stop-727712]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network scheduled-stop-727712 not found
** /stderr **
I1027 22:51:37.009271 418392 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I1027 22:51:37.028989 418392 network.go:211] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-743a90b7240a IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:ea:c0:8c:48:2b:2c} reservation:<nil>}
I1027 22:51:37.029275 418392 network.go:211] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-eb21287fdf9a IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:6a:a0:7e:f1:d9:d6} reservation:<nil>}
I1027 22:51:37.029471 418392 network.go:211] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-176a96de0236 IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:5a:24:d0:29:4c:9c} reservation:<nil>}
I1027 22:51:37.029821 418392 network.go:206] using free private subnet 192.168.76.0/24: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x40019fa0b0}
I1027 22:51:37.029837 418392 network_create.go:124] attempt to create docker network scheduled-stop-727712 192.168.76.0/24 with gateway 192.168.76.1 and MTU of 1500 ...
I1027 22:51:37.029898 418392 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.76.0/24 --gateway=192.168.76.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=scheduled-stop-727712 scheduled-stop-727712
I1027 22:51:37.087778 418392 network_create.go:108] docker network scheduled-stop-727712 192.168.76.0/24 created
I1027 22:51:37.087814 418392 kic.go:121] calculated static IP "192.168.76.2" for the "scheduled-stop-727712" container
I1027 22:51:37.087886 418392 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I1027 22:51:37.103631 418392 cli_runner.go:164] Run: docker volume create scheduled-stop-727712 --label name.minikube.sigs.k8s.io=scheduled-stop-727712 --label created_by.minikube.sigs.k8s.io=true
I1027 22:51:37.120734 418392 oci.go:103] Successfully created a docker volume scheduled-stop-727712
I1027 22:51:37.120891 418392 cli_runner.go:164] Run: docker run --rm --name scheduled-stop-727712-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-727712 --entrypoint /usr/bin/test -v scheduled-stop-727712:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 -d /var/lib
I1027 22:51:37.663990 418392 oci.go:107] Successfully prepared a docker volume scheduled-stop-727712
I1027 22:51:37.664037 418392 preload.go:183] Checking if preload exists for k8s version v1.34.1 and runtime containerd
I1027 22:51:37.664055 418392 kic.go:194] Starting extracting preloaded images to volume ...
I1027 22:51:37.664119 418392 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/21790-269600/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-727712:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 -I lz4 -xf /preloaded.tar -C /extractDir
I1027 22:51:42.070799 418392 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/21790-269600/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-727712:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 -I lz4 -xf /preloaded.tar -C /extractDir: (4.406633691s)
I1027 22:51:42.070822 418392 kic.go:203] duration metric: took 4.40676394s to extract preloaded images to volume ...
W1027 22:51:42.071027 418392 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I1027 22:51:42.071142 418392 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I1027 22:51:42.139867 418392 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname scheduled-stop-727712 --name scheduled-stop-727712 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-727712 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=scheduled-stop-727712 --network scheduled-stop-727712 --ip 192.168.76.2 --volume scheduled-stop-727712:/var --security-opt apparmor=unconfined --memory=3072mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8
I1027 22:51:42.486611 418392 cli_runner.go:164] Run: docker container inspect scheduled-stop-727712 --format={{.State.Running}}
I1027 22:51:42.513114 418392 cli_runner.go:164] Run: docker container inspect scheduled-stop-727712 --format={{.State.Status}}
I1027 22:51:42.536920 418392 cli_runner.go:164] Run: docker exec scheduled-stop-727712 stat /var/lib/dpkg/alternatives/iptables
I1027 22:51:42.592307 418392 oci.go:144] the created container "scheduled-stop-727712" has a running status.
I1027 22:51:42.592338 418392 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/21790-269600/.minikube/machines/scheduled-stop-727712/id_rsa...
I1027 22:51:43.091887 418392 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/21790-269600/.minikube/machines/scheduled-stop-727712/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I1027 22:51:43.120108 418392 cli_runner.go:164] Run: docker container inspect scheduled-stop-727712 --format={{.State.Status}}
I1027 22:51:43.149865 418392 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I1027 22:51:43.149881 418392 kic_runner.go:114] Args: [docker exec --privileged scheduled-stop-727712 chown docker:docker /home/docker/.ssh/authorized_keys]
I1027 22:51:43.213797 418392 cli_runner.go:164] Run: docker container inspect scheduled-stop-727712 --format={{.State.Status}}
I1027 22:51:43.236133 418392 machine.go:94] provisionDockerMachine start ...
I1027 22:51:43.236215 418392 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-727712
I1027 22:51:43.262558 418392 main.go:143] libmachine: Using SSH client type: native
I1027 22:51:43.262908 418392 main.go:143] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef110] 0x3f18d0 <nil> [] 0s} 127.0.0.1 33335 <nil> <nil>}
I1027 22:51:43.262922 418392 main.go:143] libmachine: About to run SSH command:
hostname
I1027 22:51:43.432725 418392 main.go:143] libmachine: SSH cmd err, output: <nil>: scheduled-stop-727712
I1027 22:51:43.432740 418392 ubuntu.go:182] provisioning hostname "scheduled-stop-727712"
I1027 22:51:43.432906 418392 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-727712
I1027 22:51:43.452521 418392 main.go:143] libmachine: Using SSH client type: native
I1027 22:51:43.452893 418392 main.go:143] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef110] 0x3f18d0 <nil> [] 0s} 127.0.0.1 33335 <nil> <nil>}
I1027 22:51:43.452903 418392 main.go:143] libmachine: About to run SSH command:
sudo hostname scheduled-stop-727712 && echo "scheduled-stop-727712" | sudo tee /etc/hostname
I1027 22:51:43.616025 418392 main.go:143] libmachine: SSH cmd err, output: <nil>: scheduled-stop-727712
I1027 22:51:43.616095 418392 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-727712
I1027 22:51:43.637377 418392 main.go:143] libmachine: Using SSH client type: native
I1027 22:51:43.637691 418392 main.go:143] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef110] 0x3f18d0 <nil> [] 0s} 127.0.0.1 33335 <nil> <nil>}
I1027 22:51:43.637706 418392 main.go:143] libmachine: About to run SSH command:
if ! grep -xq '.*\sscheduled-stop-727712' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 scheduled-stop-727712/g' /etc/hosts;
else
echo '127.0.1.1 scheduled-stop-727712' | sudo tee -a /etc/hosts;
fi
fi
I1027 22:51:43.800869 418392 main.go:143] libmachine: SSH cmd err, output: <nil>:
I1027 22:51:43.800886 418392 ubuntu.go:188] set auth options {CertDir:/home/jenkins/minikube-integration/21790-269600/.minikube CaCertPath:/home/jenkins/minikube-integration/21790-269600/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/21790-269600/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/21790-269600/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/21790-269600/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/21790-269600/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/21790-269600/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/21790-269600/.minikube}
I1027 22:51:43.800913 418392 ubuntu.go:190] setting up certificates
I1027 22:51:43.800921 418392 provision.go:84] configureAuth start
I1027 22:51:43.800986 418392 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-727712
I1027 22:51:43.818394 418392 provision.go:143] copyHostCerts
I1027 22:51:43.818453 418392 exec_runner.go:144] found /home/jenkins/minikube-integration/21790-269600/.minikube/ca.pem, removing ...
I1027 22:51:43.818461 418392 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21790-269600/.minikube/ca.pem
I1027 22:51:43.818537 418392 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21790-269600/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/21790-269600/.minikube/ca.pem (1078 bytes)
I1027 22:51:43.818620 418392 exec_runner.go:144] found /home/jenkins/minikube-integration/21790-269600/.minikube/cert.pem, removing ...
I1027 22:51:43.818624 418392 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21790-269600/.minikube/cert.pem
I1027 22:51:43.818652 418392 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21790-269600/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/21790-269600/.minikube/cert.pem (1123 bytes)
I1027 22:51:43.818699 418392 exec_runner.go:144] found /home/jenkins/minikube-integration/21790-269600/.minikube/key.pem, removing ...
I1027 22:51:43.818702 418392 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21790-269600/.minikube/key.pem
I1027 22:51:43.818724 418392 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21790-269600/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/21790-269600/.minikube/key.pem (1679 bytes)
I1027 22:51:43.818765 418392 provision.go:117] generating server cert: /home/jenkins/minikube-integration/21790-269600/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/21790-269600/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/21790-269600/.minikube/certs/ca-key.pem org=jenkins.scheduled-stop-727712 san=[127.0.0.1 192.168.76.2 localhost minikube scheduled-stop-727712]
I1027 22:51:44.583695 418392 provision.go:177] copyRemoteCerts
I1027 22:51:44.583755 418392 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I1027 22:51:44.583794 418392 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-727712
I1027 22:51:44.601212 418392 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33335 SSHKeyPath:/home/jenkins/minikube-integration/21790-269600/.minikube/machines/scheduled-stop-727712/id_rsa Username:docker}
I1027 22:51:44.704983 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I1027 22:51:44.722397 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/machines/server.pem --> /etc/docker/server.pem (1229 bytes)
I1027 22:51:44.739802 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I1027 22:51:44.758321 418392 provision.go:87] duration metric: took 957.370596ms to configureAuth
I1027 22:51:44.758339 418392 ubuntu.go:206] setting minikube options for container-runtime
I1027 22:51:44.758537 418392 config.go:182] Loaded profile config "scheduled-stop-727712": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.34.1
I1027 22:51:44.758543 418392 machine.go:97] duration metric: took 1.522400211s to provisionDockerMachine
I1027 22:51:44.758548 418392 client.go:176] duration metric: took 7.783755038s to LocalClient.Create
I1027 22:51:44.758560 418392 start.go:167] duration metric: took 7.783808996s to libmachine.API.Create "scheduled-stop-727712"
I1027 22:51:44.758565 418392 start.go:293] postStartSetup for "scheduled-stop-727712" (driver="docker")
I1027 22:51:44.758574 418392 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I1027 22:51:44.758620 418392 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I1027 22:51:44.758663 418392 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-727712
I1027 22:51:44.775447 418392 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33335 SSHKeyPath:/home/jenkins/minikube-integration/21790-269600/.minikube/machines/scheduled-stop-727712/id_rsa Username:docker}
I1027 22:51:44.881684 418392 ssh_runner.go:195] Run: cat /etc/os-release
I1027 22:51:44.884801 418392 main.go:143] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I1027 22:51:44.884820 418392 info.go:137] Remote host: Debian GNU/Linux 12 (bookworm)
I1027 22:51:44.884829 418392 filesync.go:126] Scanning /home/jenkins/minikube-integration/21790-269600/.minikube/addons for local assets ...
I1027 22:51:44.884884 418392 filesync.go:126] Scanning /home/jenkins/minikube-integration/21790-269600/.minikube/files for local assets ...
I1027 22:51:44.884966 418392 filesync.go:149] local asset: /home/jenkins/minikube-integration/21790-269600/.minikube/files/etc/ssl/certs/2714482.pem -> 2714482.pem in /etc/ssl/certs
I1027 22:51:44.885071 418392 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I1027 22:51:44.892228 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/files/etc/ssl/certs/2714482.pem --> /etc/ssl/certs/2714482.pem (1708 bytes)
I1027 22:51:44.908898 418392 start.go:296] duration metric: took 150.319381ms for postStartSetup
I1027 22:51:44.909264 418392 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-727712
I1027 22:51:44.926000 418392 profile.go:143] Saving config to /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/config.json ...
I1027 22:51:44.926284 418392 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I1027 22:51:44.926323 418392 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-727712
I1027 22:51:44.943017 418392 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33335 SSHKeyPath:/home/jenkins/minikube-integration/21790-269600/.minikube/machines/scheduled-stop-727712/id_rsa Username:docker}
I1027 22:51:45.066412 418392 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I1027 22:51:45.075235 418392 start.go:128] duration metric: took 8.105959431s to createHost
I1027 22:51:45.075253 418392 start.go:83] releasing machines lock for "scheduled-stop-727712", held for 8.106089253s
I1027 22:51:45.075356 418392 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-727712
I1027 22:51:45.104017 418392 ssh_runner.go:195] Run: cat /version.json
I1027 22:51:45.104062 418392 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-727712
I1027 22:51:45.104329 418392 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I1027 22:51:45.104384 418392 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-727712
I1027 22:51:45.148890 418392 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33335 SSHKeyPath:/home/jenkins/minikube-integration/21790-269600/.minikube/machines/scheduled-stop-727712/id_rsa Username:docker}
I1027 22:51:45.152885 418392 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33335 SSHKeyPath:/home/jenkins/minikube-integration/21790-269600/.minikube/machines/scheduled-stop-727712/id_rsa Username:docker}
I1027 22:51:45.388001 418392 ssh_runner.go:195] Run: systemctl --version
I1027 22:51:45.396994 418392 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
W1027 22:51:45.402368 418392 cni.go:209] loopback cni configuration skipped: "/etc/cni/net.d/*loopback.conf*" not found
I1027 22:51:45.402440 418392 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I1027 22:51:45.446508 418392 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/10-crio-bridge.conflist.disabled] bridge cni config(s)
I1027 22:51:45.446528 418392 start.go:496] detecting cgroup driver to use...
I1027 22:51:45.446610 418392 detect.go:187] detected "cgroupfs" cgroup driver on host os
I1027 22:51:45.446734 418392 ssh_runner.go:195] Run: sudo systemctl stop -f crio
I1027 22:51:45.470301 418392 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I1027 22:51:45.485477 418392 docker.go:218] disabling cri-docker service (if available) ...
I1027 22:51:45.485535 418392 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
I1027 22:51:45.507248 418392 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
I1027 22:51:45.526883 418392 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
I1027 22:51:45.649095 418392 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
I1027 22:51:45.774675 418392 docker.go:234] disabling docker service ...
I1027 22:51:45.774733 418392 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
I1027 22:51:45.797144 418392 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
I1027 22:51:45.810708 418392 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
I1027 22:51:45.943079 418392 ssh_runner.go:195] Run: sudo systemctl mask docker.service
I1027 22:51:46.065714 418392 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
I1027 22:51:46.079852 418392 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I1027 22:51:46.094965 418392 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10.1"|' /etc/containerd/config.toml"
I1027 22:51:46.104281 418392 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I1027 22:51:46.113671 418392 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I1027 22:51:46.113732 418392 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I1027 22:51:46.122986 418392 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I1027 22:51:46.132076 418392 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I1027 22:51:46.141345 418392 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I1027 22:51:46.150232 418392 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I1027 22:51:46.158783 418392 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I1027 22:51:46.167609 418392 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I1027 22:51:46.176686 418392 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I1027 22:51:46.186807 418392 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I1027 22:51:46.195021 418392 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I1027 22:51:46.202581 418392 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1027 22:51:46.319799 418392 ssh_runner.go:195] Run: sudo systemctl restart containerd
I1027 22:51:46.465441 418392 start.go:543] Will wait 60s for socket path /run/containerd/containerd.sock
I1027 22:51:46.465519 418392 ssh_runner.go:195] Run: stat /run/containerd/containerd.sock
I1027 22:51:46.469644 418392 start.go:564] Will wait 60s for crictl version
I1027 22:51:46.469703 418392 ssh_runner.go:195] Run: which crictl
I1027 22:51:46.473443 418392 ssh_runner.go:195] Run: sudo /usr/local/bin/crictl version
I1027 22:51:46.508097 418392 start.go:580] Version: 0.1.0
RuntimeName: containerd
RuntimeVersion: v1.7.28
RuntimeApiVersion: v1
I1027 22:51:46.508164 418392 ssh_runner.go:195] Run: containerd --version
I1027 22:51:46.530755 418392 ssh_runner.go:195] Run: containerd --version
I1027 22:51:46.562672 418392 out.go:179] * Preparing Kubernetes v1.34.1 on containerd 1.7.28 ...
I1027 22:51:46.565539 418392 cli_runner.go:164] Run: docker network inspect scheduled-stop-727712 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I1027 22:51:46.581324 418392 ssh_runner.go:195] Run: grep 192.168.76.1 host.minikube.internal$ /etc/hosts
I1027 22:51:46.585002 418392 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.76.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I1027 22:51:46.594581 418392 kubeadm.go:884] updating cluster {Name:scheduled-stop-727712 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-727712 Namespace:default APIServerHAVIP: APIServerName:miniku
beCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSH
AuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I1027 22:51:46.594688 418392 preload.go:183] Checking if preload exists for k8s version v1.34.1 and runtime containerd
I1027 22:51:46.594743 418392 ssh_runner.go:195] Run: sudo crictl images --output json
I1027 22:51:46.619379 418392 containerd.go:627] all images are preloaded for containerd runtime.
I1027 22:51:46.619392 418392 containerd.go:534] Images already preloaded, skipping extraction
I1027 22:51:46.619451 418392 ssh_runner.go:195] Run: sudo crictl images --output json
I1027 22:51:46.644016 418392 containerd.go:627] all images are preloaded for containerd runtime.
I1027 22:51:46.644028 418392 cache_images.go:86] Images are preloaded, skipping loading
I1027 22:51:46.644034 418392 kubeadm.go:935] updating node { 192.168.76.2 8443 v1.34.1 containerd true true} ...
I1027 22:51:46.644170 418392 kubeadm.go:947] kubelet [Unit]
Wants=containerd.service
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.34.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=scheduled-stop-727712 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.76.2
[Install]
config:
{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-727712 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I1027 22:51:46.644233 418392 ssh_runner.go:195] Run: sudo crictl info
I1027 22:51:46.673899 418392 cni.go:84] Creating CNI manager for ""
I1027 22:51:46.673909 418392 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I1027 22:51:46.673934 418392 kubeadm.go:85] Using pod CIDR: 10.244.0.0/16
I1027 22:51:46.673962 418392 kubeadm.go:190] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.76.2 APIServerPort:8443 KubernetesVersion:v1.34.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:scheduled-stop-727712 NodeName:scheduled-stop-727712 DNSDomain:cluster.local CRISocket:/run/containerd/containerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.76.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.76.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt St
aticPodPath:/etc/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///run/containerd/containerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I1027 22:51:46.674075 418392 kubeadm.go:196] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta4
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.76.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///run/containerd/containerd.sock
name: "scheduled-stop-727712"
kubeletExtraArgs:
- name: "node-ip"
value: "192.168.76.2"
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta4
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.76.2"]
extraArgs:
- name: "enable-admission-plugins"
value: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
- name: "allocate-node-cidrs"
value: "true"
- name: "leader-elect"
value: "false"
scheduler:
extraArgs:
- name: "leader-elect"
value: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
kubernetesVersion: v1.34.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///run/containerd/containerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I1027 22:51:46.674138 418392 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.34.1
I1027 22:51:46.682047 418392 binaries.go:44] Found k8s binaries, skipping transfer
I1027 22:51:46.682108 418392 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I1027 22:51:46.690131 418392 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (325 bytes)
I1027 22:51:46.703604 418392 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I1027 22:51:46.716769 418392 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2234 bytes)
I1027 22:51:46.729941 418392 ssh_runner.go:195] Run: grep 192.168.76.2 control-plane.minikube.internal$ /etc/hosts
I1027 22:51:46.733723 418392 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.76.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I1027 22:51:46.743793 418392 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1027 22:51:46.861840 418392 ssh_runner.go:195] Run: sudo systemctl start kubelet
I1027 22:51:46.877338 418392 certs.go:69] Setting up /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712 for IP: 192.168.76.2
I1027 22:51:46.877350 418392 certs.go:195] generating shared ca certs ...
I1027 22:51:46.877364 418392 certs.go:227] acquiring lock for ca certs: {Name:mk23112b7c069e590ec7058965e0532af7da3447 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1027 22:51:46.877490 418392 certs.go:236] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/21790-269600/.minikube/ca.key
I1027 22:51:46.877538 418392 certs.go:236] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/21790-269600/.minikube/proxy-client-ca.key
I1027 22:51:46.877543 418392 certs.go:257] generating profile certs ...
I1027 22:51:46.877600 418392 certs.go:364] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/client.key
I1027 22:51:46.877609 418392 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/client.crt with IP's: []
I1027 22:51:47.139642 418392 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/client.crt ...
I1027 22:51:47.139658 418392 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/client.crt: {Name:mk02b1fe44fe7d263d295e27b35c25679f2129b1 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1027 22:51:47.139864 418392 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/client.key ...
I1027 22:51:47.139875 418392 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/client.key: {Name:mk12d2db0690fcc29a3710dc2a09afac24477f49 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1027 22:51:47.139966 418392 certs.go:364] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/apiserver.key.c5fdf043
I1027 22:51:47.139992 418392 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/apiserver.crt.c5fdf043 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.76.2]
I1027 22:51:47.404290 418392 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/apiserver.crt.c5fdf043 ...
I1027 22:51:47.404311 418392 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/apiserver.crt.c5fdf043: {Name:mkcfcdd668803596f4fe7cf11d4d8db113b79b40 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1027 22:51:47.404492 418392 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/apiserver.key.c5fdf043 ...
I1027 22:51:47.404501 418392 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/apiserver.key.c5fdf043: {Name:mk5146e5bf103239dfe7fcafd830aba4e07f8279 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1027 22:51:47.404585 418392 certs.go:382] copying /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/apiserver.crt.c5fdf043 -> /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/apiserver.crt
I1027 22:51:47.404666 418392 certs.go:386] copying /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/apiserver.key.c5fdf043 -> /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/apiserver.key
I1027 22:51:47.404718 418392 certs.go:364] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/proxy-client.key
I1027 22:51:47.404731 418392 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/proxy-client.crt with IP's: []
I1027 22:51:47.765438 418392 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/proxy-client.crt ...
I1027 22:51:47.765456 418392 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/proxy-client.crt: {Name:mk3110cc948bc97ba66f86030060c484290252dc Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1027 22:51:47.765698 418392 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/proxy-client.key ...
I1027 22:51:47.765707 418392 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/proxy-client.key: {Name:mka437a15c8009921bf85405f8bfba897937a727 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1027 22:51:47.765918 418392 certs.go:484] found cert: /home/jenkins/minikube-integration/21790-269600/.minikube/certs/271448.pem (1338 bytes)
W1027 22:51:47.765959 418392 certs.go:480] ignoring /home/jenkins/minikube-integration/21790-269600/.minikube/certs/271448_empty.pem, impossibly tiny 0 bytes
I1027 22:51:47.765966 418392 certs.go:484] found cert: /home/jenkins/minikube-integration/21790-269600/.minikube/certs/ca-key.pem (1679 bytes)
I1027 22:51:47.765989 418392 certs.go:484] found cert: /home/jenkins/minikube-integration/21790-269600/.minikube/certs/ca.pem (1078 bytes)
I1027 22:51:47.766011 418392 certs.go:484] found cert: /home/jenkins/minikube-integration/21790-269600/.minikube/certs/cert.pem (1123 bytes)
I1027 22:51:47.766034 418392 certs.go:484] found cert: /home/jenkins/minikube-integration/21790-269600/.minikube/certs/key.pem (1679 bytes)
I1027 22:51:47.766076 418392 certs.go:484] found cert: /home/jenkins/minikube-integration/21790-269600/.minikube/files/etc/ssl/certs/2714482.pem (1708 bytes)
I1027 22:51:47.766764 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I1027 22:51:47.787636 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I1027 22:51:47.806074 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I1027 22:51:47.825368 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I1027 22:51:47.843714 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1432 bytes)
I1027 22:51:47.862718 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
I1027 22:51:47.881965 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I1027 22:51:47.900056 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/profiles/scheduled-stop-727712/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I1027 22:51:47.917613 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/certs/271448.pem --> /usr/share/ca-certificates/271448.pem (1338 bytes)
I1027 22:51:47.935806 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/files/etc/ssl/certs/2714482.pem --> /usr/share/ca-certificates/2714482.pem (1708 bytes)
I1027 22:51:47.953959 418392 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21790-269600/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I1027 22:51:47.971871 418392 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I1027 22:51:47.985415 418392 ssh_runner.go:195] Run: openssl version
I1027 22:51:47.991668 418392 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/271448.pem && ln -fs /usr/share/ca-certificates/271448.pem /etc/ssl/certs/271448.pem"
I1027 22:51:48.000184 418392 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/271448.pem
I1027 22:51:48.004755 418392 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 Oct 27 22:22 /usr/share/ca-certificates/271448.pem
I1027 22:51:48.004848 418392 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/271448.pem
I1027 22:51:48.051298 418392 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/271448.pem /etc/ssl/certs/51391683.0"
I1027 22:51:48.060362 418392 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/2714482.pem && ln -fs /usr/share/ca-certificates/2714482.pem /etc/ssl/certs/2714482.pem"
I1027 22:51:48.069230 418392 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/2714482.pem
I1027 22:51:48.073298 418392 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 Oct 27 22:22 /usr/share/ca-certificates/2714482.pem
I1027 22:51:48.073380 418392 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/2714482.pem
I1027 22:51:48.115244 418392 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/2714482.pem /etc/ssl/certs/3ec20f2e.0"
I1027 22:51:48.123910 418392 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I1027 22:51:48.132180 418392 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I1027 22:51:48.136162 418392 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Oct 27 22:15 /usr/share/ca-certificates/minikubeCA.pem
I1027 22:51:48.136222 418392 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I1027 22:51:48.177325 418392 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I1027 22:51:48.185550 418392 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I1027 22:51:48.188995 418392 certs.go:400] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I1027 22:51:48.189053 418392 kubeadm.go:401] StartCluster: {Name:scheduled-stop-727712 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-727712 Namespace:default APIServerHAVIP: APIServerName:minikubeC
A APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAut
hSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I1027 22:51:48.189128 418392 cri.go:54] listing CRI containers in root /run/containerd/runc/k8s.io: {State:paused Name: Namespaces:[kube-system]}
I1027 22:51:48.189192 418392 ssh_runner.go:195] Run: sudo -s eval "crictl ps -a --quiet --label io.kubernetes.pod.namespace=kube-system"
I1027 22:51:48.215444 418392 cri.go:89] found id: ""
I1027 22:51:48.215515 418392 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I1027 22:51:48.223638 418392 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I1027 22:51:48.231357 418392 kubeadm.go:215] ignoring SystemVerification for kubeadm because of docker driver
I1027 22:51:48.231412 418392 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I1027 22:51:48.242543 418392 kubeadm.go:156] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I1027 22:51:48.242553 418392 kubeadm.go:158] found existing configuration files:
I1027 22:51:48.242605 418392 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I1027 22:51:48.250414 418392 kubeadm.go:164] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I1027 22:51:48.250469 418392 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I1027 22:51:48.258145 418392 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I1027 22:51:48.265631 418392 kubeadm.go:164] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I1027 22:51:48.265684 418392 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I1027 22:51:48.272858 418392 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I1027 22:51:48.281561 418392 kubeadm.go:164] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I1027 22:51:48.281614 418392 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I1027 22:51:48.288716 418392 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I1027 22:51:48.296311 418392 kubeadm.go:164] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I1027 22:51:48.296374 418392 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I1027 22:51:48.303828 418392 ssh_runner.go:286] Start: sudo /bin/bash -c "env PATH="/var/lib/minikube/binaries/v1.34.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I1027 22:51:48.342700 418392 kubeadm.go:319] [init] Using Kubernetes version: v1.34.1
I1027 22:51:48.342857 418392 kubeadm.go:319] [preflight] Running pre-flight checks
I1027 22:51:48.367738 418392 kubeadm.go:319] [preflight] The system verification failed. Printing the output from the verification:
I1027 22:51:48.367801 418392 kubeadm.go:319] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1084-aws[0m
I1027 22:51:48.367834 418392 kubeadm.go:319] [0;37mOS[0m: [0;32mLinux[0m
I1027 22:51:48.367879 418392 kubeadm.go:319] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I1027 22:51:48.367926 418392 kubeadm.go:319] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I1027 22:51:48.367973 418392 kubeadm.go:319] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I1027 22:51:48.368020 418392 kubeadm.go:319] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I1027 22:51:48.368067 418392 kubeadm.go:319] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I1027 22:51:48.368114 418392 kubeadm.go:319] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I1027 22:51:48.368159 418392 kubeadm.go:319] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I1027 22:51:48.368206 418392 kubeadm.go:319] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I1027 22:51:48.368251 418392 kubeadm.go:319] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I1027 22:51:48.442152 418392 kubeadm.go:319] [preflight] Pulling images required for setting up a Kubernetes cluster
I1027 22:51:48.442260 418392 kubeadm.go:319] [preflight] This might take a minute or two, depending on the speed of your internet connection
I1027 22:51:48.442355 418392 kubeadm.go:319] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I1027 22:51:48.453175 418392 kubeadm.go:319] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I1027 22:51:48.458405 418392 out.go:252] - Generating certificates and keys ...
I1027 22:51:48.458498 418392 kubeadm.go:319] [certs] Using existing ca certificate authority
I1027 22:51:48.458566 418392 kubeadm.go:319] [certs] Using existing apiserver certificate and key on disk
I1027 22:51:49.363566 418392 kubeadm.go:319] [certs] Generating "apiserver-kubelet-client" certificate and key
I1027 22:51:50.243636 418392 kubeadm.go:319] [certs] Generating "front-proxy-ca" certificate and key
I1027 22:51:50.425142 418392 kubeadm.go:319] [certs] Generating "front-proxy-client" certificate and key
I1027 22:51:51.022010 418392 kubeadm.go:319] [certs] Generating "etcd/ca" certificate and key
I1027 22:51:51.479005 418392 kubeadm.go:319] [certs] Generating "etcd/server" certificate and key
I1027 22:51:51.479311 418392 kubeadm.go:319] [certs] etcd/server serving cert is signed for DNS names [localhost scheduled-stop-727712] and IPs [192.168.76.2 127.0.0.1 ::1]
I1027 22:51:51.969067 418392 kubeadm.go:319] [certs] Generating "etcd/peer" certificate and key
I1027 22:51:51.969410 418392 kubeadm.go:319] [certs] etcd/peer serving cert is signed for DNS names [localhost scheduled-stop-727712] and IPs [192.168.76.2 127.0.0.1 ::1]
I1027 22:51:52.212683 418392 kubeadm.go:319] [certs] Generating "etcd/healthcheck-client" certificate and key
I1027 22:51:52.316722 418392 kubeadm.go:319] [certs] Generating "apiserver-etcd-client" certificate and key
I1027 22:51:53.189804 418392 kubeadm.go:319] [certs] Generating "sa" key and public key
I1027 22:51:53.190032 418392 kubeadm.go:319] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I1027 22:51:53.931633 418392 kubeadm.go:319] [kubeconfig] Writing "admin.conf" kubeconfig file
I1027 22:51:55.000538 418392 kubeadm.go:319] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I1027 22:51:55.390938 418392 kubeadm.go:319] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I1027 22:51:55.777247 418392 kubeadm.go:319] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I1027 22:51:56.376030 418392 kubeadm.go:319] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I1027 22:51:56.376802 418392 kubeadm.go:319] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I1027 22:51:56.379521 418392 kubeadm.go:319] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I1027 22:51:56.383188 418392 out.go:252] - Booting up control plane ...
I1027 22:51:56.383289 418392 kubeadm.go:319] [control-plane] Creating static Pod manifest for "kube-apiserver"
I1027 22:51:56.383370 418392 kubeadm.go:319] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I1027 22:51:56.383437 418392 kubeadm.go:319] [control-plane] Creating static Pod manifest for "kube-scheduler"
I1027 22:51:56.399123 418392 kubeadm.go:319] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I1027 22:51:56.399479 418392 kubeadm.go:319] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/instance-config.yaml"
I1027 22:51:56.406750 418392 kubeadm.go:319] [patches] Applied patch of type "application/strategic-merge-patch+json" to target "kubeletconfiguration"
I1027 22:51:56.407220 418392 kubeadm.go:319] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I1027 22:51:56.407447 418392 kubeadm.go:319] [kubelet-start] Starting the kubelet
I1027 22:51:56.537448 418392 kubeadm.go:319] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I1027 22:51:56.537574 418392 kubeadm.go:319] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I1027 22:51:58.033397 418392 kubeadm.go:319] [kubelet-check] The kubelet is healthy after 1.500990827s
I1027 22:51:58.037110 418392 kubeadm.go:319] [control-plane-check] Waiting for healthy control plane components. This can take up to 4m0s
I1027 22:51:58.037196 418392 kubeadm.go:319] [control-plane-check] Checking kube-apiserver at https://192.168.76.2:8443/livez
I1027 22:51:58.037284 418392 kubeadm.go:319] [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz
I1027 22:51:58.037543 418392 kubeadm.go:319] [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez
I1027 22:52:00.375579 418392 kubeadm.go:319] [control-plane-check] kube-controller-manager is healthy after 2.337901548s
I1027 22:52:04.543669 418392 kubeadm.go:319] [control-plane-check] kube-apiserver is healthy after 6.506508301s
I1027 22:52:04.785138 418392 kubeadm.go:319] [control-plane-check] kube-scheduler is healthy after 6.746451592s
I1027 22:52:04.823849 418392 kubeadm.go:319] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I1027 22:52:04.839863 418392 kubeadm.go:319] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I1027 22:52:04.857746 418392 kubeadm.go:319] [upload-certs] Skipping phase. Please see --upload-certs
I1027 22:52:04.858176 418392 kubeadm.go:319] [mark-control-plane] Marking the node scheduled-stop-727712 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I1027 22:52:04.872653 418392 kubeadm.go:319] [bootstrap-token] Using token: x3pj4l.awtike51qlcy8cjp
I1027 22:52:04.875591 418392 out.go:252] - Configuring RBAC rules ...
I1027 22:52:04.875710 418392 kubeadm.go:319] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I1027 22:52:04.882516 418392 kubeadm.go:319] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I1027 22:52:04.895074 418392 kubeadm.go:319] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I1027 22:52:04.901536 418392 kubeadm.go:319] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I1027 22:52:04.907372 418392 kubeadm.go:319] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I1027 22:52:04.911786 418392 kubeadm.go:319] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I1027 22:52:05.193240 418392 kubeadm.go:319] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I1027 22:52:05.626626 418392 kubeadm.go:319] [addons] Applied essential addon: CoreDNS
I1027 22:52:06.192194 418392 kubeadm.go:319] [addons] Applied essential addon: kube-proxy
I1027 22:52:06.193272 418392 kubeadm.go:319]
I1027 22:52:06.193351 418392 kubeadm.go:319] Your Kubernetes control-plane has initialized successfully!
I1027 22:52:06.193355 418392 kubeadm.go:319]
I1027 22:52:06.193435 418392 kubeadm.go:319] To start using your cluster, you need to run the following as a regular user:
I1027 22:52:06.193439 418392 kubeadm.go:319]
I1027 22:52:06.193463 418392 kubeadm.go:319] mkdir -p $HOME/.kube
I1027 22:52:06.193524 418392 kubeadm.go:319] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I1027 22:52:06.193577 418392 kubeadm.go:319] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I1027 22:52:06.193580 418392 kubeadm.go:319]
I1027 22:52:06.193635 418392 kubeadm.go:319] Alternatively, if you are the root user, you can run:
I1027 22:52:06.193639 418392 kubeadm.go:319]
I1027 22:52:06.193688 418392 kubeadm.go:319] export KUBECONFIG=/etc/kubernetes/admin.conf
I1027 22:52:06.193691 418392 kubeadm.go:319]
I1027 22:52:06.193745 418392 kubeadm.go:319] You should now deploy a pod network to the cluster.
I1027 22:52:06.193823 418392 kubeadm.go:319] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I1027 22:52:06.193893 418392 kubeadm.go:319] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I1027 22:52:06.193900 418392 kubeadm.go:319]
I1027 22:52:06.193987 418392 kubeadm.go:319] You can now join any number of control-plane nodes by copying certificate authorities
I1027 22:52:06.194066 418392 kubeadm.go:319] and service account keys on each node and then running the following as root:
I1027 22:52:06.194070 418392 kubeadm.go:319]
I1027 22:52:06.194157 418392 kubeadm.go:319] kubeadm join control-plane.minikube.internal:8443 --token x3pj4l.awtike51qlcy8cjp \
I1027 22:52:06.194264 418392 kubeadm.go:319] --discovery-token-ca-cert-hash sha256:e5bde37e868b0a9f20cc35703c4d8ced7fe96b47e180bf7d5d1b064d5adb88da \
I1027 22:52:06.194289 418392 kubeadm.go:319] --control-plane
I1027 22:52:06.194292 418392 kubeadm.go:319]
I1027 22:52:06.194380 418392 kubeadm.go:319] Then you can join any number of worker nodes by running the following on each as root:
I1027 22:52:06.194384 418392 kubeadm.go:319]
I1027 22:52:06.194468 418392 kubeadm.go:319] kubeadm join control-plane.minikube.internal:8443 --token x3pj4l.awtike51qlcy8cjp \
I1027 22:52:06.194574 418392 kubeadm.go:319] --discovery-token-ca-cert-hash sha256:e5bde37e868b0a9f20cc35703c4d8ced7fe96b47e180bf7d5d1b064d5adb88da
I1027 22:52:06.198360 418392 kubeadm.go:319] [WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2
I1027 22:52:06.198575 418392 kubeadm.go:319] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1084-aws\n", err: exit status 1
I1027 22:52:06.198677 418392 kubeadm.go:319] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I1027 22:52:06.198691 418392 cni.go:84] Creating CNI manager for ""
I1027 22:52:06.198698 418392 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I1027 22:52:06.201949 418392 out.go:179] * Configuring CNI (Container Networking Interface) ...
I1027 22:52:06.204978 418392 ssh_runner.go:195] Run: stat /opt/cni/bin/portmap
I1027 22:52:06.209261 418392 cni.go:182] applying CNI manifest using /var/lib/minikube/binaries/v1.34.1/kubectl ...
I1027 22:52:06.209279 418392 ssh_runner.go:362] scp memory --> /var/tmp/minikube/cni.yaml (2601 bytes)
I1027 22:52:06.224963 418392 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.1/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml
I1027 22:52:06.525485 418392 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I1027 22:52:06.525622 418392 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I1027 22:52:06.525719 418392 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes scheduled-stop-727712 minikube.k8s.io/updated_at=2025_10_27T22_52_06_0700 minikube.k8s.io/version=v1.37.0 minikube.k8s.io/commit=e7da329725eb7dc274e4db0e5490c73fe54de60f minikube.k8s.io/name=scheduled-stop-727712 minikube.k8s.io/primary=true
I1027 22:52:06.542334 418392 ops.go:34] apiserver oom_adj: -16
I1027 22:52:06.764731 418392 kubeadm.go:1114] duration metric: took 239.154383ms to wait for elevateKubeSystemPrivileges
I1027 22:52:06.802055 418392 kubeadm.go:403] duration metric: took 18.613010618s to StartCluster
I1027 22:52:06.802081 418392 settings.go:142] acquiring lock: {Name:mkf2704123c96ec6115f0b73542ebb274f80c701 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1027 22:52:06.802144 418392 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/21790-269600/kubeconfig
I1027 22:52:06.802781 418392 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21790-269600/kubeconfig: {Name:mk0ee9c08e1ab37887a79c19b3bd04613966c4db Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1027 22:52:06.802974 418392 start.go:236] Will wait 6m0s for node &{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
I1027 22:52:06.803095 418392 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I1027 22:52:06.803323 418392 config.go:182] Loaded profile config "scheduled-stop-727712": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.34.1
I1027 22:52:06.803355 418392 addons.go:511] enable addons start: toEnable=map[ambassador:false amd-gpu-device-plugin:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubetail:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-rancher:false volcano:false volumesnapshots:false yakd:false]
I1027 22:52:06.803409 418392 addons.go:69] Setting storage-provisioner=true in profile "scheduled-stop-727712"
I1027 22:52:06.803422 418392 addons.go:238] Setting addon storage-provisioner=true in "scheduled-stop-727712"
I1027 22:52:06.803447 418392 host.go:66] Checking if "scheduled-stop-727712" exists ...
I1027 22:52:06.803942 418392 cli_runner.go:164] Run: docker container inspect scheduled-stop-727712 --format={{.State.Status}}
I1027 22:52:06.804363 418392 addons.go:69] Setting default-storageclass=true in profile "scheduled-stop-727712"
I1027 22:52:06.804380 418392 addons_storage_classes.go:34] enableOrDisableStorageClasses default-storageclass=true on "scheduled-stop-727712"
I1027 22:52:06.804661 418392 cli_runner.go:164] Run: docker container inspect scheduled-stop-727712 --format={{.State.Status}}
I1027 22:52:06.809623 418392 out.go:179] * Verifying Kubernetes components...
I1027 22:52:06.814874 418392 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1027 22:52:06.834480 418392 addons.go:238] Setting addon default-storageclass=true in "scheduled-stop-727712"
I1027 22:52:06.834507 418392 host.go:66] Checking if "scheduled-stop-727712" exists ...
I1027 22:52:06.834973 418392 cli_runner.go:164] Run: docker container inspect scheduled-stop-727712 --format={{.State.Status}}
I1027 22:52:06.845454 418392 out.go:179] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I1027 22:52:06.850844 418392 addons.go:435] installing /etc/kubernetes/addons/storage-provisioner.yaml
I1027 22:52:06.850856 418392 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I1027 22:52:06.850923 418392 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-727712
I1027 22:52:06.862488 418392 addons.go:435] installing /etc/kubernetes/addons/storageclass.yaml
I1027 22:52:06.862501 418392 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I1027 22:52:06.862571 418392 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-727712
I1027 22:52:06.880898 418392 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33335 SSHKeyPath:/home/jenkins/minikube-integration/21790-269600/.minikube/machines/scheduled-stop-727712/id_rsa Username:docker}
I1027 22:52:06.905114 418392 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33335 SSHKeyPath:/home/jenkins/minikube-integration/21790-269600/.minikube/machines/scheduled-stop-727712/id_rsa Username:docker}
I1027 22:52:07.048196 418392 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.76.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I1027 22:52:07.053142 418392 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.34.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I1027 22:52:07.080250 418392 ssh_runner.go:195] Run: sudo systemctl start kubelet
I1027 22:52:07.083438 418392 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.34.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I1027 22:52:07.454958 418392 start.go:977] {"host.minikube.internal": 192.168.76.1} host record injected into CoreDNS's ConfigMap
I1027 22:52:07.456440 418392 api_server.go:52] waiting for apiserver process to appear ...
I1027 22:52:07.456484 418392 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I1027 22:52:07.633932 418392 api_server.go:72] duration metric: took 830.895862ms to wait for apiserver process to appear ...
I1027 22:52:07.633944 418392 api_server.go:88] waiting for apiserver healthz status ...
I1027 22:52:07.633960 418392 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
I1027 22:52:07.636865 418392 out.go:179] * Enabled addons: default-storageclass, storage-provisioner
I1027 22:52:07.639402 418392 addons.go:514] duration metric: took 836.023799ms for enable addons: enabled=[default-storageclass storage-provisioner]
I1027 22:52:07.648979 418392 api_server.go:279] https://192.168.76.2:8443/healthz returned 200:
ok
I1027 22:52:07.649972 418392 api_server.go:141] control plane version: v1.34.1
I1027 22:52:07.649986 418392 api_server.go:131] duration metric: took 16.036948ms to wait for apiserver health ...
I1027 22:52:07.649993 418392 system_pods.go:43] waiting for kube-system pods to appear ...
I1027 22:52:07.655379 418392 system_pods.go:59] 5 kube-system pods found
I1027 22:52:07.655398 418392 system_pods.go:61] "etcd-scheduled-stop-727712" [b6f1c3d1-812d-42cb-8d23-03f8e4094eac] Running / Ready:ContainersNotReady (containers with unready status: [etcd]) / ContainersReady:ContainersNotReady (containers with unready status: [etcd])
I1027 22:52:07.655408 418392 system_pods.go:61] "kube-apiserver-scheduled-stop-727712" [9f0a2dab-bde2-4e87-bcb7-df3f075fa89d] Running / Ready:ContainersNotReady (containers with unready status: [kube-apiserver]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-apiserver])
I1027 22:52:07.655417 418392 system_pods.go:61] "kube-controller-manager-scheduled-stop-727712" [41b46a48-3c24-4f79-9b00-8fd5ec9c0938] Running / Ready:ContainersNotReady (containers with unready status: [kube-controller-manager]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-controller-manager])
I1027 22:52:07.655423 418392 system_pods.go:61] "kube-scheduler-scheduled-stop-727712" [cb8f5226-4f50-4672-a78d-7eeb0b6e7436] Running / Ready:ContainersNotReady (containers with unready status: [kube-scheduler]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-scheduler])
I1027 22:52:07.655428 418392 system_pods.go:61] "storage-provisioner" [5b1a327d-3c03-46f9-bb3e-c2ca777134ea] Pending: PodScheduled:Unschedulable (0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. no new claims to deallocate, preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.)
I1027 22:52:07.655433 418392 system_pods.go:74] duration metric: took 5.434977ms to wait for pod list to return data ...
I1027 22:52:07.655444 418392 kubeadm.go:587] duration metric: took 852.413146ms to wait for: map[apiserver:true system_pods:true]
I1027 22:52:07.655455 418392 node_conditions.go:102] verifying NodePressure condition ...
I1027 22:52:07.659669 418392 node_conditions.go:122] node storage ephemeral capacity is 203034800Ki
I1027 22:52:07.659689 418392 node_conditions.go:123] node cpu capacity is 2
I1027 22:52:07.659699 418392 node_conditions.go:105] duration metric: took 4.239436ms to run NodePressure ...
I1027 22:52:07.659710 418392 start.go:242] waiting for startup goroutines ...
I1027 22:52:07.959618 418392 kapi.go:214] "coredns" deployment in "kube-system" namespace and "scheduled-stop-727712" context rescaled to 1 replicas
I1027 22:52:07.959648 418392 start.go:247] waiting for cluster config update ...
I1027 22:52:07.959660 418392 start.go:256] writing updated cluster config ...
I1027 22:52:07.959959 418392 ssh_runner.go:195] Run: rm -f paused
I1027 22:52:08.022066 418392 start.go:626] kubectl: 1.33.2, cluster: 1.34.1 (minor skew: 1)
I1027 22:52:08.025976 418392 out.go:179] * Done! kubectl is now configured to use "scheduled-stop-727712" cluster and "default" namespace by default
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD NAMESPACE
470713fd5059f a1894772a478e 11 seconds ago Running etcd 0 73a8daf09bfc4 etcd-scheduled-stop-727712 kube-system
ea6b40d79beaa 43911e833d64d 11 seconds ago Running kube-apiserver 0 3e763dc732c25 kube-apiserver-scheduled-stop-727712 kube-system
87d46ba327bb4 7eb2c6ff0c5a7 11 seconds ago Running kube-controller-manager 0 12de0f5565b69 kube-controller-manager-scheduled-stop-727712 kube-system
3465522f0e6fe b5f57ec6b9867 11 seconds ago Running kube-scheduler 0 5233bb03777c8 kube-scheduler-scheduled-stop-727712 kube-system
==> containerd <==
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.230293306Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-scheduled-stop-727712,Uid:a11bb2e02a7fcf4365c3eceb1a732bd5,Namespace:kube-system,Attempt:0,}"
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.236302098Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:etcd-scheduled-stop-727712,Uid:f90249fd5db74f66b8c702404888f66b,Namespace:kube-system,Attempt:0,}"
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.246170147Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-scheduled-stop-727712,Uid:36b305b48e9d4f07c3e68a639a017971,Namespace:kube-system,Attempt:0,}"
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.249545566Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-scheduled-stop-727712,Uid:2f77f89dc401b2fdd5ce53b49edcb9c9,Namespace:kube-system,Attempt:0,}"
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.333200276Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-scheduled-stop-727712,Uid:a11bb2e02a7fcf4365c3eceb1a732bd5,Namespace:kube-system,Attempt:0,} returns sandbox id \"5233bb03777c853709044870245c25eecc433539ff58d8adca485bf1f7440c45\""
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.342156266Z" level=info msg="CreateContainer within sandbox \"5233bb03777c853709044870245c25eecc433539ff58d8adca485bf1f7440c45\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}"
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.363184486Z" level=info msg="CreateContainer within sandbox \"5233bb03777c853709044870245c25eecc433539ff58d8adca485bf1f7440c45\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"3465522f0e6fecaf49e0a70169a576ed0728b6b203a509748e14fb8abb0054d1\""
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.363957179Z" level=info msg="StartContainer for \"3465522f0e6fecaf49e0a70169a576ed0728b6b203a509748e14fb8abb0054d1\""
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.441675771Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-scheduled-stop-727712,Uid:36b305b48e9d4f07c3e68a639a017971,Namespace:kube-system,Attempt:0,} returns sandbox id \"12de0f5565b6977e9fd798f57b3100fda5662da154002d146ef68993b2a8e822\""
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.449316014Z" level=info msg="CreateContainer within sandbox \"12de0f5565b6977e9fd798f57b3100fda5662da154002d146ef68993b2a8e822\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}"
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.469030670Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-scheduled-stop-727712,Uid:2f77f89dc401b2fdd5ce53b49edcb9c9,Namespace:kube-system,Attempt:0,} returns sandbox id \"3e763dc732c25f72bf880a5a741acd59183f6d12a5fd3e76f66b87e05566171d\""
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.479050227Z" level=info msg="CreateContainer within sandbox \"12de0f5565b6977e9fd798f57b3100fda5662da154002d146ef68993b2a8e822\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"87d46ba327bb4fa01b948b9d8c95d50b61e6c69547b9cfc1ea1bd3ed84116de0\""
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.480435226Z" level=info msg="StartContainer for \"87d46ba327bb4fa01b948b9d8c95d50b61e6c69547b9cfc1ea1bd3ed84116de0\""
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.483575968Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:etcd-scheduled-stop-727712,Uid:f90249fd5db74f66b8c702404888f66b,Namespace:kube-system,Attempt:0,} returns sandbox id \"73a8daf09bfc4efeebffb912068785007009d6edb4a803683c67b9dffe9977eb\""
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.484659154Z" level=info msg="CreateContainer within sandbox \"3e763dc732c25f72bf880a5a741acd59183f6d12a5fd3e76f66b87e05566171d\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}"
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.506597292Z" level=info msg="CreateContainer within sandbox \"73a8daf09bfc4efeebffb912068785007009d6edb4a803683c67b9dffe9977eb\" for container &ContainerMetadata{Name:etcd,Attempt:0,}"
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.552979020Z" level=info msg="CreateContainer within sandbox \"3e763dc732c25f72bf880a5a741acd59183f6d12a5fd3e76f66b87e05566171d\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"ea6b40d79beaa878730654815431d2160efdf71c999a6f0103e1f83f6849c43b\""
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.555310695Z" level=info msg="StartContainer for \"ea6b40d79beaa878730654815431d2160efdf71c999a6f0103e1f83f6849c43b\""
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.592719482Z" level=info msg="CreateContainer within sandbox \"73a8daf09bfc4efeebffb912068785007009d6edb4a803683c67b9dffe9977eb\" for &ContainerMetadata{Name:etcd,Attempt:0,} returns container id \"470713fd5059f25961f796c5a9231ed8665ec82e7c9c9073d24d476488b86aeb\""
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.598848152Z" level=info msg="StartContainer for \"3465522f0e6fecaf49e0a70169a576ed0728b6b203a509748e14fb8abb0054d1\" returns successfully"
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.606605352Z" level=info msg="StartContainer for \"470713fd5059f25961f796c5a9231ed8665ec82e7c9c9073d24d476488b86aeb\""
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.610446704Z" level=info msg="StartContainer for \"87d46ba327bb4fa01b948b9d8c95d50b61e6c69547b9cfc1ea1bd3ed84116de0\" returns successfully"
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.741252874Z" level=info msg="StartContainer for \"ea6b40d79beaa878730654815431d2160efdf71c999a6f0103e1f83f6849c43b\" returns successfully"
Oct 27 22:51:58 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:51:58.757225460Z" level=info msg="StartContainer for \"470713fd5059f25961f796c5a9231ed8665ec82e7c9c9073d24d476488b86aeb\" returns successfully"
Oct 27 22:52:09 scheduled-stop-727712 containerd[763]: time="2025-10-27T22:52:09.466655048Z" level=info msg="No cni config template is specified, wait for other system components to drop the config."
==> describe nodes <==
Name: scheduled-stop-727712
Roles: control-plane
Labels: beta.kubernetes.io/arch=arm64
beta.kubernetes.io/os=linux
kubernetes.io/arch=arm64
kubernetes.io/hostname=scheduled-stop-727712
kubernetes.io/os=linux
minikube.k8s.io/commit=e7da329725eb7dc274e4db0e5490c73fe54de60f
minikube.k8s.io/name=scheduled-stop-727712
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2025_10_27T22_52_06_0700
minikube.k8s.io/version=v1.37.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
Annotations: volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Mon, 27 Oct 2025 22:52:02 +0000
Taints: node.kubernetes.io/not-ready:NoSchedule
Unschedulable: false
Lease:
HolderIdentity: scheduled-stop-727712
AcquireTime: <unset>
RenewTime: Mon, 27 Oct 2025 22:52:05 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Mon, 27 Oct 2025 22:52:05 +0000 Mon, 27 Oct 2025 22:51:59 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Mon, 27 Oct 2025 22:52:05 +0000 Mon, 27 Oct 2025 22:51:59 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Mon, 27 Oct 2025 22:52:05 +0000 Mon, 27 Oct 2025 22:51:59 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready False Mon, 27 Oct 2025 22:52:05 +0000 Mon, 27 Oct 2025 22:51:59 +0000 KubeletNotReady container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized
Addresses:
InternalIP: 192.168.76.2
Hostname: scheduled-stop-727712
Capacity:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022296Ki
pods: 110
Allocatable:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022296Ki
pods: 110
System Info:
Machine ID: ef38fbc8889a0e5f09e9dc0868f5cd19
System UUID: edba9d75-2e97-418e-b3c9-1495c89d2650
Boot ID: 9ceac5df-4f07-4c4c-b81a-a03ec3534783
Kernel Version: 5.15.0-1084-aws
OS Image: Debian GNU/Linux 12 (bookworm)
Operating System: linux
Architecture: arm64
Container Runtime Version: containerd://1.7.28
Kubelet Version: v1.34.1
Kube-Proxy Version:
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (4 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
kube-system etcd-scheduled-stop-727712 100m (5%) 0 (0%) 100Mi (1%) 0 (0%) 4s
kube-system kube-apiserver-scheduled-stop-727712 250m (12%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-controller-manager-scheduled-stop-727712 200m (10%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-scheduler-scheduled-stop-727712 100m (5%) 0 (0%) 0 (0%) 0 (0%) 4s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 650m (32%) 0 (0%)
memory 100Mi (1%) 0 (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
hugepages-32Mi 0 (0%) 0 (0%)
hugepages-64Ki 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal NodeAllocatableEnforced 12s kubelet Updated Node Allocatable limit across pods
Normal Starting 12s kubelet Starting kubelet.
Warning CgroupV1 12s kubelet cgroup v1 support is in maintenance mode, please migrate to cgroup v2
Normal NodeHasSufficientMemory 12s (x8 over 12s) kubelet Node scheduled-stop-727712 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12s (x8 over 12s) kubelet Node scheduled-stop-727712 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12s (x7 over 12s) kubelet Node scheduled-stop-727712 status is now: NodeHasSufficientPID
Normal Starting 4s kubelet Starting kubelet.
Warning CgroupV1 4s kubelet cgroup v1 support is in maintenance mode, please migrate to cgroup v2
Normal NodeAllocatableEnforced 4s kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 4s kubelet Node scheduled-stop-727712 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 4s kubelet Node scheduled-stop-727712 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 4s kubelet Node scheduled-stop-727712 status is now: NodeHasSufficientPID
Normal RegisteredNode 0s node-controller Node scheduled-stop-727712 event: Registered Node scheduled-stop-727712 in Controller
==> dmesg <==
[Oct27 21:49] overlayfs: idmapped layers are currently not supported
[ +30.761565] overlayfs: idmapped layers are currently not supported
[Oct27 21:51] overlayfs: idmapped layers are currently not supported
[Oct27 21:53] overlayfs: idmapped layers are currently not supported
[Oct27 21:54] overlayfs: idmapped layers are currently not supported
[Oct27 21:55] overlayfs: idmapped layers are currently not supported
[Oct27 21:59] overlayfs: idmapped layers are currently not supported
[Oct27 22:00] overlayfs: idmapped layers are currently not supported
[ +24.025643] overlayfs: idmapped layers are currently not supported
[Oct27 22:01] overlayfs: idmapped layers are currently not supported
[Oct27 22:02] overlayfs: idmapped layers are currently not supported
[ +55.889286] overlayfs: idmapped layers are currently not supported
[Oct27 22:03] overlayfs: idmapped layers are currently not supported
[Oct27 22:04] overlayfs: idmapped layers are currently not supported
[Oct27 22:05] overlayfs: idmapped layers are currently not supported
[ +46.465103] overlayfs: idmapped layers are currently not supported
[Oct27 22:06] overlayfs: idmapped layers are currently not supported
[ +0.509504] overlayfs: idmapped layers are currently not supported
[Oct27 22:07] overlayfs: idmapped layers are currently not supported
[Oct27 22:08] overlayfs: idmapped layers are currently not supported
[Oct27 22:09] overlayfs: idmapped layers are currently not supported
[Oct27 22:10] overlayfs: idmapped layers are currently not supported
[Oct27 22:11] overlayfs: idmapped layers are currently not supported
[ +52.535536] overlayfs: idmapped layers are currently not supported
[Oct27 22:14] kauditd_printk_skb: 8 callbacks suppressed
==> etcd [470713fd5059f25961f796c5a9231ed8665ec82e7c9c9073d24d476488b86aeb] <==
{"level":"warn","ts":"2025-10-27T22:52:00.958720Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51214","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:00.997714Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51260","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:00.999573Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51234","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.008578Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51270","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.025570Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51290","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.059492Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51296","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.062297Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51304","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.093534Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51322","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.106688Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51338","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.131180Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51366","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.140165Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51372","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.168850Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51402","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.190464Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51412","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.216280Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51434","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.226161Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51458","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.243611Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51490","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.270342Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51506","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.284214Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51518","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.307133Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51526","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.332316Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51546","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.345029Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51568","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.371174Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51594","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.385129Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51606","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.418658Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51616","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-27T22:52:01.510010Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:51636","server-name":"","error":"EOF"}
==> kernel <==
22:52:09 up 2:34, 0 user, load average: 2.05, 2.15, 2.56
Linux scheduled-stop-727712 5.15.0-1084-aws #91~20.04.1-Ubuntu SMP Fri May 2 07:00:04 UTC 2025 aarch64 GNU/Linux
PRETTY_NAME="Debian GNU/Linux 12 (bookworm)"
==> kube-apiserver [ea6b40d79beaa878730654815431d2160efdf71c999a6f0103e1f83f6849c43b] <==
I1027 22:52:02.512553 1 shared_informer.go:356] "Caches are synced" controller="*generic.policySource[*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicy,*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicyBinding,k8s.io/apiserver/pkg/admission/plugin/policy/validating.Validator]"
I1027 22:52:02.512637 1 policy_source.go:240] refreshing policies
I1027 22:52:02.512721 1 handler_discovery.go:451] Starting ResourceDiscoveryManager
I1027 22:52:02.514203 1 controller.go:667] quota admission added evaluator for: namespaces
I1027 22:52:02.537995 1 controller.go:667] quota admission added evaluator for: leases.coordination.k8s.io
I1027 22:52:02.626758 1 cidrallocator.go:301] created ClusterIP allocator for Service CIDR 10.96.0.0/12
I1027 22:52:02.626928 1 default_servicecidr_controller.go:228] Setting default ServiceCIDR condition Ready to True
I1027 22:52:02.725967 1 cidrallocator.go:277] updated ClusterIP allocator for Service CIDR 10.96.0.0/12
I1027 22:52:02.726078 1 default_servicecidr_controller.go:137] Shutting down kubernetes-service-cidr-controller
I1027 22:52:03.168425 1 storage_scheduling.go:95] created PriorityClass system-node-critical with value 2000001000
I1027 22:52:03.184484 1 storage_scheduling.go:95] created PriorityClass system-cluster-critical with value 2000000000
I1027 22:52:03.184734 1 storage_scheduling.go:111] all system priority classes are created successfully or already exist.
I1027 22:52:04.168873 1 controller.go:667] quota admission added evaluator for: roles.rbac.authorization.k8s.io
I1027 22:52:04.245669 1 controller.go:667] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
I1027 22:52:04.351617 1 controller.go:667] quota admission added evaluator for: serviceaccounts
I1027 22:52:04.411109 1 alloc.go:328] "allocated clusterIPs" service="default/kubernetes" clusterIPs={"IPv4":"10.96.0.1"}
W1027 22:52:04.437794 1 lease.go:265] Resetting endpoints for master service "kubernetes" to [192.168.76.2]
I1027 22:52:04.440537 1 controller.go:667] quota admission added evaluator for: endpoints
I1027 22:52:04.449167 1 controller.go:667] quota admission added evaluator for: endpointslices.discovery.k8s.io
I1027 22:52:05.605803 1 controller.go:667] quota admission added evaluator for: deployments.apps
I1027 22:52:05.625059 1 alloc.go:328] "allocated clusterIPs" service="kube-system/kube-dns" clusterIPs={"IPv4":"10.96.0.10"}
I1027 22:52:05.639484 1 controller.go:667] quota admission added evaluator for: daemonsets.apps
I1027 22:52:09.461409 1 cidrallocator.go:277] updated ClusterIP allocator for Service CIDR 10.96.0.0/12
I1027 22:52:09.468268 1 cidrallocator.go:277] updated ClusterIP allocator for Service CIDR 10.96.0.0/12
I1027 22:52:09.902704 1 controller.go:667] quota admission added evaluator for: controllerrevisions.apps
==> kube-controller-manager [87d46ba327bb4fa01b948b9d8c95d50b61e6c69547b9cfc1ea1bd3ed84116de0] <==
I1027 22:52:09.396935 1 shared_informer.go:356] "Caches are synced" controller="ReplicaSet"
I1027 22:52:09.396985 1 shared_informer.go:356] "Caches are synced" controller="expand"
I1027 22:52:09.397147 1 shared_informer.go:356] "Caches are synced" controller="TTL"
I1027 22:52:09.398190 1 shared_informer.go:356] "Caches are synced" controller="resource_claim"
I1027 22:52:09.399276 1 shared_informer.go:356] "Caches are synced" controller="job"
I1027 22:52:09.399947 1 shared_informer.go:356] "Caches are synced" controller="cronjob"
I1027 22:52:09.400045 1 shared_informer.go:356] "Caches are synced" controller="endpoint"
I1027 22:52:09.400218 1 shared_informer.go:356] "Caches are synced" controller="ephemeral"
I1027 22:52:09.400625 1 shared_informer.go:356] "Caches are synced" controller="persistent volume"
I1027 22:52:09.403836 1 shared_informer.go:356] "Caches are synced" controller="garbage collector"
I1027 22:52:09.407099 1 shared_informer.go:356] "Caches are synced" controller="resource quota"
I1027 22:52:09.407346 1 shared_informer.go:356] "Caches are synced" controller="PV protection"
I1027 22:52:09.416609 1 shared_informer.go:356] "Caches are synced" controller="validatingadmissionpolicy-status"
I1027 22:52:09.430117 1 shared_informer.go:356] "Caches are synced" controller="endpoint_slice_mirroring"
I1027 22:52:09.444874 1 shared_informer.go:356] "Caches are synced" controller="taint"
I1027 22:52:09.444972 1 node_lifecycle_controller.go:1221] "Initializing eviction metric for zone" logger="node-lifecycle-controller" zone=""
I1027 22:52:09.445044 1 node_lifecycle_controller.go:873] "Missing timestamp for Node. Assuming now as a timestamp" logger="node-lifecycle-controller" node="scheduled-stop-727712"
I1027 22:52:09.445087 1 node_lifecycle_controller.go:1025] "Controller detected that all Nodes are not-Ready. Entering master disruption mode" logger="node-lifecycle-controller"
I1027 22:52:09.445118 1 shared_informer.go:356] "Caches are synced" controller="HPA"
I1027 22:52:09.445696 1 shared_informer.go:356] "Caches are synced" controller="legacy-service-account-token-cleaner"
I1027 22:52:09.447481 1 shared_informer.go:356] "Caches are synced" controller="attach detach"
I1027 22:52:09.447910 1 shared_informer.go:356] "Caches are synced" controller="ReplicationController"
I1027 22:52:09.462253 1 shared_informer.go:356] "Caches are synced" controller="garbage collector"
I1027 22:52:09.462275 1 garbagecollector.go:154] "Garbage collector: all resource monitors have synced" logger="garbage-collector-controller"
I1027 22:52:09.462282 1 garbagecollector.go:157] "Proceeding to collect garbage" logger="garbage-collector-controller"
==> kube-scheduler [3465522f0e6fecaf49e0a70169a576ed0728b6b203a509748e14fb8abb0054d1] <==
I1027 22:52:01.519354 1 serving.go:386] Generated self-signed cert in-memory
I1027 22:52:04.766942 1 server.go:175] "Starting Kubernetes Scheduler" version="v1.34.1"
I1027 22:52:04.766981 1 server.go:177] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I1027 22:52:04.772596 1 secure_serving.go:211] Serving securely on 127.0.0.1:10259
I1027 22:52:04.772679 1 requestheader_controller.go:180] Starting RequestHeaderAuthRequestController
I1027 22:52:04.772709 1 shared_informer.go:349] "Waiting for caches to sync" controller="RequestHeaderAuthRequestController"
I1027 22:52:04.772747 1 tlsconfig.go:243] "Starting DynamicServingCertificateController"
I1027 22:52:04.778972 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file"
I1027 22:52:04.779023 1 shared_informer.go:349] "Waiting for caches to sync" controller="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file"
I1027 22:52:04.779162 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
I1027 22:52:04.779175 1 shared_informer.go:349] "Waiting for caches to sync" controller="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
I1027 22:52:04.872798 1 shared_informer.go:356] "Caches are synced" controller="RequestHeaderAuthRequestController"
I1027 22:52:04.879624 1 shared_informer.go:356] "Caches are synced" controller="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
I1027 22:52:04.879690 1 shared_informer.go:356] "Caches are synced" controller="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file"
==> kubelet <==
Oct 27 22:52:06 scheduled-stop-727712 kubelet[1505]: I1027 22:52:06.552248 1505 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world"
Oct 27 22:52:06 scheduled-stop-727712 kubelet[1505]: I1027 22:52:06.640344 1505 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-scheduled-stop-727712" podStartSLOduration=1.640324867 podStartE2EDuration="1.640324867s" podCreationTimestamp="2025-10-27 22:52:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-27 22:52:06.597495537 +0000 UTC m=+1.161614495" watchObservedRunningTime="2025-10-27 22:52:06.640324867 +0000 UTC m=+1.204443825"
Oct 27 22:52:06 scheduled-stop-727712 kubelet[1505]: I1027 22:52:06.655640 1505 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/etcd-scheduled-stop-727712"
Oct 27 22:52:06 scheduled-stop-727712 kubelet[1505]: I1027 22:52:06.662636 1505 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-scheduled-stop-727712"
Oct 27 22:52:06 scheduled-stop-727712 kubelet[1505]: I1027 22:52:06.663182 1505 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-scheduled-stop-727712" podStartSLOduration=1.663167489 podStartE2EDuration="1.663167489s" podCreationTimestamp="2025-10-27 22:52:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-27 22:52:06.640650335 +0000 UTC m=+1.204769301" watchObservedRunningTime="2025-10-27 22:52:06.663167489 +0000 UTC m=+1.227286496"
Oct 27 22:52:06 scheduled-stop-727712 kubelet[1505]: I1027 22:52:06.663339 1505 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-scheduled-stop-727712" podStartSLOduration=1.663331216 podStartE2EDuration="1.663331216s" podCreationTimestamp="2025-10-27 22:52:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-27 22:52:06.662636235 +0000 UTC m=+1.226755217" watchObservedRunningTime="2025-10-27 22:52:06.663331216 +0000 UTC m=+1.227450199"
Oct 27 22:52:06 scheduled-stop-727712 kubelet[1505]: E1027 22:52:06.668481 1505 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"etcd-scheduled-stop-727712\" already exists" pod="kube-system/etcd-scheduled-stop-727712"
Oct 27 22:52:06 scheduled-stop-727712 kubelet[1505]: E1027 22:52:06.675609 1505 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-apiserver-scheduled-stop-727712\" already exists" pod="kube-system/kube-apiserver-scheduled-stop-727712"
Oct 27 22:52:06 scheduled-stop-727712 kubelet[1505]: I1027 22:52:06.700722 1505 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/etcd-scheduled-stop-727712" podStartSLOduration=1.700701471 podStartE2EDuration="1.700701471s" podCreationTimestamp="2025-10-27 22:52:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-27 22:52:06.685528581 +0000 UTC m=+1.249647538" watchObservedRunningTime="2025-10-27 22:52:06.700701471 +0000 UTC m=+1.264820429"
Oct 27 22:52:09 scheduled-stop-727712 kubelet[1505]: I1027 22:52:09.465186 1505 kuberuntime_manager.go:1828] "Updating runtime config through cri with podcidr" CIDR="10.244.0.0/24"
Oct 27 22:52:09 scheduled-stop-727712 kubelet[1505]: I1027 22:52:09.466903 1505 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="10.244.0.0/24"
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: I1027 22:52:10.006012 1505 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3cd975eb-73fd-47bf-b537-2f092a7d019d-xtables-lock\") pod \"kindnet-jzkr4\" (UID: \"3cd975eb-73fd-47bf-b537-2f092a7d019d\") " pod="kube-system/kindnet-jzkr4"
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: I1027 22:52:10.006119 1505 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2rtg\" (UniqueName: \"kubernetes.io/projected/f3e2228d-d053-48f5-94e8-42d9fcaf5226-kube-api-access-h2rtg\") pod \"kube-proxy-mfzjd\" (UID: \"f3e2228d-d053-48f5-94e8-42d9fcaf5226\") " pod="kube-system/kube-proxy-mfzjd"
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: I1027 22:52:10.006187 1505 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-cfg\" (UniqueName: \"kubernetes.io/host-path/3cd975eb-73fd-47bf-b537-2f092a7d019d-cni-cfg\") pod \"kindnet-jzkr4\" (UID: \"3cd975eb-73fd-47bf-b537-2f092a7d019d\") " pod="kube-system/kindnet-jzkr4"
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: I1027 22:52:10.006209 1505 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/f3e2228d-d053-48f5-94e8-42d9fcaf5226-kube-proxy\") pod \"kube-proxy-mfzjd\" (UID: \"f3e2228d-d053-48f5-94e8-42d9fcaf5226\") " pod="kube-system/kube-proxy-mfzjd"
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: I1027 22:52:10.006259 1505 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3cd975eb-73fd-47bf-b537-2f092a7d019d-lib-modules\") pod \"kindnet-jzkr4\" (UID: \"3cd975eb-73fd-47bf-b537-2f092a7d019d\") " pod="kube-system/kindnet-jzkr4"
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: I1027 22:52:10.006330 1505 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f3e2228d-d053-48f5-94e8-42d9fcaf5226-xtables-lock\") pod \"kube-proxy-mfzjd\" (UID: \"f3e2228d-d053-48f5-94e8-42d9fcaf5226\") " pod="kube-system/kube-proxy-mfzjd"
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: I1027 22:52:10.006352 1505 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6jf8\" (UniqueName: \"kubernetes.io/projected/3cd975eb-73fd-47bf-b537-2f092a7d019d-kube-api-access-r6jf8\") pod \"kindnet-jzkr4\" (UID: \"3cd975eb-73fd-47bf-b537-2f092a7d019d\") " pod="kube-system/kindnet-jzkr4"
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: I1027 22:52:10.006506 1505 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f3e2228d-d053-48f5-94e8-42d9fcaf5226-lib-modules\") pod \"kube-proxy-mfzjd\" (UID: \"f3e2228d-d053-48f5-94e8-42d9fcaf5226\") " pod="kube-system/kube-proxy-mfzjd"
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: E1027 22:52:10.131375 1505 projected.go:291] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: E1027 22:52:10.131406 1505 projected.go:196] Error preparing data for projected volume kube-api-access-r6jf8 for pod kube-system/kindnet-jzkr4: configmap "kube-root-ca.crt" not found
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: E1027 22:52:10.131481 1505 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3cd975eb-73fd-47bf-b537-2f092a7d019d-kube-api-access-r6jf8 podName:3cd975eb-73fd-47bf-b537-2f092a7d019d nodeName:}" failed. No retries permitted until 2025-10-27 22:52:10.631456607 +0000 UTC m=+5.195575565 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-r6jf8" (UniqueName: "kubernetes.io/projected/3cd975eb-73fd-47bf-b537-2f092a7d019d-kube-api-access-r6jf8") pod "kindnet-jzkr4" (UID: "3cd975eb-73fd-47bf-b537-2f092a7d019d") : configmap "kube-root-ca.crt" not found
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: E1027 22:52:10.132070 1505 projected.go:291] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: E1027 22:52:10.132091 1505 projected.go:196] Error preparing data for projected volume kube-api-access-h2rtg for pod kube-system/kube-proxy-mfzjd: configmap "kube-root-ca.crt" not found
Oct 27 22:52:10 scheduled-stop-727712 kubelet[1505]: E1027 22:52:10.132141 1505 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f3e2228d-d053-48f5-94e8-42d9fcaf5226-kube-api-access-h2rtg podName:f3e2228d-d053-48f5-94e8-42d9fcaf5226 nodeName:}" failed. No retries permitted until 2025-10-27 22:52:10.632124199 +0000 UTC m=+5.196243165 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-h2rtg" (UniqueName: "kubernetes.io/projected/f3e2228d-d053-48f5-94e8-42d9fcaf5226-kube-api-access-h2rtg") pod "kube-proxy-mfzjd" (UID: "f3e2228d-d053-48f5-94e8-42d9fcaf5226") : configmap "kube-root-ca.crt" not found
-- /stdout --
helpers_test.go:262: (dbg) Run: out/minikube-linux-arm64 status --format={{.APIServer}} -p scheduled-stop-727712 -n scheduled-stop-727712
helpers_test.go:269: (dbg) Run: kubectl --context scheduled-stop-727712 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:280: non-running pods: coredns-66bc5c9577-tsptq kindnet-jzkr4 kube-proxy-mfzjd storage-provisioner
helpers_test.go:282: ======> post-mortem[TestScheduledStopUnix]: describe non-running pods <======
helpers_test.go:285: (dbg) Run: kubectl --context scheduled-stop-727712 describe pod coredns-66bc5c9577-tsptq kindnet-jzkr4 kube-proxy-mfzjd storage-provisioner
helpers_test.go:285: (dbg) Non-zero exit: kubectl --context scheduled-stop-727712 describe pod coredns-66bc5c9577-tsptq kindnet-jzkr4 kube-proxy-mfzjd storage-provisioner: exit status 1 (118.622223ms)
** stderr **
Error from server (NotFound): pods "coredns-66bc5c9577-tsptq" not found
Error from server (NotFound): pods "kindnet-jzkr4" not found
Error from server (NotFound): pods "kube-proxy-mfzjd" not found
Error from server (NotFound): pods "storage-provisioner" not found
** /stderr **
helpers_test.go:287: kubectl --context scheduled-stop-727712 describe pod coredns-66bc5c9577-tsptq kindnet-jzkr4 kube-proxy-mfzjd storage-provisioner: exit status 1
helpers_test.go:175: Cleaning up "scheduled-stop-727712" profile ...
helpers_test.go:178: (dbg) Run: out/minikube-linux-arm64 delete -p scheduled-stop-727712
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p scheduled-stop-727712: (2.188646281s)
--- FAIL: TestScheduledStopUnix (36.43s)