=== RUN TestScheduledStopUnix
scheduled_stop_test.go:128: (dbg) Run: out/minikube-linux-arm64 start -p scheduled-stop-090251 --memory=3072 --driver=docker --container-runtime=docker
scheduled_stop_test.go:128: (dbg) Done: out/minikube-linux-arm64 start -p scheduled-stop-090251 --memory=3072 --driver=docker --container-runtime=docker: (31.963122908s)
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-090251 --schedule 5m
scheduled_stop_test.go:191: (dbg) Run: out/minikube-linux-arm64 status --format={{.TimeToStop}} -p scheduled-stop-090251 -n scheduled-stop-090251
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-090251 --schedule 15s
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:98: process 209469 running but should have been killed on reschedule of stop
panic.go:636: *** TestScheduledStopUnix FAILED at 2025-09-08 14:05:38.235490252 +0000 UTC m=+2305.404562900
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:223: ======> post-mortem[TestScheduledStopUnix]: network settings <======
helpers_test.go:230: HOST ENV snapshots: PROXY env: HTTP_PROXY="<empty>" HTTPS_PROXY="<empty>" NO_PROXY="<empty>"
helpers_test.go:238: ======> post-mortem[TestScheduledStopUnix]: docker inspect <======
helpers_test.go:239: (dbg) Run: docker inspect scheduled-stop-090251
helpers_test.go:243: (dbg) docker inspect scheduled-stop-090251:
-- stdout --
[
{
"Id": "d77af57d07af59f38009762436625e7e6b80a645dfe3f16cf16fdb10825bba9b",
"Created": "2025-09-08T14:05:10.676204834Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 206605,
"ExitCode": 0,
"Error": "",
"StartedAt": "2025-09-08T14:05:10.744009856Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:1a6e5b410fd9226cf2434621073598c7c01bccc994a53260ab0a0d834a0f1815",
"ResolvConfPath": "/var/lib/docker/containers/d77af57d07af59f38009762436625e7e6b80a645dfe3f16cf16fdb10825bba9b/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/d77af57d07af59f38009762436625e7e6b80a645dfe3f16cf16fdb10825bba9b/hostname",
"HostsPath": "/var/lib/docker/containers/d77af57d07af59f38009762436625e7e6b80a645dfe3f16cf16fdb10825bba9b/hosts",
"LogPath": "/var/lib/docker/containers/d77af57d07af59f38009762436625e7e6b80a645dfe3f16cf16fdb10825bba9b/d77af57d07af59f38009762436625e7e6b80a645dfe3f16cf16fdb10825bba9b-json.log",
"Name": "/scheduled-stop-090251",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"scheduled-stop-090251:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {}
},
"NetworkMode": "scheduled-stop-090251",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 3221225472,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 6442450944,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"ID": "d77af57d07af59f38009762436625e7e6b80a645dfe3f16cf16fdb10825bba9b",
"LowerDir": "/var/lib/docker/overlay2/ef1b0576c9b0880eb09ee1261c411deb854905f94f35e9cad2a947aca72b561b-init/diff:/var/lib/docker/overlay2/570c170e295ff2789664398ebc60cb792c7d3e094959c6d22ed3c06d39e2eff9/diff",
"MergedDir": "/var/lib/docker/overlay2/ef1b0576c9b0880eb09ee1261c411deb854905f94f35e9cad2a947aca72b561b/merged",
"UpperDir": "/var/lib/docker/overlay2/ef1b0576c9b0880eb09ee1261c411deb854905f94f35e9cad2a947aca72b561b/diff",
"WorkDir": "/var/lib/docker/overlay2/ef1b0576c9b0880eb09ee1261c411deb854905f94f35e9cad2a947aca72b561b/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "scheduled-stop-090251",
"Source": "/var/lib/docker/volumes/scheduled-stop-090251/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "scheduled-stop-090251",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "scheduled-stop-090251",
"name.minikube.sigs.k8s.io": "scheduled-stop-090251",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "21f146426c2356bf34581a5e605fea55a9416ad29b36268289c18e5e293b5ca0",
"SandboxKey": "/var/run/docker/netns/21f146426c23",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32970"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32971"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32974"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32972"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32973"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"scheduled-stop-090251": {
"IPAMConfig": {
"IPv4Address": "192.168.76.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "d2:a5:4c:4b:d3:38",
"DriverOpts": null,
"GwPriority": 0,
"NetworkID": "e8719ec6d49fac0e23164632cc37196a6ace8f604b259e066a87587e861f23d7",
"EndpointID": "bc8b35bbadcc81bfac8739383710d1e6e92c66b130e53cd3dffe16ecbea2ec79",
"Gateway": "192.168.76.1",
"IPAddress": "192.168.76.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"scheduled-stop-090251",
"d77af57d07af"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:247: (dbg) Run: out/minikube-linux-arm64 status --format={{.Host}} -p scheduled-stop-090251 -n scheduled-stop-090251
helpers_test.go:252: <<< TestScheduledStopUnix FAILED: start of post-mortem logs <<<
helpers_test.go:253: ======> post-mortem[TestScheduledStopUnix]: minikube logs <======
helpers_test.go:255: (dbg) Run: out/minikube-linux-arm64 -p scheduled-stop-090251 logs -n 25
helpers_test.go:260: TestScheduledStopUnix logs:
-- stdout --
==> Audit <==
┌─────────┬─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┬───────────────────────┬─────────┬─────────┬─────────────────────┬─────────────────────┐
│ COMMAND │ ARGS │ PROFILE │ USER │ VERSION │ START TIME │ END TIME │
├─────────┼─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┼───────────────────────┼─────────┼─────────┼─────────────────────┼─────────────────────┤
│ stop │ -p multinode-025632 │ multinode-025632 │ jenkins │ v1.36.0 │ 08 Sep 25 13:59 UTC │ 08 Sep 25 14:00 UTC │
│ start │ -p multinode-025632 --wait=true -v=5 --alsologtostderr │ multinode-025632 │ jenkins │ v1.36.0 │ 08 Sep 25 14:00 UTC │ 08 Sep 25 14:01 UTC │
│ node │ list -p multinode-025632 │ multinode-025632 │ jenkins │ v1.36.0 │ 08 Sep 25 14:01 UTC │ │
│ node │ multinode-025632 node delete m03 │ multinode-025632 │ jenkins │ v1.36.0 │ 08 Sep 25 14:01 UTC │ 08 Sep 25 14:01 UTC │
│ stop │ multinode-025632 stop │ multinode-025632 │ jenkins │ v1.36.0 │ 08 Sep 25 14:01 UTC │ 08 Sep 25 14:01 UTC │
│ start │ -p multinode-025632 --wait=true -v=5 --alsologtostderr --driver=docker --container-runtime=docker │ multinode-025632 │ jenkins │ v1.36.0 │ 08 Sep 25 14:01 UTC │ 08 Sep 25 14:02 UTC │
│ node │ list -p multinode-025632 │ multinode-025632 │ jenkins │ v1.36.0 │ 08 Sep 25 14:02 UTC │ │
│ start │ -p multinode-025632-m02 --driver=docker --container-runtime=docker │ multinode-025632-m02 │ jenkins │ v1.36.0 │ 08 Sep 25 14:02 UTC │ │
│ start │ -p multinode-025632-m03 --driver=docker --container-runtime=docker │ multinode-025632-m03 │ jenkins │ v1.36.0 │ 08 Sep 25 14:02 UTC │ 08 Sep 25 14:02 UTC │
│ node │ add -p multinode-025632 │ multinode-025632 │ jenkins │ v1.36.0 │ 08 Sep 25 14:02 UTC │ │
│ delete │ -p multinode-025632-m03 │ multinode-025632-m03 │ jenkins │ v1.36.0 │ 08 Sep 25 14:02 UTC │ 08 Sep 25 14:02 UTC │
│ delete │ -p multinode-025632 │ multinode-025632 │ jenkins │ v1.36.0 │ 08 Sep 25 14:02 UTC │ 08 Sep 25 14:03 UTC │
│ start │ -p test-preload-350973 --memory=3072 --alsologtostderr --wait=true --preload=false --driver=docker --container-runtime=docker --kubernetes-version=v1.32.0 │ test-preload-350973 │ jenkins │ v1.36.0 │ 08 Sep 25 14:03 UTC │ 08 Sep 25 14:03 UTC │
│ image │ test-preload-350973 image pull gcr.io/k8s-minikube/busybox │ test-preload-350973 │ jenkins │ v1.36.0 │ 08 Sep 25 14:03 UTC │ 08 Sep 25 14:03 UTC │
│ stop │ -p test-preload-350973 │ test-preload-350973 │ jenkins │ v1.36.0 │ 08 Sep 25 14:03 UTC │ 08 Sep 25 14:04 UTC │
│ start │ -p test-preload-350973 --memory=3072 --alsologtostderr -v=1 --wait=true --driver=docker --container-runtime=docker │ test-preload-350973 │ jenkins │ v1.36.0 │ 08 Sep 25 14:04 UTC │ 08 Sep 25 14:05 UTC │
│ image │ test-preload-350973 image list │ test-preload-350973 │ jenkins │ v1.36.0 │ 08 Sep 25 14:05 UTC │ 08 Sep 25 14:05 UTC │
│ delete │ -p test-preload-350973 │ test-preload-350973 │ jenkins │ v1.36.0 │ 08 Sep 25 14:05 UTC │ 08 Sep 25 14:05 UTC │
│ start │ -p scheduled-stop-090251 --memory=3072 --driver=docker --container-runtime=docker │ scheduled-stop-090251 │ jenkins │ v1.36.0 │ 08 Sep 25 14:05 UTC │ 08 Sep 25 14:05 UTC │
│ stop │ -p scheduled-stop-090251 --schedule 5m │ scheduled-stop-090251 │ jenkins │ v1.36.0 │ 08 Sep 25 14:05 UTC │ │
│ stop │ -p scheduled-stop-090251 --schedule 5m │ scheduled-stop-090251 │ jenkins │ v1.36.0 │ 08 Sep 25 14:05 UTC │ │
│ stop │ -p scheduled-stop-090251 --schedule 5m │ scheduled-stop-090251 │ jenkins │ v1.36.0 │ 08 Sep 25 14:05 UTC │ │
│ stop │ -p scheduled-stop-090251 --schedule 15s │ scheduled-stop-090251 │ jenkins │ v1.36.0 │ 08 Sep 25 14:05 UTC │ │
│ stop │ -p scheduled-stop-090251 --schedule 15s │ scheduled-stop-090251 │ jenkins │ v1.36.0 │ 08 Sep 25 14:05 UTC │ │
│ stop │ -p scheduled-stop-090251 --schedule 15s │ scheduled-stop-090251 │ jenkins │ v1.36.0 │ 08 Sep 25 14:05 UTC │ │
└─────────┴─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┴───────────────────────┴─────────┴─────────┴─────────────────────┴─────────────────────┘
==> Last Start <==
Log file created at: 2025/09/08 14:05:05
Running on machine: ip-172-31-29-130
Binary: Built with gc go1.24.6 for linux/arm64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0908 14:05:05.768051 206209 out.go:360] Setting OutFile to fd 1 ...
I0908 14:05:05.768165 206209 out.go:408] TERM=,COLORTERM=, which probably does not support color
I0908 14:05:05.768169 206209 out.go:374] Setting ErrFile to fd 2...
I0908 14:05:05.768173 206209 out.go:408] TERM=,COLORTERM=, which probably does not support color
I0908 14:05:05.768445 206209 root.go:338] Updating PATH: /home/jenkins/minikube-integration/21504-2320/.minikube/bin
I0908 14:05:05.769086 206209 out.go:368] Setting JSON to false
I0908 14:05:05.769919 206209 start.go:130] hostinfo: {"hostname":"ip-172-31-29-130","uptime":2854,"bootTime":1757337452,"procs":154,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1084-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"36adf542-ef4f-4e2d-a0c8-6868d1383ff9"}
I0908 14:05:05.769974 206209 start.go:140] virtualization:
I0908 14:05:05.773772 206209 out.go:179] * [scheduled-stop-090251] minikube v1.36.0 on Ubuntu 20.04 (arm64)
I0908 14:05:05.778220 206209 out.go:179] - MINIKUBE_LOCATION=21504
I0908 14:05:05.778390 206209 notify.go:220] Checking for updates...
I0908 14:05:05.784855 206209 out.go:179] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0908 14:05:05.788081 206209 out.go:179] - KUBECONFIG=/home/jenkins/minikube-integration/21504-2320/kubeconfig
I0908 14:05:05.791182 206209 out.go:179] - MINIKUBE_HOME=/home/jenkins/minikube-integration/21504-2320/.minikube
I0908 14:05:05.794325 206209 out.go:179] - MINIKUBE_BIN=out/minikube-linux-arm64
I0908 14:05:05.797428 206209 out.go:179] - MINIKUBE_FORCE_SYSTEMD=
I0908 14:05:05.800621 206209 driver.go:421] Setting default libvirt URI to qemu:///system
I0908 14:05:05.834503 206209 docker.go:123] docker version: linux-28.1.1:Docker Engine - Community
I0908 14:05:05.834643 206209 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0908 14:05:05.893353 206209 info.go:266] docker info: {ID:U5VK:ZNT5:35M3:FHLW:Q7TL:ELFX:BNAG:AV4T:UD2H:SK5L:SEJV:SJJL Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:23 OomKillDisable:true NGoroutines:42 SystemTime:2025-09-08 14:05:05.883708602 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1084-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214831104 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-29-130 Labels:[] ExperimentalBuild:false ServerVersion:28.1.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:05044ec0a9a75232cad458027ca83437aae3f4da Expected:} RuncCommit:{ID:v1.2.5-0-g59923ef Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx P
ath:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.23.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.35.1]] Warnings:<nil>}}
I0908 14:05:05.893480 206209 docker.go:318] overlay module found
I0908 14:05:05.896803 206209 out.go:179] * Using the docker driver based on user configuration
I0908 14:05:05.899742 206209 start.go:304] selected driver: docker
I0908 14:05:05.899760 206209 start.go:918] validating driver "docker" against <nil>
I0908 14:05:05.899773 206209 start.go:929] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0908 14:05:05.900506 206209 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0908 14:05:05.955090 206209 info.go:266] docker info: {ID:U5VK:ZNT5:35M3:FHLW:Q7TL:ELFX:BNAG:AV4T:UD2H:SK5L:SEJV:SJJL Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:23 OomKillDisable:true NGoroutines:42 SystemTime:2025-09-08 14:05:05.944858468 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1084-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214831104 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-29-130 Labels:[] ExperimentalBuild:false ServerVersion:28.1.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:05044ec0a9a75232cad458027ca83437aae3f4da Expected:} RuncCommit:{ID:v1.2.5-0-g59923ef Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx P
ath:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.23.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.35.1]] Warnings:<nil>}}
I0908 14:05:05.955230 206209 start_flags.go:327] no existing cluster config was found, will generate one from the flags
I0908 14:05:05.955457 206209 start_flags.go:974] Wait components to verify : map[apiserver:true system_pods:true]
I0908 14:05:05.958448 206209 out.go:179] * Using Docker driver with root privileges
I0908 14:05:05.961431 206209 cni.go:84] Creating CNI manager for ""
I0908 14:05:05.961491 206209 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0908 14:05:05.961498 206209 start_flags.go:336] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0908 14:05:05.961584 206209 start.go:348] cluster config:
{Name:scheduled-stop-090251 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.0 ClusterName:scheduled-stop-090251 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local
ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.34.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0908 14:05:05.966689 206209 out.go:179] * Starting "scheduled-stop-090251" primary control-plane node in "scheduled-stop-090251" cluster
I0908 14:05:05.969556 206209 cache.go:123] Beginning downloading kic base image for docker with docker
I0908 14:05:05.972609 206209 out.go:179] * Pulling base image v0.0.47-1756980985-21488 ...
I0908 14:05:05.975569 206209 preload.go:131] Checking if preload exists for k8s version v1.34.0 and runtime docker
I0908 14:05:05.975618 206209 preload.go:146] Found local preload: /home/jenkins/minikube-integration/21504-2320/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.0-docker-overlay2-arm64.tar.lz4
I0908 14:05:05.975626 206209 cache.go:58] Caching tarball of preloaded images
I0908 14:05:05.975671 206209 image.go:81] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 in local docker daemon
I0908 14:05:05.975727 206209 preload.go:172] Found /home/jenkins/minikube-integration/21504-2320/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.0-docker-overlay2-arm64.tar.lz4 in cache, skipping download
I0908 14:05:05.975736 206209 cache.go:61] Finished verifying existence of preloaded tar for v1.34.0 on docker
I0908 14:05:05.976071 206209 profile.go:143] Saving config to /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/config.json ...
I0908 14:05:05.976088 206209 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/config.json: {Name:mkdf4bb1a5f04d2861b05fbc9b986be2c4d77f1b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:05:05.995681 206209 image.go:100] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 in local docker daemon, skipping pull
I0908 14:05:05.995695 206209 cache.go:147] gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 exists in daemon, skipping load
I0908 14:05:05.995707 206209 cache.go:232] Successfully downloaded all kic artifacts
I0908 14:05:05.995740 206209 start.go:360] acquireMachinesLock for scheduled-stop-090251: {Name:mkfd5cdcc53046b1b45b6a24241c42f463801d72 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0908 14:05:05.995839 206209 start.go:364] duration metric: took 85.513µs to acquireMachinesLock for "scheduled-stop-090251"
I0908 14:05:05.995862 206209 start.go:93] Provisioning new machine with config: &{Name:scheduled-stop-090251 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.0 ClusterName:scheduled-stop-090251 Namespace:default APIServerHAVIP: A
PIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.34.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHA
uthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.34.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0908 14:05:05.995930 206209 start.go:125] createHost starting for "" (driver="docker")
I0908 14:05:05.999495 206209 out.go:252] * Creating docker container (CPUs=2, Memory=3072MB) ...
I0908 14:05:05.999739 206209 start.go:159] libmachine.API.Create for "scheduled-stop-090251" (driver="docker")
I0908 14:05:05.999773 206209 client.go:168] LocalClient.Create starting
I0908 14:05:05.999841 206209 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21504-2320/.minikube/certs/ca.pem
I0908 14:05:05.999874 206209 main.go:141] libmachine: Decoding PEM data...
I0908 14:05:05.999889 206209 main.go:141] libmachine: Parsing certificate...
I0908 14:05:05.999939 206209 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21504-2320/.minikube/certs/cert.pem
I0908 14:05:05.999963 206209 main.go:141] libmachine: Decoding PEM data...
I0908 14:05:05.999972 206209 main.go:141] libmachine: Parsing certificate...
I0908 14:05:06.000344 206209 cli_runner.go:164] Run: docker network inspect scheduled-stop-090251 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0908 14:05:06.018493 206209 cli_runner.go:211] docker network inspect scheduled-stop-090251 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0908 14:05:06.018563 206209 network_create.go:284] running [docker network inspect scheduled-stop-090251] to gather additional debugging logs...
I0908 14:05:06.018578 206209 cli_runner.go:164] Run: docker network inspect scheduled-stop-090251
W0908 14:05:06.049052 206209 cli_runner.go:211] docker network inspect scheduled-stop-090251 returned with exit code 1
I0908 14:05:06.049074 206209 network_create.go:287] error running [docker network inspect scheduled-stop-090251]: docker network inspect scheduled-stop-090251: exit status 1
stdout:
[]
stderr:
Error response from daemon: network scheduled-stop-090251 not found
I0908 14:05:06.049085 206209 network_create.go:289] output of [docker network inspect scheduled-stop-090251]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network scheduled-stop-090251 not found
** /stderr **
I0908 14:05:06.049219 206209 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0908 14:05:06.067243 206209 network.go:211] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-128e80606eed IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:ea:f3:36:ea:cc:a6} reservation:<nil>}
I0908 14:05:06.067485 206209 network.go:211] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-43116294d4d2 IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:ee:2d:cc:c9:e9:fc} reservation:<nil>}
I0908 14:05:06.067742 206209 network.go:211] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-2707d6e37252 IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:9e:af:7b:c1:17:e8} reservation:<nil>}
I0908 14:05:06.068064 206209 network.go:206] using free private subnet 192.168.76.0/24: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x4001a2e290}
I0908 14:05:06.068083 206209 network_create.go:124] attempt to create docker network scheduled-stop-090251 192.168.76.0/24 with gateway 192.168.76.1 and MTU of 1500 ...
I0908 14:05:06.068137 206209 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.76.0/24 --gateway=192.168.76.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=scheduled-stop-090251 scheduled-stop-090251
I0908 14:05:06.129590 206209 network_create.go:108] docker network scheduled-stop-090251 192.168.76.0/24 created
I0908 14:05:06.129612 206209 kic.go:121] calculated static IP "192.168.76.2" for the "scheduled-stop-090251" container
I0908 14:05:06.129695 206209 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0908 14:05:06.145985 206209 cli_runner.go:164] Run: docker volume create scheduled-stop-090251 --label name.minikube.sigs.k8s.io=scheduled-stop-090251 --label created_by.minikube.sigs.k8s.io=true
I0908 14:05:06.164543 206209 oci.go:103] Successfully created a docker volume scheduled-stop-090251
I0908 14:05:06.164633 206209 cli_runner.go:164] Run: docker run --rm --name scheduled-stop-090251-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-090251 --entrypoint /usr/bin/test -v scheduled-stop-090251:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 -d /var/lib
I0908 14:05:06.722344 206209 oci.go:107] Successfully prepared a docker volume scheduled-stop-090251
I0908 14:05:06.722380 206209 preload.go:131] Checking if preload exists for k8s version v1.34.0 and runtime docker
I0908 14:05:06.722411 206209 kic.go:194] Starting extracting preloaded images to volume ...
I0908 14:05:06.722475 206209 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/21504-2320/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.0-docker-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-090251:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 -I lz4 -xf /preloaded.tar -C /extractDir
I0908 14:05:10.609158 206209 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/21504-2320/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.0-docker-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-090251:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 -I lz4 -xf /preloaded.tar -C /extractDir: (3.886648123s)
I0908 14:05:10.609179 206209 kic.go:203] duration metric: took 3.886765471s to extract preloaded images to volume ...
W0908 14:05:10.609310 206209 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0908 14:05:10.609423 206209 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0908 14:05:10.661425 206209 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname scheduled-stop-090251 --name scheduled-stop-090251 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-090251 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=scheduled-stop-090251 --network scheduled-stop-090251 --ip 192.168.76.2 --volume scheduled-stop-090251:/var --security-opt apparmor=unconfined --memory=3072mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79
I0908 14:05:10.956989 206209 cli_runner.go:164] Run: docker container inspect scheduled-stop-090251 --format={{.State.Running}}
I0908 14:05:10.990546 206209 cli_runner.go:164] Run: docker container inspect scheduled-stop-090251 --format={{.State.Status}}
I0908 14:05:11.015384 206209 cli_runner.go:164] Run: docker exec scheduled-stop-090251 stat /var/lib/dpkg/alternatives/iptables
I0908 14:05:11.078453 206209 oci.go:144] the created container "scheduled-stop-090251" has a running status.
I0908 14:05:11.078494 206209 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/21504-2320/.minikube/machines/scheduled-stop-090251/id_rsa...
I0908 14:05:11.403745 206209 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/21504-2320/.minikube/machines/scheduled-stop-090251/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0908 14:05:11.425633 206209 cli_runner.go:164] Run: docker container inspect scheduled-stop-090251 --format={{.State.Status}}
I0908 14:05:11.452020 206209 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0908 14:05:11.452031 206209 kic_runner.go:114] Args: [docker exec --privileged scheduled-stop-090251 chown docker:docker /home/docker/.ssh/authorized_keys]
I0908 14:05:11.521668 206209 cli_runner.go:164] Run: docker container inspect scheduled-stop-090251 --format={{.State.Status}}
I0908 14:05:11.548602 206209 machine.go:93] provisionDockerMachine start ...
I0908 14:05:11.548694 206209 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-090251
I0908 14:05:11.573588 206209 main.go:141] libmachine: Using SSH client type: native
I0908 14:05:11.573926 206209 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef840] 0x3f2000 <nil> [] 0s} 127.0.0.1 32970 <nil> <nil>}
I0908 14:05:11.573934 206209 main.go:141] libmachine: About to run SSH command:
hostname
I0908 14:05:11.746107 206209 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-090251
I0908 14:05:11.746121 206209 ubuntu.go:182] provisioning hostname "scheduled-stop-090251"
I0908 14:05:11.746190 206209 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-090251
I0908 14:05:11.767197 206209 main.go:141] libmachine: Using SSH client type: native
I0908 14:05:11.767490 206209 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef840] 0x3f2000 <nil> [] 0s} 127.0.0.1 32970 <nil> <nil>}
I0908 14:05:11.767499 206209 main.go:141] libmachine: About to run SSH command:
sudo hostname scheduled-stop-090251 && echo "scheduled-stop-090251" | sudo tee /etc/hostname
I0908 14:05:11.908412 206209 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-090251
I0908 14:05:11.908498 206209 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-090251
I0908 14:05:11.931526 206209 main.go:141] libmachine: Using SSH client type: native
I0908 14:05:11.931835 206209 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef840] 0x3f2000 <nil> [] 0s} 127.0.0.1 32970 <nil> <nil>}
I0908 14:05:11.931850 206209 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sscheduled-stop-090251' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 scheduled-stop-090251/g' /etc/hosts;
else
echo '127.0.1.1 scheduled-stop-090251' | sudo tee -a /etc/hosts;
fi
fi
I0908 14:05:12.059108 206209 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0908 14:05:12.059124 206209 ubuntu.go:188] set auth options {CertDir:/home/jenkins/minikube-integration/21504-2320/.minikube CaCertPath:/home/jenkins/minikube-integration/21504-2320/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/21504-2320/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/21504-2320/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/21504-2320/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/21504-2320/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/21504-2320/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/21504-2320/.minikube}
I0908 14:05:12.059145 206209 ubuntu.go:190] setting up certificates
I0908 14:05:12.059154 206209 provision.go:84] configureAuth start
I0908 14:05:12.059213 206209 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-090251
I0908 14:05:12.083246 206209 provision.go:143] copyHostCerts
I0908 14:05:12.083307 206209 exec_runner.go:144] found /home/jenkins/minikube-integration/21504-2320/.minikube/ca.pem, removing ...
I0908 14:05:12.083315 206209 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21504-2320/.minikube/ca.pem
I0908 14:05:12.083396 206209 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21504-2320/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/21504-2320/.minikube/ca.pem (1082 bytes)
I0908 14:05:12.083490 206209 exec_runner.go:144] found /home/jenkins/minikube-integration/21504-2320/.minikube/cert.pem, removing ...
I0908 14:05:12.083494 206209 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21504-2320/.minikube/cert.pem
I0908 14:05:12.083518 206209 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21504-2320/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/21504-2320/.minikube/cert.pem (1123 bytes)
I0908 14:05:12.083573 206209 exec_runner.go:144] found /home/jenkins/minikube-integration/21504-2320/.minikube/key.pem, removing ...
I0908 14:05:12.083576 206209 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21504-2320/.minikube/key.pem
I0908 14:05:12.083597 206209 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21504-2320/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/21504-2320/.minikube/key.pem (1675 bytes)
I0908 14:05:12.083648 206209 provision.go:117] generating server cert: /home/jenkins/minikube-integration/21504-2320/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/21504-2320/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/21504-2320/.minikube/certs/ca-key.pem org=jenkins.scheduled-stop-090251 san=[127.0.0.1 192.168.76.2 localhost minikube scheduled-stop-090251]
I0908 14:05:13.067924 206209 provision.go:177] copyRemoteCerts
I0908 14:05:13.067982 206209 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0908 14:05:13.068025 206209 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-090251
I0908 14:05:13.086081 206209 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/21504-2320/.minikube/machines/scheduled-stop-090251/id_rsa Username:docker}
I0908 14:05:13.179659 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
I0908 14:05:13.205899 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/machines/server.pem --> /etc/docker/server.pem (1233 bytes)
I0908 14:05:13.231403 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0908 14:05:13.255847 206209 provision.go:87] duration metric: took 1.196669961s to configureAuth
I0908 14:05:13.255864 206209 ubuntu.go:206] setting minikube options for container-runtime
I0908 14:05:13.256061 206209 config.go:182] Loaded profile config "scheduled-stop-090251": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.34.0
I0908 14:05:13.256114 206209 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-090251
I0908 14:05:13.273295 206209 main.go:141] libmachine: Using SSH client type: native
I0908 14:05:13.273602 206209 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef840] 0x3f2000 <nil> [] 0s} 127.0.0.1 32970 <nil> <nil>}
I0908 14:05:13.273609 206209 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0908 14:05:13.399052 206209 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0908 14:05:13.399064 206209 ubuntu.go:71] root file system type: overlay
I0908 14:05:13.399167 206209 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0908 14:05:13.399230 206209 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-090251
I0908 14:05:13.417135 206209 main.go:141] libmachine: Using SSH client type: native
I0908 14:05:13.417420 206209 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef840] 0x3f2000 <nil> [] 0s} 127.0.0.1 32970 <nil> <nil>}
I0908 14:05:13.417500 206209 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network-online.target nss-lookup.target docker.socket firewalld.service containerd.service time-set.target
Wants=network-online.target containerd.service
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=always
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 \
-H fd:// --containerd=/run/containerd/containerd.sock \
-H unix:///var/run/docker.sock \
--default-ulimit=nofile=1048576:1048576 \
--tlsverify \
--tlscacert /etc/docker/ca.pem \
--tlscert /etc/docker/server.pem \
--tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0908 14:05:13.559210 206209 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network-online.target nss-lookup.target docker.socket firewalld.service containerd.service time-set.target
Wants=network-online.target containerd.service
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=always
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H fd:// --containerd=/run/containerd/containerd.sock -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
I0908 14:05:13.559286 206209 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-090251
I0908 14:05:13.577953 206209 main.go:141] libmachine: Using SSH client type: native
I0908 14:05:13.578246 206209 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef840] 0x3f2000 <nil> [] 0s} 127.0.0.1 32970 <nil> <nil>}
I0908 14:05:13.578261 206209 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0908 14:05:14.395129 206209 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2025-09-03 20:57:01.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2025-09-08 14:05:13.554681707 +0000
@@ -9,23 +9,34 @@
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
Restart=always
+
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H fd:// --containerd=/run/containerd/containerd.sock -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
+
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0908 14:05:14.395155 206209 machine.go:96] duration metric: took 2.846540901s to provisionDockerMachine
I0908 14:05:14.395165 206209 client.go:171] duration metric: took 8.395386436s to LocalClient.Create
I0908 14:05:14.395183 206209 start.go:167] duration metric: took 8.395445224s to libmachine.API.Create "scheduled-stop-090251"
I0908 14:05:14.395189 206209 start.go:293] postStartSetup for "scheduled-stop-090251" (driver="docker")
I0908 14:05:14.395199 206209 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0908 14:05:14.395258 206209 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0908 14:05:14.395304 206209 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-090251
I0908 14:05:14.412412 206209 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/21504-2320/.minikube/machines/scheduled-stop-090251/id_rsa Username:docker}
I0908 14:05:14.503919 206209 ssh_runner.go:195] Run: cat /etc/os-release
I0908 14:05:14.507044 206209 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0908 14:05:14.507067 206209 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0908 14:05:14.507076 206209 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0908 14:05:14.507082 206209 info.go:137] Remote host: Ubuntu 22.04.5 LTS
I0908 14:05:14.507091 206209 filesync.go:126] Scanning /home/jenkins/minikube-integration/21504-2320/.minikube/addons for local assets ...
I0908 14:05:14.507147 206209 filesync.go:126] Scanning /home/jenkins/minikube-integration/21504-2320/.minikube/files for local assets ...
I0908 14:05:14.507228 206209 filesync.go:149] local asset: /home/jenkins/minikube-integration/21504-2320/.minikube/files/etc/ssl/certs/41202.pem -> 41202.pem in /etc/ssl/certs
I0908 14:05:14.507328 206209 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I0908 14:05:14.515581 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/files/etc/ssl/certs/41202.pem --> /etc/ssl/certs/41202.pem (1708 bytes)
I0908 14:05:14.539977 206209 start.go:296] duration metric: took 144.77485ms for postStartSetup
I0908 14:05:14.540351 206209 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-090251
I0908 14:05:14.557241 206209 profile.go:143] Saving config to /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/config.json ...
I0908 14:05:14.557501 206209 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0908 14:05:14.557545 206209 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-090251
I0908 14:05:14.574147 206209 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/21504-2320/.minikube/machines/scheduled-stop-090251/id_rsa Username:docker}
I0908 14:05:14.659547 206209 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0908 14:05:14.664115 206209 start.go:128] duration metric: took 8.668172627s to createHost
I0908 14:05:14.664129 206209 start.go:83] releasing machines lock for "scheduled-stop-090251", held for 8.668283139s
I0908 14:05:14.664205 206209 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-090251
I0908 14:05:14.681453 206209 ssh_runner.go:195] Run: cat /version.json
I0908 14:05:14.681489 206209 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0908 14:05:14.681494 206209 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-090251
I0908 14:05:14.681557 206209 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-090251
I0908 14:05:14.704835 206209 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/21504-2320/.minikube/machines/scheduled-stop-090251/id_rsa Username:docker}
I0908 14:05:14.717026 206209 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/21504-2320/.minikube/machines/scheduled-stop-090251/id_rsa Username:docker}
I0908 14:05:14.790897 206209 ssh_runner.go:195] Run: systemctl --version
I0908 14:05:14.918254 206209 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0908 14:05:14.922545 206209 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0908 14:05:14.948297 206209 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0908 14:05:14.948372 206209 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0908 14:05:14.979671 206209 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0908 14:05:14.979688 206209 start.go:495] detecting cgroup driver to use...
I0908 14:05:14.979719 206209 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0908 14:05:14.979813 206209 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0908 14:05:14.996027 206209 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10.1"|' /etc/containerd/config.toml"
I0908 14:05:15.006455 206209 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0908 14:05:15.016646 206209 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0908 14:05:15.016718 206209 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0908 14:05:15.040279 206209 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0908 14:05:15.060612 206209 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0908 14:05:15.081178 206209 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0908 14:05:15.092218 206209 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0908 14:05:15.102254 206209 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0908 14:05:15.113459 206209 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0908 14:05:15.124146 206209 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0908 14:05:15.135690 206209 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0908 14:05:15.145002 206209 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0908 14:05:15.154050 206209 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0908 14:05:15.247148 206209 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0908 14:05:15.336485 206209 start.go:495] detecting cgroup driver to use...
I0908 14:05:15.336522 206209 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0908 14:05:15.336580 206209 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0908 14:05:15.350894 206209 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I0908 14:05:15.363509 206209 ssh_runner.go:195] Run: sudo systemctl stop -f containerd
I0908 14:05:15.397074 206209 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I0908 14:05:15.408777 206209 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0908 14:05:15.421398 206209 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0908 14:05:15.438304 206209 ssh_runner.go:195] Run: which cri-dockerd
I0908 14:05:15.441806 206209 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0908 14:05:15.450662 206209 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (192 bytes)
I0908 14:05:15.469029 206209 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0908 14:05:15.564603 206209 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0908 14:05:15.652162 206209 docker.go:575] configuring docker to use "cgroupfs" as cgroup driver...
I0908 14:05:15.652245 206209 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0908 14:05:15.670723 206209 ssh_runner.go:195] Run: sudo systemctl reset-failed docker
I0908 14:05:15.682684 206209 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0908 14:05:15.775268 206209 ssh_runner.go:195] Run: sudo systemctl restart docker
I0908 14:05:16.161576 206209 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
I0908 14:05:16.173839 206209 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0908 14:05:16.186716 206209 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0908 14:05:16.199097 206209 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0908 14:05:16.294646 206209 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0908 14:05:16.391593 206209 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0908 14:05:16.483355 206209 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0908 14:05:16.497807 206209 ssh_runner.go:195] Run: sudo systemctl reset-failed cri-docker.service
I0908 14:05:16.509621 206209 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0908 14:05:16.602517 206209 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0908 14:05:16.677337 206209 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0908 14:05:16.691148 206209 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0908 14:05:16.691213 206209 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0908 14:05:16.695206 206209 start.go:563] Will wait 60s for crictl version
I0908 14:05:16.695260 206209 ssh_runner.go:195] Run: which crictl
I0908 14:05:16.698861 206209 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0908 14:05:16.747539 206209 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 28.4.0
RuntimeApiVersion: v1
I0908 14:05:16.747595 206209 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0908 14:05:16.769651 206209 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0908 14:05:16.795492 206209 out.go:252] * Preparing Kubernetes v1.34.0 on Docker 28.4.0 ...
I0908 14:05:16.795576 206209 cli_runner.go:164] Run: docker network inspect scheduled-stop-090251 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0908 14:05:16.811506 206209 ssh_runner.go:195] Run: grep 192.168.76.1 host.minikube.internal$ /etc/hosts
I0908 14:05:16.815195 206209 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.76.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0908 14:05:16.826066 206209 kubeadm.go:875] updating cluster {Name:scheduled-stop-090251 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.0 ClusterName:scheduled-stop-090251 Namespace:default APIServerHAVIP: APIServerName:miniku
beCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock
: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0908 14:05:16.826166 206209 preload.go:131] Checking if preload exists for k8s version v1.34.0 and runtime docker
I0908 14:05:16.826218 206209 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0908 14:05:16.844557 206209 docker.go:691] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.34.0
registry.k8s.io/kube-controller-manager:v1.34.0
registry.k8s.io/kube-scheduler:v1.34.0
registry.k8s.io/kube-proxy:v1.34.0
registry.k8s.io/etcd:3.6.4-0
registry.k8s.io/pause:3.10.1
registry.k8s.io/coredns/coredns:v1.12.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0908 14:05:16.844571 206209 docker.go:621] Images already preloaded, skipping extraction
I0908 14:05:16.844639 206209 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0908 14:05:16.861616 206209 docker.go:691] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.34.0
registry.k8s.io/kube-controller-manager:v1.34.0
registry.k8s.io/kube-scheduler:v1.34.0
registry.k8s.io/kube-proxy:v1.34.0
registry.k8s.io/etcd:3.6.4-0
registry.k8s.io/pause:3.10.1
registry.k8s.io/coredns/coredns:v1.12.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0908 14:05:16.861629 206209 cache_images.go:85] Images are preloaded, skipping loading
I0908 14:05:16.861640 206209 kubeadm.go:926] updating node { 192.168.76.2 8443 v1.34.0 docker true true} ...
I0908 14:05:16.861729 206209 kubeadm.go:938] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.34.0/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=scheduled-stop-090251 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.76.2
[Install]
config:
{KubernetesVersion:v1.34.0 ClusterName:scheduled-stop-090251 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0908 14:05:16.861796 206209 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0908 14:05:16.910230 206209 cni.go:84] Creating CNI manager for ""
I0908 14:05:16.910253 206209 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0908 14:05:16.910264 206209 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0908 14:05:16.910284 206209 kubeadm.go:189] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.76.2 APIServerPort:8443 KubernetesVersion:v1.34.0 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:scheduled-stop-090251 NodeName:scheduled-stop-090251 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.76.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.76.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPo
dPath:/etc/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0908 14:05:16.910424 206209 kubeadm.go:195] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta4
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.76.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "scheduled-stop-090251"
kubeletExtraArgs:
- name: "node-ip"
value: "192.168.76.2"
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta4
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.76.2"]
extraArgs:
- name: "enable-admission-plugins"
value: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
- name: "allocate-node-cidrs"
value: "true"
- name: "leader-elect"
value: "false"
scheduler:
extraArgs:
- name: "leader-elect"
value: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
kubernetesVersion: v1.34.0
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0908 14:05:16.910492 206209 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.34.0
I0908 14:05:16.919577 206209 binaries.go:44] Found k8s binaries, skipping transfer
I0908 14:05:16.919637 206209 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0908 14:05:16.928447 206209 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (320 bytes)
I0908 14:05:16.947245 206209 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0908 14:05:16.965590 206209 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2222 bytes)
I0908 14:05:16.984781 206209 ssh_runner.go:195] Run: grep 192.168.76.2 control-plane.minikube.internal$ /etc/hosts
I0908 14:05:16.988512 206209 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.76.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0908 14:05:16.999993 206209 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0908 14:05:17.096017 206209 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0908 14:05:17.110290 206209 certs.go:68] Setting up /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251 for IP: 192.168.76.2
I0908 14:05:17.110311 206209 certs.go:194] generating shared ca certs ...
I0908 14:05:17.110326 206209 certs.go:226] acquiring lock for ca certs: {Name:mk0021cd008d807b29f57862e5444612344fe341 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:05:17.110508 206209 certs.go:235] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/21504-2320/.minikube/ca.key
I0908 14:05:17.110556 206209 certs.go:235] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/21504-2320/.minikube/proxy-client-ca.key
I0908 14:05:17.110562 206209 certs.go:256] generating profile certs ...
I0908 14:05:17.110625 206209 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/client.key
I0908 14:05:17.110636 206209 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/client.crt with IP's: []
I0908 14:05:17.720157 206209 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/client.crt ...
I0908 14:05:17.720174 206209 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/client.crt: {Name:mke2435ce5f36c7a5f690b94048ce62a8a6a9d53 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:05:17.720391 206209 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/client.key ...
I0908 14:05:17.720400 206209 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/client.key: {Name:mkba7b9cc3132cfd3727ac0cb363fb13635fd333 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:05:17.720493 206209 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/apiserver.key.e19f5597
I0908 14:05:17.720506 206209 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/apiserver.crt.e19f5597 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.76.2]
I0908 14:05:18.382673 206209 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/apiserver.crt.e19f5597 ...
I0908 14:05:18.382690 206209 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/apiserver.crt.e19f5597: {Name:mk9e10f10b87349f96e758cb03a902117d82e002 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:05:18.382901 206209 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/apiserver.key.e19f5597 ...
I0908 14:05:18.382910 206209 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/apiserver.key.e19f5597: {Name:mke1ea6745c6c601fb2c2203d4a511fcbf61234f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:05:18.383007 206209 certs.go:381] copying /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/apiserver.crt.e19f5597 -> /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/apiserver.crt
I0908 14:05:18.383083 206209 certs.go:385] copying /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/apiserver.key.e19f5597 -> /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/apiserver.key
I0908 14:05:18.383139 206209 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/proxy-client.key
I0908 14:05:18.383150 206209 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/proxy-client.crt with IP's: []
I0908 14:05:18.985131 206209 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/proxy-client.crt ...
I0908 14:05:18.985146 206209 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/proxy-client.crt: {Name:mkd80fa368b5af97ab6f603dcce4362a79ca3103 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:05:18.985340 206209 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/proxy-client.key ...
I0908 14:05:18.985347 206209 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/proxy-client.key: {Name:mkeddac85ddc295406d8cbeaf8c4880a9c78b6b4 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:05:18.985524 206209 certs.go:484] found cert: /home/jenkins/minikube-integration/21504-2320/.minikube/certs/4120.pem (1338 bytes)
W0908 14:05:18.985558 206209 certs.go:480] ignoring /home/jenkins/minikube-integration/21504-2320/.minikube/certs/4120_empty.pem, impossibly tiny 0 bytes
I0908 14:05:18.985565 206209 certs.go:484] found cert: /home/jenkins/minikube-integration/21504-2320/.minikube/certs/ca-key.pem (1679 bytes)
I0908 14:05:18.985588 206209 certs.go:484] found cert: /home/jenkins/minikube-integration/21504-2320/.minikube/certs/ca.pem (1082 bytes)
I0908 14:05:18.985612 206209 certs.go:484] found cert: /home/jenkins/minikube-integration/21504-2320/.minikube/certs/cert.pem (1123 bytes)
I0908 14:05:18.985632 206209 certs.go:484] found cert: /home/jenkins/minikube-integration/21504-2320/.minikube/certs/key.pem (1675 bytes)
I0908 14:05:18.985674 206209 certs.go:484] found cert: /home/jenkins/minikube-integration/21504-2320/.minikube/files/etc/ssl/certs/41202.pem (1708 bytes)
I0908 14:05:18.986264 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0908 14:05:19.013665 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I0908 14:05:19.041287 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0908 14:05:19.066701 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0908 14:05:19.091989 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1432 bytes)
I0908 14:05:19.117317 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0908 14:05:19.143786 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0908 14:05:19.169406 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/profiles/scheduled-stop-090251/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0908 14:05:19.194549 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/files/etc/ssl/certs/41202.pem --> /usr/share/ca-certificates/41202.pem (1708 bytes)
I0908 14:05:19.219926 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0908 14:05:19.244889 206209 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2320/.minikube/certs/4120.pem --> /usr/share/ca-certificates/4120.pem (1338 bytes)
I0908 14:05:19.270157 206209 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0908 14:05:19.289026 206209 ssh_runner.go:195] Run: openssl version
I0908 14:05:19.295589 206209 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0908 14:05:19.306029 206209 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0908 14:05:19.310708 206209 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 8 13:28 /usr/share/ca-certificates/minikubeCA.pem
I0908 14:05:19.310788 206209 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0908 14:05:19.318736 206209 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0908 14:05:19.328657 206209 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/4120.pem && ln -fs /usr/share/ca-certificates/4120.pem /etc/ssl/certs/4120.pem"
I0908 14:05:19.338565 206209 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/4120.pem
I0908 14:05:19.342238 206209 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 Sep 8 13:34 /usr/share/ca-certificates/4120.pem
I0908 14:05:19.342302 206209 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/4120.pem
I0908 14:05:19.350041 206209 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/4120.pem /etc/ssl/certs/51391683.0"
I0908 14:05:19.359801 206209 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/41202.pem && ln -fs /usr/share/ca-certificates/41202.pem /etc/ssl/certs/41202.pem"
I0908 14:05:19.369955 206209 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/41202.pem
I0908 14:05:19.378602 206209 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 Sep 8 13:34 /usr/share/ca-certificates/41202.pem
I0908 14:05:19.378660 206209 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/41202.pem
I0908 14:05:19.386148 206209 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/41202.pem /etc/ssl/certs/3ec20f2e.0"
I0908 14:05:19.396118 206209 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0908 14:05:19.400383 206209 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0908 14:05:19.400425 206209 kubeadm.go:392] StartCluster: {Name:scheduled-stop-090251 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.0 ClusterName:scheduled-stop-090251 Namespace:default APIServerHAVIP: APIServerName:minikubeC
A APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: S
SHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0908 14:05:19.400530 206209 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0908 14:05:19.424834 206209 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0908 14:05:19.439170 206209 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0908 14:05:19.448469 206209 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0908 14:05:19.448526 206209 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0908 14:05:19.457429 206209 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0908 14:05:19.457438 206209 kubeadm.go:157] found existing configuration files:
I0908 14:05:19.457488 206209 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0908 14:05:19.467170 206209 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0908 14:05:19.467225 206209 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0908 14:05:19.476349 206209 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0908 14:05:19.485308 206209 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0908 14:05:19.485365 206209 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0908 14:05:19.494656 206209 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0908 14:05:19.504008 206209 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0908 14:05:19.504080 206209 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0908 14:05:19.512743 206209 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0908 14:05:19.522140 206209 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0908 14:05:19.522198 206209 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0908 14:05:19.531264 206209 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.34.0:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0908 14:05:19.575373 206209 kubeadm.go:310] [init] Using Kubernetes version: v1.34.0
I0908 14:05:19.575571 206209 kubeadm.go:310] [preflight] Running pre-flight checks
I0908 14:05:19.596778 206209 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0908 14:05:19.596844 206209 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1084-aws[0m
I0908 14:05:19.596879 206209 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0908 14:05:19.596924 206209 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0908 14:05:19.596972 206209 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0908 14:05:19.597020 206209 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0908 14:05:19.597068 206209 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0908 14:05:19.597116 206209 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0908 14:05:19.597174 206209 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0908 14:05:19.597219 206209 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0908 14:05:19.597268 206209 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0908 14:05:19.597315 206209 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0908 14:05:19.663587 206209 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0908 14:05:19.663691 206209 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0908 14:05:19.663781 206209 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0908 14:05:19.682400 206209 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0908 14:05:19.689334 206209 out.go:252] - Generating certificates and keys ...
I0908 14:05:19.689424 206209 kubeadm.go:310] [certs] Using existing ca certificate authority
I0908 14:05:19.689488 206209 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0908 14:05:20.879092 206209 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0908 14:05:21.153566 206209 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0908 14:05:21.542985 206209 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0908 14:05:22.551967 206209 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0908 14:05:23.331151 206209 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0908 14:05:23.331295 206209 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [localhost scheduled-stop-090251] and IPs [192.168.76.2 127.0.0.1 ::1]
I0908 14:05:23.832518 206209 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0908 14:05:23.832826 206209 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [localhost scheduled-stop-090251] and IPs [192.168.76.2 127.0.0.1 ::1]
I0908 14:05:24.084919 206209 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0908 14:05:24.585887 206209 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0908 14:05:24.680438 206209 kubeadm.go:310] [certs] Generating "sa" key and public key
I0908 14:05:24.680565 206209 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0908 14:05:25.041455 206209 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0908 14:05:25.810223 206209 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0908 14:05:25.981768 206209 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0908 14:05:26.620759 206209 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0908 14:05:27.116685 206209 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0908 14:05:27.117451 206209 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0908 14:05:27.120120 206209 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0908 14:05:27.123576 206209 out.go:252] - Booting up control plane ...
I0908 14:05:27.123717 206209 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0908 14:05:27.123814 206209 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0908 14:05:27.123894 206209 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0908 14:05:27.137903 206209 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0908 14:05:27.138006 206209 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/instance-config.yaml"
I0908 14:05:27.145181 206209 kubeadm.go:310] [patches] Applied patch of type "application/strategic-merge-patch+json" to target "kubeletconfiguration"
I0908 14:05:27.145456 206209 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0908 14:05:27.145500 206209 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0908 14:05:27.251252 206209 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0908 14:05:27.251366 206209 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0908 14:05:28.251060 206209 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 1.00169163s
I0908 14:05:28.254605 206209 kubeadm.go:310] [control-plane-check] Waiting for healthy control plane components. This can take up to 4m0s
I0908 14:05:28.254694 206209 kubeadm.go:310] [control-plane-check] Checking kube-apiserver at https://192.168.76.2:8443/livez
I0908 14:05:28.254833 206209 kubeadm.go:310] [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz
I0908 14:05:28.254927 206209 kubeadm.go:310] [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez
I0908 14:05:32.051395 206209 kubeadm.go:310] [control-plane-check] kube-controller-manager is healthy after 3.795784703s
I0908 14:05:33.214181 206209 kubeadm.go:310] [control-plane-check] kube-scheduler is healthy after 4.959554331s
I0908 14:05:34.757348 206209 kubeadm.go:310] [control-plane-check] kube-apiserver is healthy after 6.502174073s
I0908 14:05:34.778715 206209 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0908 14:05:34.792996 206209 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0908 14:05:34.809535 206209 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0908 14:05:34.809767 206209 kubeadm.go:310] [mark-control-plane] Marking the node scheduled-stop-090251 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0908 14:05:34.822862 206209 kubeadm.go:310] [bootstrap-token] Using token: usvw21.x5azmvbxk8j4zltx
I0908 14:05:34.825630 206209 out.go:252] - Configuring RBAC rules ...
I0908 14:05:34.825747 206209 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0908 14:05:34.833866 206209 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0908 14:05:34.845044 206209 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0908 14:05:34.853608 206209 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0908 14:05:34.858915 206209 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0908 14:05:34.864703 206209 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0908 14:05:35.170119 206209 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0908 14:05:35.600051 206209 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0908 14:05:36.167194 206209 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0908 14:05:36.168492 206209 kubeadm.go:310]
I0908 14:05:36.168561 206209 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0908 14:05:36.168565 206209 kubeadm.go:310]
I0908 14:05:36.168642 206209 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0908 14:05:36.168645 206209 kubeadm.go:310]
I0908 14:05:36.168670 206209 kubeadm.go:310] mkdir -p $HOME/.kube
I0908 14:05:36.168815 206209 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0908 14:05:36.168873 206209 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0908 14:05:36.168876 206209 kubeadm.go:310]
I0908 14:05:36.168930 206209 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0908 14:05:36.168933 206209 kubeadm.go:310]
I0908 14:05:36.168980 206209 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0908 14:05:36.168984 206209 kubeadm.go:310]
I0908 14:05:36.169035 206209 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0908 14:05:36.169110 206209 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0908 14:05:36.169178 206209 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0908 14:05:36.169181 206209 kubeadm.go:310]
I0908 14:05:36.169265 206209 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0908 14:05:36.169342 206209 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0908 14:05:36.169345 206209 kubeadm.go:310]
I0908 14:05:36.169429 206209 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token usvw21.x5azmvbxk8j4zltx \
I0908 14:05:36.169533 206209 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:7412d166a96d82f93918d03ec09dbc9aa58e761d95c1602d1d02b657bf9086f8 \
I0908 14:05:36.169553 206209 kubeadm.go:310] --control-plane
I0908 14:05:36.169557 206209 kubeadm.go:310]
I0908 14:05:36.169642 206209 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0908 14:05:36.169645 206209 kubeadm.go:310]
I0908 14:05:36.169727 206209 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token usvw21.x5azmvbxk8j4zltx \
I0908 14:05:36.169829 206209 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:7412d166a96d82f93918d03ec09dbc9aa58e761d95c1602d1d02b657bf9086f8
I0908 14:05:36.174978 206209 kubeadm.go:310] [WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2
I0908 14:05:36.175198 206209 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1084-aws\n", err: exit status 1
I0908 14:05:36.175302 206209 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0908 14:05:36.175319 206209 cni.go:84] Creating CNI manager for ""
I0908 14:05:36.175331 206209 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0908 14:05:36.178447 206209 out.go:179] * Configuring bridge CNI (Container Networking Interface) ...
I0908 14:05:36.181236 206209 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0908 14:05:36.190148 206209 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0908 14:05:36.208718 206209 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0908 14:05:36.208794 206209 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.0/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0908 14:05:36.208827 206209 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes scheduled-stop-090251 minikube.k8s.io/updated_at=2025_09_08T14_05_36_0700 minikube.k8s.io/version=v1.36.0 minikube.k8s.io/commit=4237956cfce90d4ab760d817400bd4c89cad50d6 minikube.k8s.io/name=scheduled-stop-090251 minikube.k8s.io/primary=true
I0908 14:05:36.225442 206209 ops.go:34] apiserver oom_adj: -16
I0908 14:05:36.327757 206209 kubeadm.go:1105] duration metric: took 119.034105ms to wait for elevateKubeSystemPrivileges
I0908 14:05:36.356269 206209 kubeadm.go:394] duration metric: took 16.95583859s to StartCluster
I0908 14:05:36.356292 206209 settings.go:142] acquiring lock: {Name:mk6466197568b454a152c74d528145484fbc55b6 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:05:36.356358 206209 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/21504-2320/kubeconfig
I0908 14:05:36.357032 206209 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2320/kubeconfig: {Name:mk789dd53c90d53c14c1d24f2d2a926103526048 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:05:36.357253 206209 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.34.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0908 14:05:36.357269 206209 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0908 14:05:36.357601 206209 config.go:182] Loaded profile config "scheduled-stop-090251": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.34.0
I0908 14:05:36.357640 206209 addons.go:511] enable addons start: toEnable=map[ambassador:false amd-gpu-device-plugin:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubetail:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:false volcano:false volumesnapshots:false yakd:false]
I0908 14:05:36.357713 206209 addons.go:69] Setting storage-provisioner=true in profile "scheduled-stop-090251"
I0908 14:05:36.357720 206209 addons.go:69] Setting default-storageclass=true in profile "scheduled-stop-090251"
I0908 14:05:36.357726 206209 addons.go:238] Setting addon storage-provisioner=true in "scheduled-stop-090251"
I0908 14:05:36.357734 206209 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "scheduled-stop-090251"
I0908 14:05:36.357747 206209 host.go:66] Checking if "scheduled-stop-090251" exists ...
I0908 14:05:36.358097 206209 cli_runner.go:164] Run: docker container inspect scheduled-stop-090251 --format={{.State.Status}}
I0908 14:05:36.358241 206209 cli_runner.go:164] Run: docker container inspect scheduled-stop-090251 --format={{.State.Status}}
I0908 14:05:36.363276 206209 out.go:179] * Verifying Kubernetes components...
I0908 14:05:36.366908 206209 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0908 14:05:36.398153 206209 addons.go:238] Setting addon default-storageclass=true in "scheduled-stop-090251"
I0908 14:05:36.398180 206209 host.go:66] Checking if "scheduled-stop-090251" exists ...
I0908 14:05:36.398623 206209 cli_runner.go:164] Run: docker container inspect scheduled-stop-090251 --format={{.State.Status}}
I0908 14:05:36.406003 206209 out.go:179] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0908 14:05:36.408850 206209 addons.go:435] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0908 14:05:36.408861 206209 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0908 14:05:36.408946 206209 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-090251
I0908 14:05:36.438645 206209 addons.go:435] installing /etc/kubernetes/addons/storageclass.yaml
I0908 14:05:36.438657 206209 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0908 14:05:36.438729 206209 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-090251
I0908 14:05:36.460014 206209 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/21504-2320/.minikube/machines/scheduled-stop-090251/id_rsa Username:docker}
I0908 14:05:36.472356 206209 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/21504-2320/.minikube/machines/scheduled-stop-090251/id_rsa Username:docker}
I0908 14:05:36.600023 206209 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.34.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.76.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.34.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0908 14:05:36.600114 206209 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0908 14:05:36.772678 206209 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.34.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0908 14:05:36.781339 206209 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.34.0/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0908 14:05:37.088633 206209 api_server.go:52] waiting for apiserver process to appear ...
I0908 14:05:37.088693 206209 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0908 14:05:37.088800 206209 start.go:976] {"host.minikube.internal": 192.168.76.1} host record injected into CoreDNS's ConfigMap
I0908 14:05:37.350492 206209 api_server.go:72] duration metric: took 993.200014ms to wait for apiserver process to appear ...
I0908 14:05:37.350503 206209 api_server.go:88] waiting for apiserver healthz status ...
I0908 14:05:37.350518 206209 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
I0908 14:05:37.362926 206209 api_server.go:279] https://192.168.76.2:8443/healthz returned 200:
ok
I0908 14:05:37.365682 206209 api_server.go:141] control plane version: v1.34.0
I0908 14:05:37.365699 206209 api_server.go:131] duration metric: took 15.190209ms to wait for apiserver health ...
I0908 14:05:37.365706 206209 system_pods.go:43] waiting for kube-system pods to appear ...
I0908 14:05:37.369754 206209 out.go:179] * Enabled addons: storage-provisioner, default-storageclass
I0908 14:05:37.371065 206209 system_pods.go:59] 5 kube-system pods found
I0908 14:05:37.371088 206209 system_pods.go:61] "etcd-scheduled-stop-090251" [f314ed1c-8434-47bd-97a2-7306a37b907c] Running / Ready:ContainersNotReady (containers with unready status: [etcd]) / ContainersReady:ContainersNotReady (containers with unready status: [etcd])
I0908 14:05:37.371095 206209 system_pods.go:61] "kube-apiserver-scheduled-stop-090251" [636679bd-6a19-492c-9adb-07621a8057d3] Running / Ready:ContainersNotReady (containers with unready status: [kube-apiserver]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-apiserver])
I0908 14:05:37.371102 206209 system_pods.go:61] "kube-controller-manager-scheduled-stop-090251" [f745426e-be5c-41f2-932d-70a63d628d1f] Running / Ready:ContainersNotReady (containers with unready status: [kube-controller-manager]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-controller-manager])
I0908 14:05:37.371107 206209 system_pods.go:61] "kube-scheduler-scheduled-stop-090251" [488b6e6b-8390-47e6-aeb6-470f0946f73a] Running / Ready:ContainersNotReady (containers with unready status: [kube-scheduler]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-scheduler])
I0908 14:05:37.371117 206209 system_pods.go:61] "storage-provisioner" [e2fc7477-2fb4-4f90-8ed3-967c5e40ca66] Pending: PodScheduled:Unschedulable (0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. no new claims to deallocate, preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.)
I0908 14:05:37.371122 206209 system_pods.go:74] duration metric: took 5.410927ms to wait for pod list to return data ...
I0908 14:05:37.371132 206209 kubeadm.go:578] duration metric: took 1.01384421s to wait for: map[apiserver:true system_pods:true]
I0908 14:05:37.371143 206209 node_conditions.go:102] verifying NodePressure condition ...
I0908 14:05:37.372643 206209 addons.go:514] duration metric: took 1.014985398s for enable addons: enabled=[storage-provisioner default-storageclass]
I0908 14:05:37.374030 206209 node_conditions.go:122] node storage ephemeral capacity is 203034800Ki
I0908 14:05:37.374048 206209 node_conditions.go:123] node cpu capacity is 2
I0908 14:05:37.374063 206209 node_conditions.go:105] duration metric: took 2.916239ms to run NodePressure ...
I0908 14:05:37.374075 206209 start.go:241] waiting for startup goroutines ...
I0908 14:05:37.592892 206209 kapi.go:214] "coredns" deployment in "kube-system" namespace and "scheduled-stop-090251" context rescaled to 1 replicas
I0908 14:05:37.592914 206209 start.go:246] waiting for cluster config update ...
I0908 14:05:37.592936 206209 start.go:255] writing updated cluster config ...
I0908 14:05:37.593217 206209 ssh_runner.go:195] Run: rm -f paused
I0908 14:05:37.653839 206209 start.go:617] kubectl: 1.33.2, cluster: 1.34.0 (minor skew: 1)
I0908 14:05:37.657156 206209 out.go:179] * Done! kubectl is now configured to use "scheduled-stop-090251" cluster and "default" namespace by default
==> Docker <==
Sep 08 14:05:16 scheduled-stop-090251 dockerd[1185]: time="2025-09-08T14:05:16.124129067Z" level=info msg="Loading containers: done."
Sep 08 14:05:16 scheduled-stop-090251 dockerd[1185]: time="2025-09-08T14:05:16.134855754Z" level=info msg="Docker daemon" commit=249d679 containerd-snapshotter=false storage-driver=overlay2 version=28.4.0
Sep 08 14:05:16 scheduled-stop-090251 dockerd[1185]: time="2025-09-08T14:05:16.135062103Z" level=info msg="Initializing buildkit"
Sep 08 14:05:16 scheduled-stop-090251 dockerd[1185]: time="2025-09-08T14:05:16.150970672Z" level=info msg="Completed buildkit initialization"
Sep 08 14:05:16 scheduled-stop-090251 dockerd[1185]: time="2025-09-08T14:05:16.158544818Z" level=info msg="Daemon has completed initialization"
Sep 08 14:05:16 scheduled-stop-090251 dockerd[1185]: time="2025-09-08T14:05:16.158642138Z" level=info msg="API listen on /run/docker.sock"
Sep 08 14:05:16 scheduled-stop-090251 dockerd[1185]: time="2025-09-08T14:05:16.158816092Z" level=info msg="API listen on /var/run/docker.sock"
Sep 08 14:05:16 scheduled-stop-090251 dockerd[1185]: time="2025-09-08T14:05:16.158940521Z" level=info msg="API listen on [::]:2376"
Sep 08 14:05:16 scheduled-stop-090251 systemd[1]: Started Docker Application Container Engine.
Sep 08 14:05:16 scheduled-stop-090251 systemd[1]: Starting CRI Interface for Docker Application Container Engine...
Sep 08 14:05:16 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:16Z" level=info msg="Starting cri-dockerd dev (HEAD)"
Sep 08 14:05:16 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:16Z" level=info msg="Connecting to docker on the Endpoint unix:///var/run/docker.sock"
Sep 08 14:05:16 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:16Z" level=info msg="Start docker client with request timeout 0s"
Sep 08 14:05:16 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:16Z" level=info msg="Hairpin mode is set to hairpin-veth"
Sep 08 14:05:16 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:16Z" level=info msg="Loaded network plugin cni"
Sep 08 14:05:16 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:16Z" level=info msg="Docker cri networking managed by network plugin cni"
Sep 08 14:05:16 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:16Z" level=info msg="Setting cgroupDriver cgroupfs"
Sep 08 14:05:16 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:16Z" level=info msg="Docker cri received runtime config &RuntimeConfig{NetworkConfig:&NetworkConfig{PodCidr:,},}"
Sep 08 14:05:16 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:16Z" level=info msg="Starting the GRPC backend for the Docker CRI interface."
Sep 08 14:05:16 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:16Z" level=info msg="Start cri-dockerd grpc backend"
Sep 08 14:05:16 scheduled-stop-090251 systemd[1]: Started CRI Interface for Docker Application Container Engine.
Sep 08 14:05:28 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:28Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/c1fb1de0a74c1a2b6985b4fb9b0e567f799396703066fe817491839e3af8bde8/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options edns0 trust-ad ndots:0]"
Sep 08 14:05:28 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:28Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/e89fa36ba0c7006de8be12de69ae7efa119759812fabec5eaa6a80a7674561f6/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options edns0 trust-ad ndots:0]"
Sep 08 14:05:28 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:28Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/7ab11ddfac51692bd825918837941486179091711a250cc70314d6c6001f1490/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options ndots:0 edns0 trust-ad]"
Sep 08 14:05:28 scheduled-stop-090251 cri-dockerd[1482]: time="2025-09-08T14:05:28Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/a1fd5eff6da81b7a03da2aba3eb8ed3f0e0e446f6a06d516051ef557042acd4e/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options edns0 trust-ad ndots:0]"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
e90e2091959d1 996be7e86d9b3 11 seconds ago Running kube-controller-manager 0 a1fd5eff6da81 kube-controller-manager-scheduled-stop-090251
3631e75a66a29 a1894772a478e 11 seconds ago Running etcd 0 7ab11ddfac516 etcd-scheduled-stop-090251
588f57c847642 a25f5ef9c34c3 11 seconds ago Running kube-scheduler 0 e89fa36ba0c70 kube-scheduler-scheduled-stop-090251
eb3c15f01e993 d291939e99406 11 seconds ago Running kube-apiserver 0 c1fb1de0a74c1 kube-apiserver-scheduled-stop-090251
==> describe nodes <==
Name: scheduled-stop-090251
Roles: control-plane
Labels: beta.kubernetes.io/arch=arm64
beta.kubernetes.io/os=linux
kubernetes.io/arch=arm64
kubernetes.io/hostname=scheduled-stop-090251
kubernetes.io/os=linux
minikube.k8s.io/commit=4237956cfce90d4ab760d817400bd4c89cad50d6
minikube.k8s.io/name=scheduled-stop-090251
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2025_09_08T14_05_36_0700
minikube.k8s.io/version=v1.36.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
Annotations: volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Mon, 08 Sep 2025 14:05:32 +0000
Taints: node.kubernetes.io/not-ready:NoSchedule
Unschedulable: false
Lease:
HolderIdentity: scheduled-stop-090251
AcquireTime: <unset>
RenewTime: Mon, 08 Sep 2025 14:05:35 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Mon, 08 Sep 2025 14:05:35 +0000 Mon, 08 Sep 2025 14:05:29 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Mon, 08 Sep 2025 14:05:35 +0000 Mon, 08 Sep 2025 14:05:29 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Mon, 08 Sep 2025 14:05:35 +0000 Mon, 08 Sep 2025 14:05:29 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Mon, 08 Sep 2025 14:05:35 +0000 Mon, 08 Sep 2025 14:05:33 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.76.2
Hostname: scheduled-stop-090251
Capacity:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022296Ki
pods: 110
Allocatable:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022296Ki
pods: 110
System Info:
Machine ID: 37e32435ecd1479d84a62bfcc574b345
System UUID: f4816ee7-6992-4ba8-b5d7-bf5a41c6efba
Boot ID: bea2a2bf-dfac-4586-ae30-d33ee0d10246
Kernel Version: 5.15.0-1084-aws
OS Image: Ubuntu 22.04.5 LTS
Operating System: linux
Architecture: arm64
Container Runtime Version: docker://28.4.0
Kubelet Version: v1.34.0
Kube-Proxy Version:
Non-terminated Pods: (4 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
kube-system etcd-scheduled-stop-090251 100m (5%) 0 (0%) 100Mi (1%) 0 (0%) 6s
kube-system kube-apiserver-scheduled-stop-090251 250m (12%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-controller-manager-scheduled-stop-090251 200m (10%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-scheduler-scheduled-stop-090251 100m (5%) 0 (0%) 0 (0%) 0 (0%) 5s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 650m (32%) 0 (0%)
memory 100Mi (1%) 0 (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
hugepages-32Mi 0 (0%) 0 (0%)
hugepages-64Ki 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal NodeHasSufficientMemory 11s (x8 over 11s) kubelet Node scheduled-stop-090251 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 11s (x8 over 11s) kubelet Node scheduled-stop-090251 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 11s (x7 over 11s) kubelet Node scheduled-stop-090251 status is now: NodeHasSufficientPID
Normal NodeAllocatableEnforced 11s kubelet Updated Node Allocatable limit across pods
Normal Starting 4s kubelet Starting kubelet.
Warning CgroupV1 4s kubelet cgroup v1 support is in maintenance mode, please migrate to cgroup v2
Normal NodeAllocatableEnforced 4s kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 4s kubelet Node scheduled-stop-090251 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 4s kubelet Node scheduled-stop-090251 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 4s kubelet Node scheduled-stop-090251 status is now: NodeHasSufficientPID
==> dmesg <==
[Sep 8 13:17] ACPI: SRAT not present
[ +0.000000] ACPI: SRAT not present
[ +0.000000] SPI driver altr_a10sr has no spi_device_id for altr,a10sr
[ +0.014241] device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log.
[ +0.506765] systemd[1]: Configuration file /run/systemd/system/netplan-ovs-cleanup.service is marked world-inaccessible. This has no effect as configuration data is accessible via APIs without restrictions. Proceeding anyway.
[ +0.034639] systemd[1]: /lib/systemd/system/snapd.service:23: Unknown key name 'RestartMode' in section 'Service', ignoring.
[ +0.800626] ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy.
[ +6.784965] kauditd_printk_skb: 36 callbacks suppressed
==> etcd [3631e75a66a2] <==
{"level":"warn","ts":"2025-09-08T14:05:31.124021Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:52930","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.145387Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:52944","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.162667Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:52958","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.202671Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:52984","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.208962Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:52970","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.227504Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53000","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.240246Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53026","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.258246Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53046","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.277808Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53064","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.316198Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53086","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.340135Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53094","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.380682Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53136","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.406485Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53156","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.415915Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53106","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.421871Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53186","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.458041Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53210","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.500328Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53234","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.514861Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53250","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.528866Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53268","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.546097Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53284","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.607665Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53304","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.633106Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53310","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.656935Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53342","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.685650Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53360","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:05:31.862642Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:53376","server-name":"","error":"EOF"}
==> kernel <==
14:05:39 up 48 min, 0 users, load average: 2.61, 2.71, 2.89
Linux scheduled-stop-090251 5.15.0-1084-aws #91~20.04.1-Ubuntu SMP Fri May 2 07:00:04 UTC 2025 aarch64 aarch64 aarch64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.5 LTS"
==> kube-apiserver [eb3c15f01e99] <==
I0908 14:05:32.848918 1 shared_informer.go:356] "Caches are synced" controller="*generic.policySource[*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicy,*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicyBinding,k8s.io/apiserver/pkg/admission/plugin/policy/validating.Validator]"
I0908 14:05:32.849015 1 policy_source.go:240] refreshing policies
I0908 14:05:32.858170 1 shared_informer.go:356] "Caches are synced" controller="ipallocator-repair-controller"
I0908 14:05:32.858384 1 shared_informer.go:356] "Caches are synced" controller="kubernetes-service-cidr-controller"
I0908 14:05:32.858653 1 default_servicecidr_controller.go:166] Creating default ServiceCIDR with CIDRs: [10.96.0.0/12]
I0908 14:05:32.864084 1 default_servicecidr_controller.go:228] Setting default ServiceCIDR condition Ready to True
I0908 14:05:32.864085 1 cidrallocator.go:301] created ClusterIP allocator for Service CIDR 10.96.0.0/12
I0908 14:05:32.872206 1 default_servicecidr_controller.go:137] Shutting down kubernetes-service-cidr-controller
I0908 14:05:32.874689 1 cidrallocator.go:277] updated ClusterIP allocator for Service CIDR 10.96.0.0/12
E0908 14:05:32.881278 1 controller.go:145] "Failed to ensure lease exists, will retry" err="namespaces \"kube-system\" not found" interval="200ms"
I0908 14:05:32.905084 1 controller.go:667] quota admission added evaluator for: namespaces
I0908 14:05:33.101084 1 controller.go:667] quota admission added evaluator for: leases.coordination.k8s.io
I0908 14:05:33.637373 1 storage_scheduling.go:95] created PriorityClass system-node-critical with value 2000001000
I0908 14:05:33.642850 1 storage_scheduling.go:95] created PriorityClass system-cluster-critical with value 2000000000
I0908 14:05:33.642948 1 storage_scheduling.go:111] all system priority classes are created successfully or already exist.
I0908 14:05:34.412247 1 controller.go:667] quota admission added evaluator for: roles.rbac.authorization.k8s.io
I0908 14:05:34.467601 1 controller.go:667] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
I0908 14:05:34.548981 1 alloc.go:328] "allocated clusterIPs" service="default/kubernetes" clusterIPs={"IPv4":"10.96.0.1"}
W0908 14:05:34.557263 1 lease.go:265] Resetting endpoints for master service "kubernetes" to [192.168.76.2]
I0908 14:05:34.558548 1 controller.go:667] quota admission added evaluator for: endpoints
I0908 14:05:34.564153 1 controller.go:667] quota admission added evaluator for: endpointslices.discovery.k8s.io
I0908 14:05:34.722312 1 controller.go:667] quota admission added evaluator for: serviceaccounts
I0908 14:05:35.573577 1 controller.go:667] quota admission added evaluator for: deployments.apps
I0908 14:05:35.598380 1 alloc.go:328] "allocated clusterIPs" service="kube-system/kube-dns" clusterIPs={"IPv4":"10.96.0.10"}
I0908 14:05:35.610153 1 controller.go:667] quota admission added evaluator for: daemonsets.apps
==> kube-controller-manager [e90e2091959d] <==
I0908 14:05:38.268548 1 node_lifecycle_controller.go:453] "Sending events to api server" logger="node-lifecycle-controller"
I0908 14:05:38.268572 1 node_lifecycle_controller.go:464] "Starting node controller" logger="node-lifecycle-controller"
I0908 14:05:38.268578 1 shared_informer.go:349] "Waiting for caches to sync" controller="taint"
I0908 14:05:38.420089 1 controllermanager.go:781] "Started controller" controller="persistentvolume-binder-controller"
I0908 14:05:38.420262 1 pv_controller_base.go:308] "Starting persistent volume controller" logger="persistentvolume-binder-controller"
I0908 14:05:38.420277 1 shared_informer.go:349] "Waiting for caches to sync" controller="persistent volume"
I0908 14:05:38.571072 1 controllermanager.go:781] "Started controller" controller="resourceclaim-controller"
I0908 14:05:38.571101 1 controllermanager.go:733] "Controller is disabled by a feature gate" controller="device-taint-eviction-controller" requiredFeatureGates=["DynamicResourceAllocation","DRADeviceTaints"]
I0908 14:05:38.571150 1 controller.go:397] "Starting resource claim controller" logger="resourceclaim-controller"
I0908 14:05:38.571190 1 shared_informer.go:349] "Waiting for caches to sync" controller="resource_claim"
I0908 14:05:38.725654 1 controllermanager.go:781] "Started controller" controller="service-cidr-controller"
I0908 14:05:38.725826 1 servicecidrs_controller.go:137] "Starting" logger="service-cidr-controller" controller="service-cidr-controller"
I0908 14:05:38.725837 1 shared_informer.go:349] "Waiting for caches to sync" controller="service-cidr-controller"
I0908 14:05:38.870814 1 controllermanager.go:781] "Started controller" controller="replicationcontroller-controller"
I0908 14:05:38.870924 1 replica_set.go:243] "Starting controller" logger="replicationcontroller-controller" name="replicationcontroller"
I0908 14:05:38.870934 1 shared_informer.go:349] "Waiting for caches to sync" controller="ReplicationController"
I0908 14:05:39.020009 1 controllermanager.go:781] "Started controller" controller="daemonset-controller"
I0908 14:05:39.020173 1 daemon_controller.go:310] "Starting daemon sets controller" logger="daemonset-controller"
I0908 14:05:39.020186 1 shared_informer.go:349] "Waiting for caches to sync" controller="daemon sets"
I0908 14:05:39.318641 1 controllermanager.go:781] "Started controller" controller="horizontal-pod-autoscaler-controller"
I0908 14:05:39.318697 1 horizontal.go:205] "Starting HPA controller" logger="horizontal-pod-autoscaler-controller"
I0908 14:05:39.318706 1 shared_informer.go:349] "Waiting for caches to sync" controller="HPA"
I0908 14:05:39.469914 1 controllermanager.go:781] "Started controller" controller="cronjob-controller"
I0908 14:05:39.470011 1 cronjob_controllerv2.go:145] "Starting cronjob controller v2" logger="cronjob-controller"
I0908 14:05:39.470024 1 shared_informer.go:349] "Waiting for caches to sync" controller="cronjob"
==> kube-scheduler [588f57c84764] <==
I0908 14:05:33.205660 1 shared_informer.go:349] "Waiting for caches to sync" controller="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
I0908 14:05:33.205479 1 secure_serving.go:211] Serving securely on 127.0.0.1:10259
I0908 14:05:33.205509 1 tlsconfig.go:243] "Starting DynamicServingCertificateController"
E0908 14:05:33.216266 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSINode"
E0908 14:05:33.216594 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIStorageCapacity"
E0908 14:05:33.222402 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace"
E0908 14:05:33.222872 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolume"
E0908 14:05:33.223143 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod"
E0908 14:05:33.223374 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node"
E0908 14:05:33.223573 1 reflector.go:205] "Failed to watch" err="failed to list *v1.DeviceClass: deviceclasses.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"deviceclasses\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.DeviceClass"
E0908 14:05:33.223771 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service"
E0908 14:05:33.223949 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicationController"
E0908 14:05:33.224121 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PodDisruptionBudget"
E0908 14:05:33.224298 1 reflector.go:205] "Failed to watch" err="failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.StatefulSet"
E0908 14:05:33.224575 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError" reflector="runtime/asm_arm64.s:1223" type="*v1.ConfigMap"
E0908 14:05:33.224772 1 reflector.go:205] "Failed to watch" err="failed to list *v1.VolumeAttachment: volumeattachments.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"volumeattachments\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.VolumeAttachment"
E0908 14:05:33.228591 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver"
E0908 14:05:33.228892 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ResourceClaim: resourceclaims.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"resourceclaims\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ResourceClaim"
E0908 14:05:33.229146 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ResourceSlice: resourceslices.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"resourceslices\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ResourceSlice"
E0908 14:05:33.229471 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicaSet"
E0908 14:05:33.229759 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolumeClaim"
E0908 14:05:33.229981 1 reflector.go:205] "Failed to watch" err="failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.StorageClass"
E0908 14:05:34.043181 1 reflector.go:205] "Failed to watch" err="failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.StorageClass"
E0908 14:05:34.099315 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node"
I0908 14:05:34.706839 1 shared_informer.go:356] "Caches are synced" controller="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
==> kubelet <==
Sep 08 14:05:35 scheduled-stop-090251 kubelet[2349]: I0908 14:05:35.906588 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-data\" (UniqueName: \"kubernetes.io/host-path/44abaa4290b67c228c69c43d79fbe80b-etcd-data\") pod \"etcd-scheduled-stop-090251\" (UID: \"44abaa4290b67c228c69c43d79fbe80b\") " pod="kube-system/etcd-scheduled-stop-090251"
Sep 08 14:05:35 scheduled-stop-090251 kubelet[2349]: I0908 14:05:35.906606 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a75ace79cf9a5bb5217509871fcd1cb3-k8s-certs\") pod \"kube-controller-manager-scheduled-stop-090251\" (UID: \"a75ace79cf9a5bb5217509871fcd1cb3\") " pod="kube-system/kube-controller-manager-scheduled-stop-090251"
Sep 08 14:05:35 scheduled-stop-090251 kubelet[2349]: I0908 14:05:35.906622 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a75ace79cf9a5bb5217509871fcd1cb3-usr-share-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-090251\" (UID: \"a75ace79cf9a5bb5217509871fcd1cb3\") " pod="kube-system/kube-controller-manager-scheduled-stop-090251"
Sep 08 14:05:35 scheduled-stop-090251 kubelet[2349]: I0908 14:05:35.906638 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/9c7d48a7794008853495fc4736306752-k8s-certs\") pod \"kube-apiserver-scheduled-stop-090251\" (UID: \"9c7d48a7794008853495fc4736306752\") " pod="kube-system/kube-apiserver-scheduled-stop-090251"
Sep 08 14:05:35 scheduled-stop-090251 kubelet[2349]: I0908 14:05:35.906655 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/9c7d48a7794008853495fc4736306752-usr-share-ca-certificates\") pod \"kube-apiserver-scheduled-stop-090251\" (UID: \"9c7d48a7794008853495fc4736306752\") " pod="kube-system/kube-apiserver-scheduled-stop-090251"
Sep 08 14:05:35 scheduled-stop-090251 kubelet[2349]: I0908 14:05:35.906684 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a75ace79cf9a5bb5217509871fcd1cb3-ca-certs\") pod \"kube-controller-manager-scheduled-stop-090251\" (UID: \"a75ace79cf9a5bb5217509871fcd1cb3\") " pod="kube-system/kube-controller-manager-scheduled-stop-090251"
Sep 08 14:05:35 scheduled-stop-090251 kubelet[2349]: I0908 14:05:35.906706 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a75ace79cf9a5bb5217509871fcd1cb3-etc-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-090251\" (UID: \"a75ace79cf9a5bb5217509871fcd1cb3\") " pod="kube-system/kube-controller-manager-scheduled-stop-090251"
Sep 08 14:05:35 scheduled-stop-090251 kubelet[2349]: I0908 14:05:35.906728 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/3ca8911316dfc033eef34347a8dba1d3-kubeconfig\") pod \"kube-scheduler-scheduled-stop-090251\" (UID: \"3ca8911316dfc033eef34347a8dba1d3\") " pod="kube-system/kube-scheduler-scheduled-stop-090251"
Sep 08 14:05:35 scheduled-stop-090251 kubelet[2349]: I0908 14:05:35.906744 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-certs\" (UniqueName: \"kubernetes.io/host-path/44abaa4290b67c228c69c43d79fbe80b-etcd-certs\") pod \"etcd-scheduled-stop-090251\" (UID: \"44abaa4290b67c228c69c43d79fbe80b\") " pod="kube-system/etcd-scheduled-stop-090251"
Sep 08 14:05:35 scheduled-stop-090251 kubelet[2349]: I0908 14:05:35.906841 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/9c7d48a7794008853495fc4736306752-ca-certs\") pod \"kube-apiserver-scheduled-stop-090251\" (UID: \"9c7d48a7794008853495fc4736306752\") " pod="kube-system/kube-apiserver-scheduled-stop-090251"
Sep 08 14:05:35 scheduled-stop-090251 kubelet[2349]: I0908 14:05:35.906872 2349 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a75ace79cf9a5bb5217509871fcd1cb3-usr-local-share-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-090251\" (UID: \"a75ace79cf9a5bb5217509871fcd1cb3\") " pod="kube-system/kube-controller-manager-scheduled-stop-090251"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: I0908 14:05:36.491332 2349 apiserver.go:52] "Watching apiserver"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: I0908 14:05:36.600680 2349 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: I0908 14:05:36.655746 2349 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/etcd-scheduled-stop-090251" podStartSLOduration=3.6557246340000003 podStartE2EDuration="3.655724634s" podCreationTimestamp="2025-09-08 14:05:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 14:05:36.641801904 +0000 UTC m=+1.247308390" watchObservedRunningTime="2025-09-08 14:05:36.655724634 +0000 UTC m=+1.261231120"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: I0908 14:05:36.661379 2349 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/etcd-scheduled-stop-090251"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: I0908 14:05:36.661882 2349 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-scheduled-stop-090251"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: I0908 14:05:36.662418 2349 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-scheduled-stop-090251"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: I0908 14:05:36.662981 2349 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-scheduled-stop-090251"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: E0908 14:05:36.686920 2349 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-scheduler-scheduled-stop-090251\" already exists" pod="kube-system/kube-scheduler-scheduled-stop-090251"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: I0908 14:05:36.687039 2349 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-scheduled-stop-090251" podStartSLOduration=1.687023883 podStartE2EDuration="1.687023883s" podCreationTimestamp="2025-09-08 14:05:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 14:05:36.656233575 +0000 UTC m=+1.261740053" watchObservedRunningTime="2025-09-08 14:05:36.687023883 +0000 UTC m=+1.292530361"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: E0908 14:05:36.687183 2349 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-controller-manager-scheduled-stop-090251\" already exists" pod="kube-system/kube-controller-manager-scheduled-stop-090251"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: E0908 14:05:36.687333 2349 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-apiserver-scheduled-stop-090251\" already exists" pod="kube-system/kube-apiserver-scheduled-stop-090251"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: E0908 14:05:36.687421 2349 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"etcd-scheduled-stop-090251\" already exists" pod="kube-system/etcd-scheduled-stop-090251"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: I0908 14:05:36.701328 2349 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-scheduled-stop-090251" podStartSLOduration=2.701307666 podStartE2EDuration="2.701307666s" podCreationTimestamp="2025-09-08 14:05:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 14:05:36.699568521 +0000 UTC m=+1.305075007" watchObservedRunningTime="2025-09-08 14:05:36.701307666 +0000 UTC m=+1.306814136"
Sep 08 14:05:36 scheduled-stop-090251 kubelet[2349]: I0908 14:05:36.701596 2349 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-scheduled-stop-090251" podStartSLOduration=1.701587253 podStartE2EDuration="1.701587253s" podCreationTimestamp="2025-09-08 14:05:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 14:05:36.688120698 +0000 UTC m=+1.293627184" watchObservedRunningTime="2025-09-08 14:05:36.701587253 +0000 UTC m=+1.307093739"
-- /stdout --
helpers_test.go:262: (dbg) Run: out/minikube-linux-arm64 status --format={{.APIServer}} -p scheduled-stop-090251 -n scheduled-stop-090251
helpers_test.go:269: (dbg) Run: kubectl --context scheduled-stop-090251 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:280: non-running pods: storage-provisioner
helpers_test.go:282: ======> post-mortem[TestScheduledStopUnix]: describe non-running pods <======
helpers_test.go:285: (dbg) Run: kubectl --context scheduled-stop-090251 describe pod storage-provisioner
helpers_test.go:285: (dbg) Non-zero exit: kubectl --context scheduled-stop-090251 describe pod storage-provisioner: exit status 1 (118.644161ms)
** stderr **
Error from server (NotFound): pods "storage-provisioner" not found
** /stderr **
helpers_test.go:287: kubectl --context scheduled-stop-090251 describe pod storage-provisioner: exit status 1
helpers_test.go:175: Cleaning up "scheduled-stop-090251" profile ...
helpers_test.go:178: (dbg) Run: out/minikube-linux-arm64 delete -p scheduled-stop-090251
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p scheduled-stop-090251: (2.194673736s)
--- FAIL: TestScheduledStopUnix (37.07s)