=== RUN TestScheduledStopUnix
scheduled_stop_test.go:128: (dbg) Run: out/minikube-linux-arm64 start -p scheduled-stop-273808 --memory=3072 --driver=docker --container-runtime=docker
scheduled_stop_test.go:128: (dbg) Done: out/minikube-linux-arm64 start -p scheduled-stop-273808 --memory=3072 --driver=docker --container-runtime=docker: (37.273196733s)
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-273808 --schedule 5m
scheduled_stop_test.go:191: (dbg) Run: out/minikube-linux-arm64 status --format={{.TimeToStop}} -p scheduled-stop-273808 -n scheduled-stop-273808
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-273808 --schedule 15s
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:98: process 1489330 running but should have been killed on reschedule of stop
panic.go:636: *** TestScheduledStopUnix FAILED at 2025-10-02 07:01:06.923295614 +0000 UTC m=+2428.094611752
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:223: ======> post-mortem[TestScheduledStopUnix]: network settings <======
helpers_test.go:230: HOST ENV snapshots: PROXY env: HTTP_PROXY="<empty>" HTTPS_PROXY="<empty>" NO_PROXY="<empty>"
helpers_test.go:238: ======> post-mortem[TestScheduledStopUnix]: docker inspect <======
helpers_test.go:239: (dbg) Run: docker inspect scheduled-stop-273808
helpers_test.go:243: (dbg) docker inspect scheduled-stop-273808:
-- stdout --
[
{
"Id": "385edf8f269ac45cb2c2712ce9c6ea5c659c2bc387459b09df2369c48cc3968d",
"Created": "2025-10-02T07:00:34.162399099Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 1486522,
"ExitCode": 0,
"Error": "",
"StartedAt": "2025-10-02T07:00:34.229066637Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:5f534d1f6dbdc6822bb3d07eb55e2a83d08e94cbdcc855a877b4f3dd1ac1278e",
"ResolvConfPath": "/var/lib/docker/containers/385edf8f269ac45cb2c2712ce9c6ea5c659c2bc387459b09df2369c48cc3968d/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/385edf8f269ac45cb2c2712ce9c6ea5c659c2bc387459b09df2369c48cc3968d/hostname",
"HostsPath": "/var/lib/docker/containers/385edf8f269ac45cb2c2712ce9c6ea5c659c2bc387459b09df2369c48cc3968d/hosts",
"LogPath": "/var/lib/docker/containers/385edf8f269ac45cb2c2712ce9c6ea5c659c2bc387459b09df2369c48cc3968d/385edf8f269ac45cb2c2712ce9c6ea5c659c2bc387459b09df2369c48cc3968d-json.log",
"Name": "/scheduled-stop-273808",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"scheduled-stop-273808:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {}
},
"NetworkMode": "scheduled-stop-273808",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 3221225472,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 6442450944,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"ID": "385edf8f269ac45cb2c2712ce9c6ea5c659c2bc387459b09df2369c48cc3968d",
"LowerDir": "/var/lib/docker/overlay2/70c00a18e9d10e89e06a785f0f9373c5b0cc82248bb423a3be7479c1f1b6e631-init/diff:/var/lib/docker/overlay2/e75aeb731217e4929bbe543c44bed11f3df1ccbcd034bec040802dc1e2cd58a4/diff",
"MergedDir": "/var/lib/docker/overlay2/70c00a18e9d10e89e06a785f0f9373c5b0cc82248bb423a3be7479c1f1b6e631/merged",
"UpperDir": "/var/lib/docker/overlay2/70c00a18e9d10e89e06a785f0f9373c5b0cc82248bb423a3be7479c1f1b6e631/diff",
"WorkDir": "/var/lib/docker/overlay2/70c00a18e9d10e89e06a785f0f9373c5b0cc82248bb423a3be7479c1f1b6e631/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "scheduled-stop-273808",
"Source": "/var/lib/docker/volumes/scheduled-stop-273808/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "scheduled-stop-273808",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759382731-21643@sha256:ca1b4db171879edd6bbb9546a4b1afac2eb5be94a0f5528496e62d2ff99de37d",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "scheduled-stop-273808",
"name.minikube.sigs.k8s.io": "scheduled-stop-273808",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "3e3ec3cfbf696815c5fb01758e4e4cc0b22a19e82f035cd941dc099850290912",
"SandboxKey": "/var/run/docker/netns/3e3ec3cfbf69",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "34154"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "34155"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "34158"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "34156"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "34157"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"scheduled-stop-273808": {
"IPAMConfig": {
"IPv4Address": "192.168.76.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "06:8b:b0:35:16:b4",
"DriverOpts": null,
"GwPriority": 0,
"NetworkID": "64f3d9b6c1db52ad9baa5e5becbb7016464c592c46a36ca76947e3bc955715f4",
"EndpointID": "8d47ae71d2849adfb76b01b92ff6188d83c96c08ec69edf629f8d5d3697f50f7",
"Gateway": "192.168.76.1",
"IPAddress": "192.168.76.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"scheduled-stop-273808",
"385edf8f269a"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:247: (dbg) Run: out/minikube-linux-arm64 status --format={{.Host}} -p scheduled-stop-273808 -n scheduled-stop-273808
helpers_test.go:252: <<< TestScheduledStopUnix FAILED: start of post-mortem logs <<<
helpers_test.go:253: ======> post-mortem[TestScheduledStopUnix]: minikube logs <======
helpers_test.go:255: (dbg) Run: out/minikube-linux-arm64 -p scheduled-stop-273808 logs -n 25
helpers_test.go:260: TestScheduledStopUnix logs:
-- stdout --
==> Audit <==
┌─────────┬─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┬───────────────────────┬─────────┬─────────┬─────────────────────┬─────────────────────┐
│ COMMAND │ ARGS │ PROFILE │ USER │ VERSION │ START TIME │ END TIME │
├─────────┼─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┼───────────────────────┼─────────┼─────────┼─────────────────────┼─────────────────────┤
│ stop │ -p multinode-355238 │ multinode-355238 │ jenkins │ v1.37.0 │ 02 Oct 25 06:55 UTC │ 02 Oct 25 06:55 UTC │
│ start │ -p multinode-355238 --wait=true -v=5 --alsologtostderr │ multinode-355238 │ jenkins │ v1.37.0 │ 02 Oct 25 06:55 UTC │ 02 Oct 25 06:56 UTC │
│ node │ list -p multinode-355238 │ multinode-355238 │ jenkins │ v1.37.0 │ 02 Oct 25 06:56 UTC │ │
│ node │ multinode-355238 node delete m03 │ multinode-355238 │ jenkins │ v1.37.0 │ 02 Oct 25 06:56 UTC │ 02 Oct 25 06:56 UTC │
│ stop │ multinode-355238 stop │ multinode-355238 │ jenkins │ v1.37.0 │ 02 Oct 25 06:56 UTC │ 02 Oct 25 06:56 UTC │
│ start │ -p multinode-355238 --wait=true -v=5 --alsologtostderr --driver=docker --container-runtime=docker │ multinode-355238 │ jenkins │ v1.37.0 │ 02 Oct 25 06:56 UTC │ 02 Oct 25 06:57 UTC │
│ node │ list -p multinode-355238 │ multinode-355238 │ jenkins │ v1.37.0 │ 02 Oct 25 06:57 UTC │ │
│ start │ -p multinode-355238-m02 --driver=docker --container-runtime=docker │ multinode-355238-m02 │ jenkins │ v1.37.0 │ 02 Oct 25 06:57 UTC │ │
│ start │ -p multinode-355238-m03 --driver=docker --container-runtime=docker │ multinode-355238-m03 │ jenkins │ v1.37.0 │ 02 Oct 25 06:57 UTC │ 02 Oct 25 06:58 UTC │
│ node │ add -p multinode-355238 │ multinode-355238 │ jenkins │ v1.37.0 │ 02 Oct 25 06:58 UTC │ │
│ delete │ -p multinode-355238-m03 │ multinode-355238-m03 │ jenkins │ v1.37.0 │ 02 Oct 25 06:58 UTC │ 02 Oct 25 06:58 UTC │
│ delete │ -p multinode-355238 │ multinode-355238 │ jenkins │ v1.37.0 │ 02 Oct 25 06:58 UTC │ 02 Oct 25 06:58 UTC │
│ start │ -p test-preload-876119 --memory=3072 --alsologtostderr --wait=true --preload=false --driver=docker --container-runtime=docker --kubernetes-version=v1.32.0 │ test-preload-876119 │ jenkins │ v1.37.0 │ 02 Oct 25 06:58 UTC │ 02 Oct 25 06:59 UTC │
│ image │ test-preload-876119 image pull gcr.io/k8s-minikube/busybox │ test-preload-876119 │ jenkins │ v1.37.0 │ 02 Oct 25 06:59 UTC │ 02 Oct 25 06:59 UTC │
│ stop │ -p test-preload-876119 │ test-preload-876119 │ jenkins │ v1.37.0 │ 02 Oct 25 06:59 UTC │ 02 Oct 25 06:59 UTC │
│ start │ -p test-preload-876119 --memory=3072 --alsologtostderr -v=1 --wait=true --driver=docker --container-runtime=docker │ test-preload-876119 │ jenkins │ v1.37.0 │ 02 Oct 25 06:59 UTC │ 02 Oct 25 07:00 UTC │
│ image │ test-preload-876119 image list │ test-preload-876119 │ jenkins │ v1.37.0 │ 02 Oct 25 07:00 UTC │ 02 Oct 25 07:00 UTC │
│ delete │ -p test-preload-876119 │ test-preload-876119 │ jenkins │ v1.37.0 │ 02 Oct 25 07:00 UTC │ 02 Oct 25 07:00 UTC │
│ start │ -p scheduled-stop-273808 --memory=3072 --driver=docker --container-runtime=docker │ scheduled-stop-273808 │ jenkins │ v1.37.0 │ 02 Oct 25 07:00 UTC │ 02 Oct 25 07:01 UTC │
│ stop │ -p scheduled-stop-273808 --schedule 5m │ scheduled-stop-273808 │ jenkins │ v1.37.0 │ 02 Oct 25 07:01 UTC │ │
│ stop │ -p scheduled-stop-273808 --schedule 5m │ scheduled-stop-273808 │ jenkins │ v1.37.0 │ 02 Oct 25 07:01 UTC │ │
│ stop │ -p scheduled-stop-273808 --schedule 5m │ scheduled-stop-273808 │ jenkins │ v1.37.0 │ 02 Oct 25 07:01 UTC │ │
│ stop │ -p scheduled-stop-273808 --schedule 15s │ scheduled-stop-273808 │ jenkins │ v1.37.0 │ 02 Oct 25 07:01 UTC │ │
│ stop │ -p scheduled-stop-273808 --schedule 15s │ scheduled-stop-273808 │ jenkins │ v1.37.0 │ 02 Oct 25 07:01 UTC │ │
│ stop │ -p scheduled-stop-273808 --schedule 15s │ scheduled-stop-273808 │ jenkins │ v1.37.0 │ 02 Oct 25 07:01 UTC │ │
└─────────┴─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┴───────────────────────┴─────────┴─────────┴─────────────────────┴─────────────────────┘
==> Last Start <==
Log file created at: 2025/10/02 07:00:29
Running on machine: ip-172-31-29-130
Binary: Built with gc go1.24.6 for linux/arm64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I1002 07:00:29.178951 1486138 out.go:360] Setting OutFile to fd 1 ...
I1002 07:00:29.179069 1486138 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1002 07:00:29.179072 1486138 out.go:374] Setting ErrFile to fd 2...
I1002 07:00:29.179076 1486138 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1002 07:00:29.179316 1486138 root.go:338] Updating PATH: /home/jenkins/minikube-integration/21643-1281649/.minikube/bin
I1002 07:00:29.179709 1486138 out.go:368] Setting JSON to false
I1002 07:00:29.180575 1486138 start.go:130] hostinfo: {"hostname":"ip-172-31-29-130","uptime":24164,"bootTime":1759364266,"procs":154,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1084-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"36adf542-ef4f-4e2d-a0c8-6868d1383ff9"}
I1002 07:00:29.180631 1486138 start.go:140] virtualization:
I1002 07:00:29.184686 1486138 out.go:179] * [scheduled-stop-273808] minikube v1.37.0 on Ubuntu 20.04 (arm64)
I1002 07:00:29.189271 1486138 out.go:179] - MINIKUBE_LOCATION=21643
I1002 07:00:29.189383 1486138 notify.go:220] Checking for updates...
I1002 07:00:29.196136 1486138 out.go:179] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I1002 07:00:29.199445 1486138 out.go:179] - KUBECONFIG=/home/jenkins/minikube-integration/21643-1281649/kubeconfig
I1002 07:00:29.202520 1486138 out.go:179] - MINIKUBE_HOME=/home/jenkins/minikube-integration/21643-1281649/.minikube
I1002 07:00:29.205703 1486138 out.go:179] - MINIKUBE_BIN=out/minikube-linux-arm64
I1002 07:00:29.208749 1486138 out.go:179] - MINIKUBE_FORCE_SYSTEMD=
I1002 07:00:29.212129 1486138 driver.go:421] Setting default libvirt URI to qemu:///system
I1002 07:00:29.238829 1486138 docker.go:123] docker version: linux-28.1.1:Docker Engine - Community
I1002 07:00:29.238949 1486138 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1002 07:00:29.298181 1486138 info.go:266] docker info: {ID:U5VK:ZNT5:35M3:FHLW:Q7TL:ELFX:BNAG:AV4T:UD2H:SK5L:SEJV:SJJL Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:23 OomKillDisable:true NGoroutines:42 SystemTime:2025-10-02 07:00:29.288918219 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1084-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214835200 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-29-130 Labels:[] ExperimentalBuild:false ServerVersion:28.1.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:05044ec0a9a75232cad458027ca83437aae3f4da Expected:} RuncCommit:{ID:v1.2.5-0-g59923ef Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx P
ath:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.23.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.35.1]] Warnings:<nil>}}
I1002 07:00:29.298277 1486138 docker.go:318] overlay module found
I1002 07:00:29.301693 1486138 out.go:179] * Using the docker driver based on user configuration
I1002 07:00:29.304649 1486138 start.go:304] selected driver: docker
I1002 07:00:29.304659 1486138 start.go:924] validating driver "docker" against <nil>
I1002 07:00:29.304671 1486138 start.go:935] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I1002 07:00:29.305408 1486138 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1002 07:00:29.362568 1486138 info.go:266] docker info: {ID:U5VK:ZNT5:35M3:FHLW:Q7TL:ELFX:BNAG:AV4T:UD2H:SK5L:SEJV:SJJL Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:23 OomKillDisable:true NGoroutines:42 SystemTime:2025-10-02 07:00:29.35351188 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1084-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:aa
rch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214835200 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-29-130 Labels:[] ExperimentalBuild:false ServerVersion:28.1.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:05044ec0a9a75232cad458027ca83437aae3f4da Expected:} RuncCommit:{ID:v1.2.5-0-g59923ef Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Pa
th:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.23.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.35.1]] Warnings:<nil>}}
I1002 07:00:29.362711 1486138 start_flags.go:327] no existing cluster config was found, will generate one from the flags
I1002 07:00:29.362926 1486138 start_flags.go:974] Wait components to verify : map[apiserver:true system_pods:true]
I1002 07:00:29.365922 1486138 out.go:179] * Using Docker driver with root privileges
I1002 07:00:29.368884 1486138 cni.go:84] Creating CNI manager for ""
I1002 07:00:29.368956 1486138 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I1002 07:00:29.368964 1486138 start_flags.go:336] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I1002 07:00:29.369043 1486138 start.go:348] cluster config:
{Name:scheduled-stop-273808 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759382731-21643@sha256:ca1b4db171879edd6bbb9546a4b1afac2eb5be94a0f5528496e62d2ff99de37d Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-273808 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local
ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I1002 07:00:29.372253 1486138 out.go:179] * Starting "scheduled-stop-273808" primary control-plane node in "scheduled-stop-273808" cluster
I1002 07:00:29.375133 1486138 cache.go:123] Beginning downloading kic base image for docker with docker
I1002 07:00:29.378156 1486138 out.go:179] * Pulling base image v0.0.48-1759382731-21643 ...
I1002 07:00:29.380965 1486138 preload.go:183] Checking if preload exists for k8s version v1.34.1 and runtime docker
I1002 07:00:29.381007 1486138 preload.go:198] Found local preload: /home/jenkins/minikube-integration/21643-1281649/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-docker-overlay2-arm64.tar.lz4
I1002 07:00:29.381015 1486138 cache.go:58] Caching tarball of preloaded images
I1002 07:00:29.381112 1486138 preload.go:233] Found /home/jenkins/minikube-integration/21643-1281649/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-docker-overlay2-arm64.tar.lz4 in cache, skipping download
I1002 07:00:29.381120 1486138 cache.go:61] Finished verifying existence of preloaded tar for v1.34.1 on docker
I1002 07:00:29.381435 1486138 profile.go:143] Saving config to /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/config.json ...
I1002 07:00:29.381452 1486138 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/config.json: {Name:mk5bbe684956a2b7bb552b021092cdae80a8efc7 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1002 07:00:29.381625 1486138 image.go:81] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759382731-21643@sha256:ca1b4db171879edd6bbb9546a4b1afac2eb5be94a0f5528496e62d2ff99de37d in local docker daemon
I1002 07:00:29.400406 1486138 image.go:100] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759382731-21643@sha256:ca1b4db171879edd6bbb9546a4b1afac2eb5be94a0f5528496e62d2ff99de37d in local docker daemon, skipping pull
I1002 07:00:29.400417 1486138 cache.go:147] gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759382731-21643@sha256:ca1b4db171879edd6bbb9546a4b1afac2eb5be94a0f5528496e62d2ff99de37d exists in daemon, skipping load
I1002 07:00:29.400444 1486138 cache.go:232] Successfully downloaded all kic artifacts
I1002 07:00:29.400469 1486138 start.go:360] acquireMachinesLock for scheduled-stop-273808: {Name:mkda6f7c4d465507031dbfae99ef6247f3fd0af5 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I1002 07:00:29.400582 1486138 start.go:364] duration metric: took 99.362µs to acquireMachinesLock for "scheduled-stop-273808"
I1002 07:00:29.400606 1486138 start.go:93] Provisioning new machine with config: &{Name:scheduled-stop-273808 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759382731-21643@sha256:ca1b4db171879edd6bbb9546a4b1afac2eb5be94a0f5528496e62d2ff99de37d Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-273808 Namespace:default APIServerHAVIP: A
PIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHA
uthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I1002 07:00:29.400668 1486138 start.go:125] createHost starting for "" (driver="docker")
I1002 07:00:29.404119 1486138 out.go:252] * Creating docker container (CPUs=2, Memory=3072MB) ...
I1002 07:00:29.404345 1486138 start.go:159] libmachine.API.Create for "scheduled-stop-273808" (driver="docker")
I1002 07:00:29.404388 1486138 client.go:168] LocalClient.Create starting
I1002 07:00:29.404483 1486138 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21643-1281649/.minikube/certs/ca.pem
I1002 07:00:29.404516 1486138 main.go:141] libmachine: Decoding PEM data...
I1002 07:00:29.404528 1486138 main.go:141] libmachine: Parsing certificate...
I1002 07:00:29.404575 1486138 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21643-1281649/.minikube/certs/cert.pem
I1002 07:00:29.404594 1486138 main.go:141] libmachine: Decoding PEM data...
I1002 07:00:29.404602 1486138 main.go:141] libmachine: Parsing certificate...
I1002 07:00:29.404944 1486138 cli_runner.go:164] Run: docker network inspect scheduled-stop-273808 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W1002 07:00:29.420944 1486138 cli_runner.go:211] docker network inspect scheduled-stop-273808 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I1002 07:00:29.421015 1486138 network_create.go:284] running [docker network inspect scheduled-stop-273808] to gather additional debugging logs...
I1002 07:00:29.421030 1486138 cli_runner.go:164] Run: docker network inspect scheduled-stop-273808
W1002 07:00:29.436771 1486138 cli_runner.go:211] docker network inspect scheduled-stop-273808 returned with exit code 1
I1002 07:00:29.436791 1486138 network_create.go:287] error running [docker network inspect scheduled-stop-273808]: docker network inspect scheduled-stop-273808: exit status 1
stdout:
[]
stderr:
Error response from daemon: network scheduled-stop-273808 not found
I1002 07:00:29.436818 1486138 network_create.go:289] output of [docker network inspect scheduled-stop-273808]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network scheduled-stop-273808 not found
** /stderr **
I1002 07:00:29.436914 1486138 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I1002 07:00:29.453399 1486138 network.go:211] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-200b88fe63d3 IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:3a:44:19:92:75:0d} reservation:<nil>}
I1002 07:00:29.453686 1486138 network.go:211] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-e62f03852c34 IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:aa:41:83:14:3f:fc} reservation:<nil>}
I1002 07:00:29.453942 1486138 network.go:211] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-52024f7d4aea IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:8a:75:f8:11:6c:e1} reservation:<nil>}
I1002 07:00:29.454265 1486138 network.go:206] using free private subnet 192.168.76.0/24: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x4001951d50}
I1002 07:00:29.454279 1486138 network_create.go:124] attempt to create docker network scheduled-stop-273808 192.168.76.0/24 with gateway 192.168.76.1 and MTU of 1500 ...
I1002 07:00:29.454334 1486138 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.76.0/24 --gateway=192.168.76.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=scheduled-stop-273808 scheduled-stop-273808
I1002 07:00:29.516274 1486138 network_create.go:108] docker network scheduled-stop-273808 192.168.76.0/24 created
I1002 07:00:29.516293 1486138 kic.go:121] calculated static IP "192.168.76.2" for the "scheduled-stop-273808" container
I1002 07:00:29.516381 1486138 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I1002 07:00:29.530734 1486138 cli_runner.go:164] Run: docker volume create scheduled-stop-273808 --label name.minikube.sigs.k8s.io=scheduled-stop-273808 --label created_by.minikube.sigs.k8s.io=true
I1002 07:00:29.548478 1486138 oci.go:103] Successfully created a docker volume scheduled-stop-273808
I1002 07:00:29.548578 1486138 cli_runner.go:164] Run: docker run --rm --name scheduled-stop-273808-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-273808 --entrypoint /usr/bin/test -v scheduled-stop-273808:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759382731-21643@sha256:ca1b4db171879edd6bbb9546a4b1afac2eb5be94a0f5528496e62d2ff99de37d -d /var/lib
I1002 07:00:30.066861 1486138 oci.go:107] Successfully prepared a docker volume scheduled-stop-273808
I1002 07:00:30.067130 1486138 preload.go:183] Checking if preload exists for k8s version v1.34.1 and runtime docker
I1002 07:00:30.067154 1486138 kic.go:194] Starting extracting preloaded images to volume ...
I1002 07:00:30.067237 1486138 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/21643-1281649/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-docker-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-273808:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759382731-21643@sha256:ca1b4db171879edd6bbb9546a4b1afac2eb5be94a0f5528496e62d2ff99de37d -I lz4 -xf /preloaded.tar -C /extractDir
I1002 07:00:34.090325 1486138 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/21643-1281649/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-docker-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-273808:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759382731-21643@sha256:ca1b4db171879edd6bbb9546a4b1afac2eb5be94a0f5528496e62d2ff99de37d -I lz4 -xf /preloaded.tar -C /extractDir: (4.023043428s)
I1002 07:00:34.090347 1486138 kic.go:203] duration metric: took 4.023190387s to extract preloaded images to volume ...
W1002 07:00:34.090486 1486138 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I1002 07:00:34.090603 1486138 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I1002 07:00:34.147926 1486138 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname scheduled-stop-273808 --name scheduled-stop-273808 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-273808 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=scheduled-stop-273808 --network scheduled-stop-273808 --ip 192.168.76.2 --volume scheduled-stop-273808:/var --security-opt apparmor=unconfined --memory=3072mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759382731-21643@sha256:ca1b4db171879edd6bbb9546a4b1afac2eb5be94a0f5528496e62d2ff99de37d
I1002 07:00:34.444430 1486138 cli_runner.go:164] Run: docker container inspect scheduled-stop-273808 --format={{.State.Running}}
I1002 07:00:34.468608 1486138 cli_runner.go:164] Run: docker container inspect scheduled-stop-273808 --format={{.State.Status}}
I1002 07:00:34.494284 1486138 cli_runner.go:164] Run: docker exec scheduled-stop-273808 stat /var/lib/dpkg/alternatives/iptables
I1002 07:00:34.546787 1486138 oci.go:144] the created container "scheduled-stop-273808" has a running status.
I1002 07:00:34.546825 1486138 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/21643-1281649/.minikube/machines/scheduled-stop-273808/id_rsa...
I1002 07:00:34.636229 1486138 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/21643-1281649/.minikube/machines/scheduled-stop-273808/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I1002 07:00:34.661798 1486138 cli_runner.go:164] Run: docker container inspect scheduled-stop-273808 --format={{.State.Status}}
I1002 07:00:34.684676 1486138 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I1002 07:00:34.684687 1486138 kic_runner.go:114] Args: [docker exec --privileged scheduled-stop-273808 chown docker:docker /home/docker/.ssh/authorized_keys]
I1002 07:00:34.736196 1486138 cli_runner.go:164] Run: docker container inspect scheduled-stop-273808 --format={{.State.Status}}
I1002 07:00:34.756667 1486138 machine.go:93] provisionDockerMachine start ...
I1002 07:00:34.756748 1486138 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-273808
I1002 07:00:34.793705 1486138 main.go:141] libmachine: Using SSH client type: native
I1002 07:00:34.794044 1486138 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef110] 0x3f18d0 <nil> [] 0s} 127.0.0.1 34154 <nil> <nil>}
I1002 07:00:34.794051 1486138 main.go:141] libmachine: About to run SSH command:
hostname
I1002 07:00:34.794852 1486138 main.go:141] libmachine: Error dialing TCP: ssh: handshake failed: read tcp 127.0.0.1:48734->127.0.0.1:34154: read: connection reset by peer
I1002 07:00:37.927912 1486138 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-273808
I1002 07:00:37.927926 1486138 ubuntu.go:182] provisioning hostname "scheduled-stop-273808"
I1002 07:00:37.927997 1486138 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-273808
I1002 07:00:37.945637 1486138 main.go:141] libmachine: Using SSH client type: native
I1002 07:00:37.945945 1486138 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef110] 0x3f18d0 <nil> [] 0s} 127.0.0.1 34154 <nil> <nil>}
I1002 07:00:37.945954 1486138 main.go:141] libmachine: About to run SSH command:
sudo hostname scheduled-stop-273808 && echo "scheduled-stop-273808" | sudo tee /etc/hostname
I1002 07:00:38.099171 1486138 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-273808
I1002 07:00:38.099244 1486138 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-273808
I1002 07:00:38.118974 1486138 main.go:141] libmachine: Using SSH client type: native
I1002 07:00:38.119296 1486138 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef110] 0x3f18d0 <nil> [] 0s} 127.0.0.1 34154 <nil> <nil>}
I1002 07:00:38.119319 1486138 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sscheduled-stop-273808' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 scheduled-stop-273808/g' /etc/hosts;
else
echo '127.0.1.1 scheduled-stop-273808' | sudo tee -a /etc/hosts;
fi
fi
I1002 07:00:38.252780 1486138 main.go:141] libmachine: SSH cmd err, output: <nil>:
I1002 07:00:38.252796 1486138 ubuntu.go:188] set auth options {CertDir:/home/jenkins/minikube-integration/21643-1281649/.minikube CaCertPath:/home/jenkins/minikube-integration/21643-1281649/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/21643-1281649/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/21643-1281649/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/21643-1281649/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/21643-1281649/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/21643-1281649/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/21643-1281649/.minikube}
I1002 07:00:38.252815 1486138 ubuntu.go:190] setting up certificates
I1002 07:00:38.252822 1486138 provision.go:84] configureAuth start
I1002 07:00:38.252882 1486138 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-273808
I1002 07:00:38.271176 1486138 provision.go:143] copyHostCerts
I1002 07:00:38.271239 1486138 exec_runner.go:144] found /home/jenkins/minikube-integration/21643-1281649/.minikube/cert.pem, removing ...
I1002 07:00:38.271246 1486138 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21643-1281649/.minikube/cert.pem
I1002 07:00:38.271326 1486138 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21643-1281649/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/21643-1281649/.minikube/cert.pem (1123 bytes)
I1002 07:00:38.271420 1486138 exec_runner.go:144] found /home/jenkins/minikube-integration/21643-1281649/.minikube/key.pem, removing ...
I1002 07:00:38.271424 1486138 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21643-1281649/.minikube/key.pem
I1002 07:00:38.271448 1486138 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21643-1281649/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/21643-1281649/.minikube/key.pem (1675 bytes)
I1002 07:00:38.271499 1486138 exec_runner.go:144] found /home/jenkins/minikube-integration/21643-1281649/.minikube/ca.pem, removing ...
I1002 07:00:38.271502 1486138 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21643-1281649/.minikube/ca.pem
I1002 07:00:38.271523 1486138 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21643-1281649/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/21643-1281649/.minikube/ca.pem (1082 bytes)
I1002 07:00:38.271567 1486138 provision.go:117] generating server cert: /home/jenkins/minikube-integration/21643-1281649/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/21643-1281649/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/21643-1281649/.minikube/certs/ca-key.pem org=jenkins.scheduled-stop-273808 san=[127.0.0.1 192.168.76.2 localhost minikube scheduled-stop-273808]
I1002 07:00:38.428487 1486138 provision.go:177] copyRemoteCerts
I1002 07:00:38.428545 1486138 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I1002 07:00:38.428582 1486138 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-273808
I1002 07:00:38.449290 1486138 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34154 SSHKeyPath:/home/jenkins/minikube-integration/21643-1281649/.minikube/machines/scheduled-stop-273808/id_rsa Username:docker}
I1002 07:00:38.551781 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/machines/server.pem --> /etc/docker/server.pem (1233 bytes)
I1002 07:00:38.568837 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I1002 07:00:38.585918 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
I1002 07:00:38.603108 1486138 provision.go:87] duration metric: took 350.262649ms to configureAuth
I1002 07:00:38.603139 1486138 ubuntu.go:206] setting minikube options for container-runtime
I1002 07:00:38.603307 1486138 config.go:182] Loaded profile config "scheduled-stop-273808": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.34.1
I1002 07:00:38.603370 1486138 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-273808
I1002 07:00:38.620240 1486138 main.go:141] libmachine: Using SSH client type: native
I1002 07:00:38.620549 1486138 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef110] 0x3f18d0 <nil> [] 0s} 127.0.0.1 34154 <nil> <nil>}
I1002 07:00:38.620556 1486138 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I1002 07:00:38.752885 1486138 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I1002 07:00:38.752896 1486138 ubuntu.go:71] root file system type: overlay
I1002 07:00:38.753025 1486138 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I1002 07:00:38.753091 1486138 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-273808
I1002 07:00:38.771401 1486138 main.go:141] libmachine: Using SSH client type: native
I1002 07:00:38.771706 1486138 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef110] 0x3f18d0 <nil> [] 0s} 127.0.0.1 34154 <nil> <nil>}
I1002 07:00:38.771785 1486138 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network-online.target nss-lookup.target docker.socket firewalld.service containerd.service time-set.target
Wants=network-online.target containerd.service
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=always
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 \
-H fd:// --containerd=/run/containerd/containerd.sock \
-H unix:///var/run/docker.sock \
--default-ulimit=nofile=1048576:1048576 \
--tlsverify \
--tlscacert /etc/docker/ca.pem \
--tlscert /etc/docker/server.pem \
--tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I1002 07:00:38.913533 1486138 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network-online.target nss-lookup.target docker.socket firewalld.service containerd.service time-set.target
Wants=network-online.target containerd.service
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=always
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H fd:// --containerd=/run/containerd/containerd.sock -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
I1002 07:00:38.913620 1486138 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-273808
I1002 07:00:38.935130 1486138 main.go:141] libmachine: Using SSH client type: native
I1002 07:00:38.935448 1486138 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef110] 0x3f18d0 <nil> [] 0s} 127.0.0.1 34154 <nil> <nil>}
I1002 07:00:38.935464 1486138 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I1002 07:00:39.880364 1486138 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2025-09-03 20:56:55.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2025-10-02 07:00:38.908454311 +0000
@@ -9,23 +9,34 @@
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
Restart=always
+
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H fd:// --containerd=/run/containerd/containerd.sock -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
+
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I1002 07:00:39.880380 1486138 machine.go:96] duration metric: took 5.12370194s to provisionDockerMachine
I1002 07:00:39.880391 1486138 client.go:171] duration metric: took 10.475997745s to LocalClient.Create
I1002 07:00:39.880408 1486138 start.go:167] duration metric: took 10.476059594s to libmachine.API.Create "scheduled-stop-273808"
I1002 07:00:39.880414 1486138 start.go:293] postStartSetup for "scheduled-stop-273808" (driver="docker")
I1002 07:00:39.880423 1486138 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I1002 07:00:39.880495 1486138 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I1002 07:00:39.880541 1486138 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-273808
I1002 07:00:39.898903 1486138 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34154 SSHKeyPath:/home/jenkins/minikube-integration/21643-1281649/.minikube/machines/scheduled-stop-273808/id_rsa Username:docker}
I1002 07:00:39.996453 1486138 ssh_runner.go:195] Run: cat /etc/os-release
I1002 07:00:40.000452 1486138 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I1002 07:00:40.000472 1486138 info.go:137] Remote host: Debian GNU/Linux 12 (bookworm)
I1002 07:00:40.000484 1486138 filesync.go:126] Scanning /home/jenkins/minikube-integration/21643-1281649/.minikube/addons for local assets ...
I1002 07:00:40.000553 1486138 filesync.go:126] Scanning /home/jenkins/minikube-integration/21643-1281649/.minikube/files for local assets ...
I1002 07:00:40.000636 1486138 filesync.go:149] local asset: /home/jenkins/minikube-integration/21643-1281649/.minikube/files/etc/ssl/certs/12835082.pem -> 12835082.pem in /etc/ssl/certs
I1002 07:00:40.000739 1486138 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I1002 07:00:40.058756 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/files/etc/ssl/certs/12835082.pem --> /etc/ssl/certs/12835082.pem (1708 bytes)
I1002 07:00:40.079308 1486138 start.go:296] duration metric: took 198.87888ms for postStartSetup
I1002 07:00:40.079706 1486138 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-273808
I1002 07:00:40.100591 1486138 profile.go:143] Saving config to /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/config.json ...
I1002 07:00:40.100895 1486138 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I1002 07:00:40.100937 1486138 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-273808
I1002 07:00:40.124879 1486138 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34154 SSHKeyPath:/home/jenkins/minikube-integration/21643-1281649/.minikube/machines/scheduled-stop-273808/id_rsa Username:docker}
I1002 07:00:40.221553 1486138 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I1002 07:00:40.227049 1486138 start.go:128] duration metric: took 10.826366919s to createHost
I1002 07:00:40.227063 1486138 start.go:83] releasing machines lock for "scheduled-stop-273808", held for 10.826473361s
I1002 07:00:40.227146 1486138 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-273808
I1002 07:00:40.244685 1486138 ssh_runner.go:195] Run: cat /version.json
I1002 07:00:40.244730 1486138 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-273808
I1002 07:00:40.244769 1486138 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I1002 07:00:40.244817 1486138 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-273808
I1002 07:00:40.262783 1486138 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34154 SSHKeyPath:/home/jenkins/minikube-integration/21643-1281649/.minikube/machines/scheduled-stop-273808/id_rsa Username:docker}
I1002 07:00:40.265401 1486138 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34154 SSHKeyPath:/home/jenkins/minikube-integration/21643-1281649/.minikube/machines/scheduled-stop-273808/id_rsa Username:docker}
I1002 07:00:40.452350 1486138 ssh_runner.go:195] Run: systemctl --version
I1002 07:00:40.459065 1486138 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
W1002 07:00:40.463682 1486138 cni.go:209] loopback cni configuration skipped: "/etc/cni/net.d/*loopback.conf*" not found
I1002 07:00:40.463747 1486138 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I1002 07:00:40.492800 1486138 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/10-crio-bridge.conflist.disabled] bridge cni config(s)
I1002 07:00:40.492817 1486138 start.go:495] detecting cgroup driver to use...
I1002 07:00:40.492850 1486138 detect.go:187] detected "cgroupfs" cgroup driver on host os
I1002 07:00:40.492947 1486138 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I1002 07:00:40.507710 1486138 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10.1"|' /etc/containerd/config.toml"
I1002 07:00:40.516872 1486138 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I1002 07:00:40.525991 1486138 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I1002 07:00:40.526054 1486138 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I1002 07:00:40.535019 1486138 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I1002 07:00:40.543889 1486138 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I1002 07:00:40.553079 1486138 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I1002 07:00:40.562407 1486138 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I1002 07:00:40.571018 1486138 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I1002 07:00:40.580198 1486138 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I1002 07:00:40.589765 1486138 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I1002 07:00:40.598813 1486138 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I1002 07:00:40.606214 1486138 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I1002 07:00:40.613592 1486138 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1002 07:00:40.728676 1486138 ssh_runner.go:195] Run: sudo systemctl restart containerd
I1002 07:00:40.825040 1486138 start.go:495] detecting cgroup driver to use...
I1002 07:00:40.825078 1486138 detect.go:187] detected "cgroupfs" cgroup driver on host os
I1002 07:00:40.825147 1486138 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I1002 07:00:40.840636 1486138 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I1002 07:00:40.854073 1486138 ssh_runner.go:195] Run: sudo systemctl stop -f containerd
I1002 07:00:40.877051 1486138 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I1002 07:00:40.890077 1486138 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I1002 07:00:40.903312 1486138 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I1002 07:00:40.917515 1486138 ssh_runner.go:195] Run: which cri-dockerd
I1002 07:00:40.921206 1486138 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I1002 07:00:40.929210 1486138 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (192 bytes)
I1002 07:00:40.943074 1486138 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I1002 07:00:41.054613 1486138 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I1002 07:00:41.184329 1486138 docker.go:575] configuring docker to use "cgroupfs" as cgroup driver...
I1002 07:00:41.184431 1486138 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I1002 07:00:41.199309 1486138 ssh_runner.go:195] Run: sudo systemctl reset-failed docker
I1002 07:00:41.213443 1486138 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1002 07:00:41.340309 1486138 ssh_runner.go:195] Run: sudo systemctl restart docker
I1002 07:00:41.721341 1486138 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
I1002 07:00:41.735221 1486138 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I1002 07:00:41.749348 1486138 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I1002 07:00:41.763203 1486138 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I1002 07:00:41.889514 1486138 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I1002 07:00:42.018034 1486138 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1002 07:00:42.152173 1486138 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I1002 07:00:42.173643 1486138 ssh_runner.go:195] Run: sudo systemctl reset-failed cri-docker.service
I1002 07:00:42.190441 1486138 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1002 07:00:42.322450 1486138 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I1002 07:00:42.394467 1486138 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I1002 07:00:42.410296 1486138 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I1002 07:00:42.410361 1486138 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I1002 07:00:42.414772 1486138 start.go:563] Will wait 60s for crictl version
I1002 07:00:42.414902 1486138 ssh_runner.go:195] Run: which crictl
I1002 07:00:42.419015 1486138 ssh_runner.go:195] Run: sudo /usr/local/bin/crictl version
I1002 07:00:42.444787 1486138 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 28.4.0
RuntimeApiVersion: v1
I1002 07:00:42.444857 1486138 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I1002 07:00:42.472493 1486138 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I1002 07:00:42.506380 1486138 out.go:252] * Preparing Kubernetes v1.34.1 on Docker 28.4.0 ...
I1002 07:00:42.506469 1486138 cli_runner.go:164] Run: docker network inspect scheduled-stop-273808 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I1002 07:00:42.523082 1486138 ssh_runner.go:195] Run: grep 192.168.76.1 host.minikube.internal$ /etc/hosts
I1002 07:00:42.527128 1486138 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.76.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I1002 07:00:42.537123 1486138 kubeadm.go:883] updating cluster {Name:scheduled-stop-273808 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759382731-21643@sha256:ca1b4db171879edd6bbb9546a4b1afac2eb5be94a0f5528496e62d2ff99de37d Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-273808 Namespace:default APIServerHAVIP: APIServerName:miniku
beCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock
: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I1002 07:00:42.537218 1486138 preload.go:183] Checking if preload exists for k8s version v1.34.1 and runtime docker
I1002 07:00:42.537270 1486138 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I1002 07:00:42.558093 1486138 docker.go:691] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.34.1
registry.k8s.io/kube-scheduler:v1.34.1
registry.k8s.io/kube-controller-manager:v1.34.1
registry.k8s.io/kube-proxy:v1.34.1
registry.k8s.io/etcd:3.6.4-0
registry.k8s.io/pause:3.10.1
registry.k8s.io/coredns/coredns:v1.12.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I1002 07:00:42.558107 1486138 docker.go:621] Images already preloaded, skipping extraction
I1002 07:00:42.558181 1486138 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I1002 07:00:42.576358 1486138 docker.go:691] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.34.1
registry.k8s.io/kube-controller-manager:v1.34.1
registry.k8s.io/kube-scheduler:v1.34.1
registry.k8s.io/kube-proxy:v1.34.1
registry.k8s.io/etcd:3.6.4-0
registry.k8s.io/pause:3.10.1
registry.k8s.io/coredns/coredns:v1.12.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I1002 07:00:42.576372 1486138 cache_images.go:85] Images are preloaded, skipping loading
I1002 07:00:42.576380 1486138 kubeadm.go:934] updating node { 192.168.76.2 8443 v1.34.1 docker true true} ...
I1002 07:00:42.576472 1486138 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.34.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=scheduled-stop-273808 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.76.2
[Install]
config:
{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-273808 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I1002 07:00:42.576537 1486138 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I1002 07:00:42.629523 1486138 cni.go:84] Creating CNI manager for ""
I1002 07:00:42.629537 1486138 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I1002 07:00:42.629551 1486138 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I1002 07:00:42.629571 1486138 kubeadm.go:189] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.76.2 APIServerPort:8443 KubernetesVersion:v1.34.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:scheduled-stop-273808 NodeName:scheduled-stop-273808 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.76.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.76.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPo
dPath:/etc/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I1002 07:00:42.629694 1486138 kubeadm.go:195] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta4
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.76.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "scheduled-stop-273808"
kubeletExtraArgs:
- name: "node-ip"
value: "192.168.76.2"
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta4
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.76.2"]
extraArgs:
- name: "enable-admission-plugins"
value: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
- name: "allocate-node-cidrs"
value: "true"
- name: "leader-elect"
value: "false"
scheduler:
extraArgs:
- name: "leader-elect"
value: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
kubernetesVersion: v1.34.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I1002 07:00:42.629758 1486138 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.34.1
I1002 07:00:42.637676 1486138 binaries.go:44] Found k8s binaries, skipping transfer
I1002 07:00:42.637749 1486138 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I1002 07:00:42.645984 1486138 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (320 bytes)
I1002 07:00:42.658748 1486138 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I1002 07:00:42.671461 1486138 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2222 bytes)
I1002 07:00:42.685392 1486138 ssh_runner.go:195] Run: grep 192.168.76.2 control-plane.minikube.internal$ /etc/hosts
I1002 07:00:42.689160 1486138 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.76.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I1002 07:00:42.698746 1486138 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1002 07:00:42.810554 1486138 ssh_runner.go:195] Run: sudo systemctl start kubelet
I1002 07:00:42.826654 1486138 certs.go:69] Setting up /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808 for IP: 192.168.76.2
I1002 07:00:42.826665 1486138 certs.go:195] generating shared ca certs ...
I1002 07:00:42.826680 1486138 certs.go:227] acquiring lock for ca certs: {Name:mkbfd31f90356176653bc4b00cb70c47296e672d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1002 07:00:42.826818 1486138 certs.go:236] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/21643-1281649/.minikube/ca.key
I1002 07:00:42.826868 1486138 certs.go:236] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/21643-1281649/.minikube/proxy-client-ca.key
I1002 07:00:42.826875 1486138 certs.go:257] generating profile certs ...
I1002 07:00:42.826931 1486138 certs.go:364] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/client.key
I1002 07:00:42.826948 1486138 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/client.crt with IP's: []
I1002 07:00:43.828499 1486138 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/client.crt ...
I1002 07:00:43.828516 1486138 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/client.crt: {Name:mkad8bf6bf5df4956e6e276d92e2cf162291b199 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1002 07:00:43.828731 1486138 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/client.key ...
I1002 07:00:43.828739 1486138 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/client.key: {Name:mk9b893fbf8fc5cd29a71a087c44b1227daeedaa Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1002 07:00:43.828835 1486138 certs.go:364] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/apiserver.key.f78285d2
I1002 07:00:43.828847 1486138 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/apiserver.crt.f78285d2 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.76.2]
I1002 07:00:44.576670 1486138 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/apiserver.crt.f78285d2 ...
I1002 07:00:44.576692 1486138 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/apiserver.crt.f78285d2: {Name:mk0c18c587ff003bf548ec4f72e78a054f31b17a Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1002 07:00:44.576887 1486138 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/apiserver.key.f78285d2 ...
I1002 07:00:44.576899 1486138 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/apiserver.key.f78285d2: {Name:mk43a37e7a10824e4f0b540bb3e0fef5a36bd2bf Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1002 07:00:44.576993 1486138 certs.go:382] copying /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/apiserver.crt.f78285d2 -> /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/apiserver.crt
I1002 07:00:44.577073 1486138 certs.go:386] copying /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/apiserver.key.f78285d2 -> /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/apiserver.key
I1002 07:00:44.577132 1486138 certs.go:364] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/proxy-client.key
I1002 07:00:44.577144 1486138 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/proxy-client.crt with IP's: []
I1002 07:00:45.086232 1486138 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/proxy-client.crt ...
I1002 07:00:45.086250 1486138 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/proxy-client.crt: {Name:mk8af19aa416a487eef66ba6b52167a9e3ae4eb2 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1002 07:00:45.086480 1486138 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/proxy-client.key ...
I1002 07:00:45.086490 1486138 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/proxy-client.key: {Name:mk7b4cc0f5f596b1a133130becc7235ba6aced70 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1002 07:00:45.086728 1486138 certs.go:484] found cert: /home/jenkins/minikube-integration/21643-1281649/.minikube/certs/1283508.pem (1338 bytes)
W1002 07:00:45.086774 1486138 certs.go:480] ignoring /home/jenkins/minikube-integration/21643-1281649/.minikube/certs/1283508_empty.pem, impossibly tiny 0 bytes
I1002 07:00:45.086784 1486138 certs.go:484] found cert: /home/jenkins/minikube-integration/21643-1281649/.minikube/certs/ca-key.pem (1675 bytes)
I1002 07:00:45.086820 1486138 certs.go:484] found cert: /home/jenkins/minikube-integration/21643-1281649/.minikube/certs/ca.pem (1082 bytes)
I1002 07:00:45.086855 1486138 certs.go:484] found cert: /home/jenkins/minikube-integration/21643-1281649/.minikube/certs/cert.pem (1123 bytes)
I1002 07:00:45.086891 1486138 certs.go:484] found cert: /home/jenkins/minikube-integration/21643-1281649/.minikube/certs/key.pem (1675 bytes)
I1002 07:00:45.086953 1486138 certs.go:484] found cert: /home/jenkins/minikube-integration/21643-1281649/.minikube/files/etc/ssl/certs/12835082.pem (1708 bytes)
I1002 07:00:45.087624 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I1002 07:00:45.119920 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I1002 07:00:45.143527 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I1002 07:00:45.171085 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I1002 07:00:45.206750 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1432 bytes)
I1002 07:00:45.237123 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I1002 07:00:45.277288 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I1002 07:00:45.310323 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/profiles/scheduled-stop-273808/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
I1002 07:00:45.344251 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/certs/1283508.pem --> /usr/share/ca-certificates/1283508.pem (1338 bytes)
I1002 07:00:45.378775 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/files/etc/ssl/certs/12835082.pem --> /usr/share/ca-certificates/12835082.pem (1708 bytes)
I1002 07:00:45.411523 1486138 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21643-1281649/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I1002 07:00:45.435470 1486138 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I1002 07:00:45.450252 1486138 ssh_runner.go:195] Run: openssl version
I1002 07:00:45.456970 1486138 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/1283508.pem && ln -fs /usr/share/ca-certificates/1283508.pem /etc/ssl/certs/1283508.pem"
I1002 07:00:45.466194 1486138 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/1283508.pem
I1002 07:00:45.470471 1486138 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 Oct 2 06:27 /usr/share/ca-certificates/1283508.pem
I1002 07:00:45.470536 1486138 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/1283508.pem
I1002 07:00:45.513901 1486138 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/1283508.pem /etc/ssl/certs/51391683.0"
I1002 07:00:45.522560 1486138 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/12835082.pem && ln -fs /usr/share/ca-certificates/12835082.pem /etc/ssl/certs/12835082.pem"
I1002 07:00:45.531247 1486138 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/12835082.pem
I1002 07:00:45.535042 1486138 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 Oct 2 06:27 /usr/share/ca-certificates/12835082.pem
I1002 07:00:45.535103 1486138 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/12835082.pem
I1002 07:00:45.577706 1486138 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/12835082.pem /etc/ssl/certs/3ec20f2e.0"
I1002 07:00:45.586703 1486138 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I1002 07:00:45.595566 1486138 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I1002 07:00:45.599979 1486138 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Oct 2 06:21 /usr/share/ca-certificates/minikubeCA.pem
I1002 07:00:45.600041 1486138 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I1002 07:00:45.641670 1486138 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I1002 07:00:45.649898 1486138 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I1002 07:00:45.653400 1486138 certs.go:400] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I1002 07:00:45.653440 1486138 kubeadm.go:400] StartCluster: {Name:scheduled-stop-273808 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759382731-21643@sha256:ca1b4db171879edd6bbb9546a4b1afac2eb5be94a0f5528496e62d2ff99de37d Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-273808 Namespace:default APIServerHAVIP: APIServerName:minikubeC
A APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: S
SHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I1002 07:00:45.653557 1486138 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I1002 07:00:45.672345 1486138 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I1002 07:00:45.680753 1486138 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I1002 07:00:45.688489 1486138 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I1002 07:00:45.688567 1486138 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I1002 07:00:45.696501 1486138 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I1002 07:00:45.696509 1486138 kubeadm.go:157] found existing configuration files:
I1002 07:00:45.696566 1486138 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I1002 07:00:45.704462 1486138 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I1002 07:00:45.704525 1486138 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I1002 07:00:45.711977 1486138 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I1002 07:00:45.720070 1486138 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I1002 07:00:45.720125 1486138 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I1002 07:00:45.727531 1486138 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I1002 07:00:45.735181 1486138 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I1002 07:00:45.735241 1486138 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I1002 07:00:45.742426 1486138 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I1002 07:00:45.749962 1486138 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I1002 07:00:45.750018 1486138 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I1002 07:00:45.757466 1486138 ssh_runner.go:286] Start: sudo /bin/bash -c "env PATH="/var/lib/minikube/binaries/v1.34.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I1002 07:00:45.835774 1486138 kubeadm.go:318] [init] Using Kubernetes version: v1.34.1
I1002 07:00:45.839286 1486138 kubeadm.go:318] [preflight] Running pre-flight checks
I1002 07:00:45.866193 1486138 kubeadm.go:318] [preflight] The system verification failed. Printing the output from the verification:
I1002 07:00:45.866259 1486138 kubeadm.go:318] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1084-aws[0m
I1002 07:00:45.866295 1486138 kubeadm.go:318] [0;37mOS[0m: [0;32mLinux[0m
I1002 07:00:45.866342 1486138 kubeadm.go:318] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I1002 07:00:45.866391 1486138 kubeadm.go:318] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I1002 07:00:45.866439 1486138 kubeadm.go:318] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I1002 07:00:45.866488 1486138 kubeadm.go:318] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I1002 07:00:45.866538 1486138 kubeadm.go:318] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I1002 07:00:45.866587 1486138 kubeadm.go:318] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I1002 07:00:45.866634 1486138 kubeadm.go:318] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I1002 07:00:45.866683 1486138 kubeadm.go:318] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I1002 07:00:45.866742 1486138 kubeadm.go:318] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I1002 07:00:45.937988 1486138 kubeadm.go:318] [preflight] Pulling images required for setting up a Kubernetes cluster
I1002 07:00:45.938108 1486138 kubeadm.go:318] [preflight] This might take a minute or two, depending on the speed of your internet connection
I1002 07:00:45.938210 1486138 kubeadm.go:318] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I1002 07:00:45.952373 1486138 kubeadm.go:318] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I1002 07:00:45.958935 1486138 out.go:252] - Generating certificates and keys ...
I1002 07:00:45.959049 1486138 kubeadm.go:318] [certs] Using existing ca certificate authority
I1002 07:00:45.959124 1486138 kubeadm.go:318] [certs] Using existing apiserver certificate and key on disk
I1002 07:00:46.905997 1486138 kubeadm.go:318] [certs] Generating "apiserver-kubelet-client" certificate and key
I1002 07:00:47.876546 1486138 kubeadm.go:318] [certs] Generating "front-proxy-ca" certificate and key
I1002 07:00:48.437150 1486138 kubeadm.go:318] [certs] Generating "front-proxy-client" certificate and key
I1002 07:00:48.813811 1486138 kubeadm.go:318] [certs] Generating "etcd/ca" certificate and key
I1002 07:00:49.578289 1486138 kubeadm.go:318] [certs] Generating "etcd/server" certificate and key
I1002 07:00:49.578576 1486138 kubeadm.go:318] [certs] etcd/server serving cert is signed for DNS names [localhost scheduled-stop-273808] and IPs [192.168.76.2 127.0.0.1 ::1]
I1002 07:00:50.161165 1486138 kubeadm.go:318] [certs] Generating "etcd/peer" certificate and key
I1002 07:00:50.161479 1486138 kubeadm.go:318] [certs] etcd/peer serving cert is signed for DNS names [localhost scheduled-stop-273808] and IPs [192.168.76.2 127.0.0.1 ::1]
I1002 07:00:50.379788 1486138 kubeadm.go:318] [certs] Generating "etcd/healthcheck-client" certificate and key
I1002 07:00:50.978125 1486138 kubeadm.go:318] [certs] Generating "apiserver-etcd-client" certificate and key
I1002 07:00:51.357695 1486138 kubeadm.go:318] [certs] Generating "sa" key and public key
I1002 07:00:51.357943 1486138 kubeadm.go:318] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I1002 07:00:52.428268 1486138 kubeadm.go:318] [kubeconfig] Writing "admin.conf" kubeconfig file
I1002 07:00:53.334705 1486138 kubeadm.go:318] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I1002 07:00:54.493494 1486138 kubeadm.go:318] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I1002 07:00:54.580501 1486138 kubeadm.go:318] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I1002 07:00:55.220905 1486138 kubeadm.go:318] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I1002 07:00:55.221667 1486138 kubeadm.go:318] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I1002 07:00:55.226007 1486138 kubeadm.go:318] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I1002 07:00:55.229646 1486138 out.go:252] - Booting up control plane ...
I1002 07:00:55.229765 1486138 kubeadm.go:318] [control-plane] Creating static Pod manifest for "kube-apiserver"
I1002 07:00:55.229862 1486138 kubeadm.go:318] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I1002 07:00:55.230347 1486138 kubeadm.go:318] [control-plane] Creating static Pod manifest for "kube-scheduler"
I1002 07:00:55.249081 1486138 kubeadm.go:318] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I1002 07:00:55.249193 1486138 kubeadm.go:318] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/instance-config.yaml"
I1002 07:00:55.258310 1486138 kubeadm.go:318] [patches] Applied patch of type "application/strategic-merge-patch+json" to target "kubeletconfiguration"
I1002 07:00:55.258595 1486138 kubeadm.go:318] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I1002 07:00:55.258638 1486138 kubeadm.go:318] [kubelet-start] Starting the kubelet
I1002 07:00:55.409908 1486138 kubeadm.go:318] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I1002 07:00:55.410026 1486138 kubeadm.go:318] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I1002 07:00:56.910786 1486138 kubeadm.go:318] [kubelet-check] The kubelet is healthy after 1.501189538s
I1002 07:00:56.917503 1486138 kubeadm.go:318] [control-plane-check] Waiting for healthy control plane components. This can take up to 4m0s
I1002 07:00:56.917605 1486138 kubeadm.go:318] [control-plane-check] Checking kube-apiserver at https://192.168.76.2:8443/livez
I1002 07:00:56.917947 1486138 kubeadm.go:318] [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz
I1002 07:00:56.918032 1486138 kubeadm.go:318] [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez
I1002 07:01:00.796981 1486138 kubeadm.go:318] [control-plane-check] kube-controller-manager is healthy after 3.878962763s
I1002 07:01:02.148716 1486138 kubeadm.go:318] [control-plane-check] kube-scheduler is healthy after 5.231228845s
I1002 07:01:03.419066 1486138 kubeadm.go:318] [control-plane-check] kube-apiserver is healthy after 6.501311437s
I1002 07:01:03.438401 1486138 kubeadm.go:318] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I1002 07:01:03.464316 1486138 kubeadm.go:318] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I1002 07:01:03.478659 1486138 kubeadm.go:318] [upload-certs] Skipping phase. Please see --upload-certs
I1002 07:01:03.478864 1486138 kubeadm.go:318] [mark-control-plane] Marking the node scheduled-stop-273808 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I1002 07:01:03.490461 1486138 kubeadm.go:318] [bootstrap-token] Using token: 5a9p6v.4zs45lxm4jmuczna
I1002 07:01:03.493359 1486138 out.go:252] - Configuring RBAC rules ...
I1002 07:01:03.493499 1486138 kubeadm.go:318] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I1002 07:01:03.497673 1486138 kubeadm.go:318] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I1002 07:01:03.509512 1486138 kubeadm.go:318] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I1002 07:01:03.514185 1486138 kubeadm.go:318] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I1002 07:01:03.520276 1486138 kubeadm.go:318] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I1002 07:01:03.524932 1486138 kubeadm.go:318] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I1002 07:01:03.828719 1486138 kubeadm.go:318] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I1002 07:01:04.256393 1486138 kubeadm.go:318] [addons] Applied essential addon: CoreDNS
I1002 07:01:04.826510 1486138 kubeadm.go:318] [addons] Applied essential addon: kube-proxy
I1002 07:01:04.827539 1486138 kubeadm.go:318]
I1002 07:01:04.827607 1486138 kubeadm.go:318] Your Kubernetes control-plane has initialized successfully!
I1002 07:01:04.827611 1486138 kubeadm.go:318]
I1002 07:01:04.827691 1486138 kubeadm.go:318] To start using your cluster, you need to run the following as a regular user:
I1002 07:01:04.827695 1486138 kubeadm.go:318]
I1002 07:01:04.827721 1486138 kubeadm.go:318] mkdir -p $HOME/.kube
I1002 07:01:04.827782 1486138 kubeadm.go:318] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I1002 07:01:04.827834 1486138 kubeadm.go:318] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I1002 07:01:04.827838 1486138 kubeadm.go:318]
I1002 07:01:04.827894 1486138 kubeadm.go:318] Alternatively, if you are the root user, you can run:
I1002 07:01:04.827897 1486138 kubeadm.go:318]
I1002 07:01:04.827946 1486138 kubeadm.go:318] export KUBECONFIG=/etc/kubernetes/admin.conf
I1002 07:01:04.827950 1486138 kubeadm.go:318]
I1002 07:01:04.828003 1486138 kubeadm.go:318] You should now deploy a pod network to the cluster.
I1002 07:01:04.828127 1486138 kubeadm.go:318] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I1002 07:01:04.828198 1486138 kubeadm.go:318] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I1002 07:01:04.828202 1486138 kubeadm.go:318]
I1002 07:01:04.828289 1486138 kubeadm.go:318] You can now join any number of control-plane nodes by copying certificate authorities
I1002 07:01:04.828368 1486138 kubeadm.go:318] and service account keys on each node and then running the following as root:
I1002 07:01:04.828372 1486138 kubeadm.go:318]
I1002 07:01:04.828459 1486138 kubeadm.go:318] kubeadm join control-plane.minikube.internal:8443 --token 5a9p6v.4zs45lxm4jmuczna \
I1002 07:01:04.828565 1486138 kubeadm.go:318] --discovery-token-ca-cert-hash sha256:08445900f37e2db8182acd70d8e2a7711d9c9dbd967d5e887a6f3a131bca817f \
I1002 07:01:04.828585 1486138 kubeadm.go:318] --control-plane
I1002 07:01:04.828589 1486138 kubeadm.go:318]
I1002 07:01:04.828676 1486138 kubeadm.go:318] Then you can join any number of worker nodes by running the following on each as root:
I1002 07:01:04.828680 1486138 kubeadm.go:318]
I1002 07:01:04.828765 1486138 kubeadm.go:318] kubeadm join control-plane.minikube.internal:8443 --token 5a9p6v.4zs45lxm4jmuczna \
I1002 07:01:04.828871 1486138 kubeadm.go:318] --discovery-token-ca-cert-hash sha256:08445900f37e2db8182acd70d8e2a7711d9c9dbd967d5e887a6f3a131bca817f
I1002 07:01:04.832787 1486138 kubeadm.go:318] [WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2
I1002 07:01:04.833032 1486138 kubeadm.go:318] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1084-aws\n", err: exit status 1
I1002 07:01:04.833144 1486138 kubeadm.go:318] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I1002 07:01:04.833172 1486138 cni.go:84] Creating CNI manager for ""
I1002 07:01:04.833184 1486138 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I1002 07:01:04.836425 1486138 out.go:179] * Configuring bridge CNI (Container Networking Interface) ...
I1002 07:01:04.839280 1486138 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I1002 07:01:04.846926 1486138 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I1002 07:01:04.861914 1486138 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I1002 07:01:04.862000 1486138 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I1002 07:01:04.862035 1486138 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes scheduled-stop-273808 minikube.k8s.io/updated_at=2025_10_02T07_01_04_0700 minikube.k8s.io/version=v1.37.0 minikube.k8s.io/commit=e67b65c0f4e92b22cf6bb9baed3c99d519c7afdb minikube.k8s.io/name=scheduled-stop-273808 minikube.k8s.io/primary=true
I1002 07:01:05.014014 1486138 kubeadm.go:1113] duration metric: took 152.073025ms to wait for elevateKubeSystemPrivileges
I1002 07:01:05.014049 1486138 ops.go:34] apiserver oom_adj: -16
I1002 07:01:05.014057 1486138 kubeadm.go:402] duration metric: took 19.360620014s to StartCluster
I1002 07:01:05.014072 1486138 settings.go:142] acquiring lock: {Name:mk549d445ee28d1b957693d6dbb26b038b2321bc Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1002 07:01:05.014139 1486138 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/21643-1281649/kubeconfig
I1002 07:01:05.014884 1486138 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21643-1281649/kubeconfig: {Name:mk9b20f6b5831bf91495a692140571471f3eef6b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1002 07:01:05.015157 1486138 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I1002 07:01:05.015269 1486138 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I1002 07:01:05.015537 1486138 config.go:182] Loaded profile config "scheduled-stop-273808": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.34.1
I1002 07:01:05.015573 1486138 addons.go:511] enable addons start: toEnable=map[ambassador:false amd-gpu-device-plugin:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubetail:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-rancher:false volcano:false volumesnapshots:false yakd:false]
I1002 07:01:05.015636 1486138 addons.go:69] Setting storage-provisioner=true in profile "scheduled-stop-273808"
I1002 07:01:05.015654 1486138 addons.go:238] Setting addon storage-provisioner=true in "scheduled-stop-273808"
I1002 07:01:05.015675 1486138 host.go:66] Checking if "scheduled-stop-273808" exists ...
I1002 07:01:05.016074 1486138 addons.go:69] Setting default-storageclass=true in profile "scheduled-stop-273808"
I1002 07:01:05.016089 1486138 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "scheduled-stop-273808"
I1002 07:01:05.016388 1486138 cli_runner.go:164] Run: docker container inspect scheduled-stop-273808 --format={{.State.Status}}
I1002 07:01:05.016639 1486138 cli_runner.go:164] Run: docker container inspect scheduled-stop-273808 --format={{.State.Status}}
I1002 07:01:05.019006 1486138 out.go:179] * Verifying Kubernetes components...
I1002 07:01:05.022052 1486138 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1002 07:01:05.070033 1486138 out.go:179] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I1002 07:01:05.072213 1486138 addons.go:238] Setting addon default-storageclass=true in "scheduled-stop-273808"
I1002 07:01:05.072242 1486138 host.go:66] Checking if "scheduled-stop-273808" exists ...
I1002 07:01:05.072672 1486138 cli_runner.go:164] Run: docker container inspect scheduled-stop-273808 --format={{.State.Status}}
I1002 07:01:05.073095 1486138 addons.go:435] installing /etc/kubernetes/addons/storage-provisioner.yaml
I1002 07:01:05.073104 1486138 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I1002 07:01:05.073151 1486138 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-273808
I1002 07:01:05.108741 1486138 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34154 SSHKeyPath:/home/jenkins/minikube-integration/21643-1281649/.minikube/machines/scheduled-stop-273808/id_rsa Username:docker}
I1002 07:01:05.116194 1486138 addons.go:435] installing /etc/kubernetes/addons/storageclass.yaml
I1002 07:01:05.116207 1486138 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I1002 07:01:05.116272 1486138 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-273808
I1002 07:01:05.141398 1486138 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34154 SSHKeyPath:/home/jenkins/minikube-integration/21643-1281649/.minikube/machines/scheduled-stop-273808/id_rsa Username:docker}
I1002 07:01:05.307808 1486138 ssh_runner.go:195] Run: sudo systemctl start kubelet
I1002 07:01:05.308026 1486138 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.76.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I1002 07:01:05.337651 1486138 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.34.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I1002 07:01:05.416106 1486138 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.34.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I1002 07:01:05.808267 1486138 start.go:976] {"host.minikube.internal": 192.168.76.1} host record injected into CoreDNS's ConfigMap
I1002 07:01:05.809988 1486138 api_server.go:52] waiting for apiserver process to appear ...
I1002 07:01:05.810036 1486138 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I1002 07:01:06.005305 1486138 api_server.go:72] duration metric: took 990.119406ms to wait for apiserver process to appear ...
I1002 07:01:06.005319 1486138 api_server.go:88] waiting for apiserver healthz status ...
I1002 07:01:06.005337 1486138 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
I1002 07:01:06.021187 1486138 api_server.go:279] https://192.168.76.2:8443/healthz returned 200:
ok
I1002 07:01:06.022554 1486138 api_server.go:141] control plane version: v1.34.1
I1002 07:01:06.022572 1486138 api_server.go:131] duration metric: took 17.248151ms to wait for apiserver health ...
I1002 07:01:06.022580 1486138 system_pods.go:43] waiting for kube-system pods to appear ...
I1002 07:01:06.027098 1486138 system_pods.go:59] 5 kube-system pods found
I1002 07:01:06.027121 1486138 system_pods.go:61] "etcd-scheduled-stop-273808" [1c2a2789-fa1c-4bf0-ada8-41f2d004d52e] Running / Ready:ContainersNotReady (containers with unready status: [etcd]) / ContainersReady:ContainersNotReady (containers with unready status: [etcd])
I1002 07:01:06.027129 1486138 system_pods.go:61] "kube-apiserver-scheduled-stop-273808" [3fc5b5cb-05f0-4bf1-b437-069c7bcc10d1] Running / Ready:ContainersNotReady (containers with unready status: [kube-apiserver]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-apiserver])
I1002 07:01:06.027136 1486138 system_pods.go:61] "kube-controller-manager-scheduled-stop-273808" [cbd59581-ddb6-4cd6-beb6-79c028093c37] Running / Ready:ContainersNotReady (containers with unready status: [kube-controller-manager]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-controller-manager])
I1002 07:01:06.027142 1486138 system_pods.go:61] "kube-scheduler-scheduled-stop-273808" [336f5874-5d3e-4158-8f2e-923382d3d81e] Running / Ready:ContainersNotReady (containers with unready status: [kube-scheduler]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-scheduler])
I1002 07:01:06.027149 1486138 system_pods.go:61] "storage-provisioner" [1f35ade9-b08c-4065-9962-964aa478420d] Pending: PodScheduled:Unschedulable (0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. no new claims to deallocate, preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.)
I1002 07:01:06.027158 1486138 system_pods.go:74] duration metric: took 4.569448ms to wait for pod list to return data ...
I1002 07:01:06.027169 1486138 kubeadm.go:586] duration metric: took 1.011989442s to wait for: map[apiserver:true system_pods:true]
I1002 07:01:06.027182 1486138 node_conditions.go:102] verifying NodePressure condition ...
I1002 07:01:06.027436 1486138 out.go:179] * Enabled addons: storage-provisioner, default-storageclass
I1002 07:01:06.030794 1486138 addons.go:514] duration metric: took 1.015200411s for enable addons: enabled=[storage-provisioner default-storageclass]
I1002 07:01:06.031354 1486138 node_conditions.go:122] node storage ephemeral capacity is 203034800Ki
I1002 07:01:06.031376 1486138 node_conditions.go:123] node cpu capacity is 2
I1002 07:01:06.031407 1486138 node_conditions.go:105] duration metric: took 4.203596ms to run NodePressure ...
I1002 07:01:06.031419 1486138 start.go:241] waiting for startup goroutines ...
I1002 07:01:06.314988 1486138 kapi.go:214] "coredns" deployment in "kube-system" namespace and "scheduled-stop-273808" context rescaled to 1 replicas
I1002 07:01:06.315009 1486138 start.go:246] waiting for cluster config update ...
I1002 07:01:06.315019 1486138 start.go:255] writing updated cluster config ...
I1002 07:01:06.315307 1486138 ssh_runner.go:195] Run: rm -f paused
I1002 07:01:06.381275 1486138 start.go:623] kubectl: 1.33.2, cluster: 1.34.1 (minor skew: 1)
I1002 07:01:06.384385 1486138 out.go:179] * Done! kubectl is now configured to use "scheduled-stop-273808" cluster and "default" namespace by default
==> Docker <==
Oct 02 07:00:41 scheduled-stop-273808 dockerd[1130]: time="2025-10-02T07:00:41.682966338Z" level=info msg="Loading containers: done."
Oct 02 07:00:41 scheduled-stop-273808 dockerd[1130]: time="2025-10-02T07:00:41.693883888Z" level=info msg="Docker daemon" commit=249d679 containerd-snapshotter=false storage-driver=overlay2 version=28.4.0
Oct 02 07:00:41 scheduled-stop-273808 dockerd[1130]: time="2025-10-02T07:00:41.693960949Z" level=info msg="Initializing buildkit"
Oct 02 07:00:41 scheduled-stop-273808 dockerd[1130]: time="2025-10-02T07:00:41.713273918Z" level=info msg="Completed buildkit initialization"
Oct 02 07:00:41 scheduled-stop-273808 dockerd[1130]: time="2025-10-02T07:00:41.718695229Z" level=info msg="Daemon has completed initialization"
Oct 02 07:00:41 scheduled-stop-273808 dockerd[1130]: time="2025-10-02T07:00:41.718791137Z" level=info msg="API listen on /run/docker.sock"
Oct 02 07:00:41 scheduled-stop-273808 dockerd[1130]: time="2025-10-02T07:00:41.718885994Z" level=info msg="API listen on /var/run/docker.sock"
Oct 02 07:00:41 scheduled-stop-273808 dockerd[1130]: time="2025-10-02T07:00:41.719002472Z" level=info msg="API listen on [::]:2376"
Oct 02 07:00:41 scheduled-stop-273808 systemd[1]: Started docker.service - Docker Application Container Engine.
Oct 02 07:00:42 scheduled-stop-273808 systemd[1]: Starting cri-docker.service - CRI Interface for Docker Application Container Engine...
Oct 02 07:00:42 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:42Z" level=info msg="Starting cri-dockerd dev (HEAD)"
Oct 02 07:00:42 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:42Z" level=info msg="Connecting to docker on the Endpoint unix:///var/run/docker.sock"
Oct 02 07:00:42 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:42Z" level=info msg="Start docker client with request timeout 0s"
Oct 02 07:00:42 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:42Z" level=info msg="Hairpin mode is set to hairpin-veth"
Oct 02 07:00:42 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:42Z" level=info msg="Loaded network plugin cni"
Oct 02 07:00:42 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:42Z" level=info msg="Docker cri networking managed by network plugin cni"
Oct 02 07:00:42 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:42Z" level=info msg="Setting cgroupDriver cgroupfs"
Oct 02 07:00:42 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:42Z" level=info msg="Docker cri received runtime config &RuntimeConfig{NetworkConfig:&NetworkConfig{PodCidr:,},}"
Oct 02 07:00:42 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:42Z" level=info msg="Starting the GRPC backend for the Docker CRI interface."
Oct 02 07:00:42 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:42Z" level=info msg="Start cri-dockerd grpc backend"
Oct 02 07:00:42 scheduled-stop-273808 systemd[1]: Started cri-docker.service - CRI Interface for Docker Application Container Engine.
Oct 02 07:00:57 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:57Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/b9c970430a8cd508cf9d1c623b60f95d5123c49f79941046ec1965501ac5eb89/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options edns0 trust-ad ndots:0]"
Oct 02 07:00:57 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:57Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/73c8c95438f48a3b3a90dca8ee25ff47ebaa9df7fae1e6aad925e9a97d1aebf6/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options trust-ad ndots:0 edns0]"
Oct 02 07:00:57 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:57Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/c96edf8923403862cb48d037c6e44c2045eafc8777fa10d285555dfd227792f8/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options ndots:0 edns0 trust-ad]"
Oct 02 07:00:57 scheduled-stop-273808 cri-dockerd[1431]: time="2025-10-02T07:00:57Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/18b9f417dac0d1afa00eaaab11ae503b02d6275c5260f061044e9b4d578a487b/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options edns0 trust-ad ndots:0]"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD NAMESPACE
139929cebf074 b5f57ec6b9867 10 seconds ago Running kube-scheduler 0 18b9f417dac0d kube-scheduler-scheduled-stop-273808 kube-system
2023557716beb a1894772a478e 10 seconds ago Running etcd 0 c96edf8923403 etcd-scheduled-stop-273808 kube-system
45e4555fe66b7 43911e833d64d 10 seconds ago Running kube-apiserver 0 73c8c95438f48 kube-apiserver-scheduled-stop-273808 kube-system
d0fb3d5ca8854 7eb2c6ff0c5a7 10 seconds ago Running kube-controller-manager 0 b9c970430a8cd kube-controller-manager-scheduled-stop-273808 kube-system
==> describe nodes <==
Name: scheduled-stop-273808
Roles: control-plane
Labels: beta.kubernetes.io/arch=arm64
beta.kubernetes.io/os=linux
kubernetes.io/arch=arm64
kubernetes.io/hostname=scheduled-stop-273808
kubernetes.io/os=linux
minikube.k8s.io/commit=e67b65c0f4e92b22cf6bb9baed3c99d519c7afdb
minikube.k8s.io/name=scheduled-stop-273808
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2025_10_02T07_01_04_0700
minikube.k8s.io/version=v1.37.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
Annotations: volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Thu, 02 Oct 2025 07:01:01 +0000
Taints: node.kubernetes.io/not-ready:NoSchedule
Unschedulable: false
Lease:
HolderIdentity: scheduled-stop-273808
AcquireTime: <unset>
RenewTime: Thu, 02 Oct 2025 07:01:04 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Thu, 02 Oct 2025 07:01:07 +0000 Thu, 02 Oct 2025 07:00:58 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Thu, 02 Oct 2025 07:01:07 +0000 Thu, 02 Oct 2025 07:00:58 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Thu, 02 Oct 2025 07:01:07 +0000 Thu, 02 Oct 2025 07:00:58 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Thu, 02 Oct 2025 07:01:07 +0000 Thu, 02 Oct 2025 07:01:07 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.76.2
Hostname: scheduled-stop-273808
Capacity:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022300Ki
pods: 110
Allocatable:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022300Ki
pods: 110
System Info:
Machine ID: 56c28623810142b3b2babf1c6bdeee00
System UUID: 99de0c25-7938-4411-88a3-aa41d41e6d49
Boot ID: 07f149ce-ad12-470a-acc5-7e688ae5314a
Kernel Version: 5.15.0-1084-aws
OS Image: Debian GNU/Linux 12 (bookworm)
Operating System: linux
Architecture: arm64
Container Runtime Version: docker://28.4.0
Kubelet Version: v1.34.1
Kube-Proxy Version:
Non-terminated Pods: (4 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
kube-system etcd-scheduled-stop-273808 100m (5%) 0 (0%) 100Mi (1%) 0 (0%) 4s
kube-system kube-apiserver-scheduled-stop-273808 250m (12%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-controller-manager-scheduled-stop-273808 200m (10%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-scheduler-scheduled-stop-273808 100m (5%) 0 (0%) 0 (0%) 0 (0%) 6s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 650m (32%) 0 (0%)
memory 100Mi (1%) 0 (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
hugepages-32Mi 0 (0%) 0 (0%)
hugepages-64Ki 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 4s kubelet Starting kubelet.
Warning CgroupV1 4s kubelet cgroup v1 support is in maintenance mode, please migrate to cgroup v2
Normal NodeAllocatableEnforced 4s kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 4s kubelet Node scheduled-stop-273808 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 4s kubelet Node scheduled-stop-273808 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 4s kubelet Node scheduled-stop-273808 status is now: NodeHasSufficientPID
Normal NodeReady 1s kubelet Node scheduled-stop-273808 status is now: NodeReady
==> dmesg <==
[Oct 2 06:06] systemd-journald[223]: Failed to send stream file descriptor to service manager: Connection refused
[Oct 2 06:20] kauditd_printk_skb: 8 callbacks suppressed
==> etcd [2023557716be] <==
{"level":"warn","ts":"2025-10-02T07:00:59.413615Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39760","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.429319Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39788","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.453257Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39796","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.465419Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39810","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.482789Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39830","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.499148Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39840","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.536428Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39854","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.545421Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39866","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.561990Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39882","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.585445Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39896","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.625521Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39922","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.651004Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39932","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.663960Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39950","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.686421Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39966","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.697241Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39974","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.716789Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:39990","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.734127Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:40002","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.770128Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:40016","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.802826Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:40036","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.830309Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:40056","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.881473Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:40086","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.899814Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:40116","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.925916Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:40132","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:00:59.972114Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:40148","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-02T07:01:00.165781Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:40172","server-name":"","error":"EOF"}
==> kernel <==
07:01:08 up 6:43, 0 user, load average: 2.67, 2.73, 3.15
Linux scheduled-stop-273808 5.15.0-1084-aws #91~20.04.1-Ubuntu SMP Fri May 2 07:00:04 UTC 2025 aarch64 GNU/Linux
PRETTY_NAME="Debian GNU/Linux 12 (bookworm)"
==> kube-apiserver [45e4555fe66b] <==
I1002 07:01:01.372651 1 cache.go:39] Caches are synced for autoregister controller
I1002 07:01:01.373000 1 shared_informer.go:356] "Caches are synced" controller="node_authorizer"
I1002 07:01:01.380844 1 shared_informer.go:356] "Caches are synced" controller="ipallocator-repair-controller"
I1002 07:01:01.384927 1 shared_informer.go:356] "Caches are synced" controller="*generic.policySource[*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicy,*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicyBinding,k8s.io/apiserver/pkg/admission/plugin/policy/validating.Validator]"
I1002 07:01:01.385061 1 policy_source.go:240] refreshing policies
E1002 07:01:01.422850 1 controller.go:148] "Unhandled Error" err="while syncing ConfigMap \"kube-system/kube-apiserver-legacy-service-account-token-tracking\", err: namespaces \"kube-system\" not found" logger="UnhandledError"
I1002 07:01:01.438875 1 controller.go:667] quota admission added evaluator for: namespaces
I1002 07:01:01.484338 1 controller.go:667] quota admission added evaluator for: leases.coordination.k8s.io
I1002 07:01:01.486876 1 cidrallocator.go:301] created ClusterIP allocator for Service CIDR 10.96.0.0/12
I1002 07:01:01.487386 1 default_servicecidr_controller.go:228] Setting default ServiceCIDR condition Ready to True
I1002 07:01:01.505760 1 cidrallocator.go:277] updated ClusterIP allocator for Service CIDR 10.96.0.0/12
I1002 07:01:01.527831 1 default_servicecidr_controller.go:137] Shutting down kubernetes-service-cidr-controller
I1002 07:01:02.159729 1 storage_scheduling.go:95] created PriorityClass system-node-critical with value 2000001000
I1002 07:01:02.166706 1 storage_scheduling.go:95] created PriorityClass system-cluster-critical with value 2000000000
I1002 07:01:02.166733 1 storage_scheduling.go:111] all system priority classes are created successfully or already exist.
I1002 07:01:02.837170 1 controller.go:667] quota admission added evaluator for: roles.rbac.authorization.k8s.io
I1002 07:01:02.894785 1 controller.go:667] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
I1002 07:01:02.969051 1 alloc.go:328] "allocated clusterIPs" service="default/kubernetes" clusterIPs={"IPv4":"10.96.0.1"}
W1002 07:01:02.977098 1 lease.go:265] Resetting endpoints for master service "kubernetes" to [192.168.76.2]
I1002 07:01:02.978449 1 controller.go:667] quota admission added evaluator for: endpoints
I1002 07:01:02.983709 1 controller.go:667] quota admission added evaluator for: endpointslices.discovery.k8s.io
I1002 07:01:03.258052 1 controller.go:667] quota admission added evaluator for: serviceaccounts
I1002 07:01:04.236210 1 controller.go:667] quota admission added evaluator for: deployments.apps
I1002 07:01:04.253935 1 alloc.go:328] "allocated clusterIPs" service="kube-system/kube-dns" clusterIPs={"IPv4":"10.96.0.10"}
I1002 07:01:04.265659 1 controller.go:667] quota admission added evaluator for: daemonsets.apps
==> kube-controller-manager [d0fb3d5ca885] <==
I1002 07:01:07.319762 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="poddisruptionbudgets.policy"
I1002 07:01:07.319782 1 controllermanager.go:781] "Started controller" controller="resourcequota-controller"
I1002 07:01:07.319850 1 resource_quota_controller.go:300] "Starting resource quota controller" logger="resourcequota-controller"
I1002 07:01:07.319861 1 shared_informer.go:349] "Waiting for caches to sync" controller="resource quota"
I1002 07:01:07.319880 1 resource_quota_monitor.go:308] "QuotaMonitor running" logger="resourcequota-controller"
I1002 07:01:07.359567 1 controllermanager.go:781] "Started controller" controller="certificatesigningrequest-approving-controller"
I1002 07:01:07.359596 1 controllermanager.go:739] "Skipping a cloud provider controller" controller="cloud-node-lifecycle-controller"
I1002 07:01:07.364319 1 certificate_controller.go:120] "Starting certificate controller" logger="certificatesigningrequest-approving-controller" name="csrapproving"
I1002 07:01:07.364347 1 shared_informer.go:349] "Waiting for caches to sync" controller="certificate-csrapproving"
I1002 07:01:07.554500 1 controllermanager.go:781] "Started controller" controller="validatingadmissionpolicy-status-controller"
I1002 07:01:07.554536 1 controllermanager.go:739] "Skipping a cloud provider controller" controller="service-lb-controller"
I1002 07:01:07.554588 1 shared_informer.go:349] "Waiting for caches to sync" controller="validatingadmissionpolicy-status"
I1002 07:01:07.705970 1 controllermanager.go:781] "Started controller" controller="replicationcontroller-controller"
I1002 07:01:07.706123 1 replica_set.go:243] "Starting controller" logger="replicationcontroller-controller" name="replicationcontroller"
I1002 07:01:07.706133 1 shared_informer.go:349] "Waiting for caches to sync" controller="ReplicationController"
I1002 07:01:07.856433 1 controllermanager.go:781] "Started controller" controller="pod-garbage-collector-controller"
I1002 07:01:07.856556 1 gc_controller.go:99] "Starting GC controller" logger="pod-garbage-collector-controller"
I1002 07:01:07.856572 1 shared_informer.go:349] "Waiting for caches to sync" controller="GC"
I1002 07:01:08.010505 1 controllermanager.go:781] "Started controller" controller="replicaset-controller"
I1002 07:01:08.010579 1 replica_set.go:243] "Starting controller" logger="replicaset-controller" name="replicaset"
I1002 07:01:08.010769 1 shared_informer.go:349] "Waiting for caches to sync" controller="ReplicaSet"
I1002 07:01:08.156020 1 controllermanager.go:781] "Started controller" controller="clusterrole-aggregation-controller"
I1002 07:01:08.160360 1 clusterroleaggregation_controller.go:194] "Starting ClusterRoleAggregator controller" logger="clusterrole-aggregation-controller"
I1002 07:01:08.160384 1 shared_informer.go:349] "Waiting for caches to sync" controller="ClusterRoleAggregator"
I1002 07:01:08.172969 1 shared_informer.go:349] "Waiting for caches to sync" controller="resource quota"
==> kube-scheduler [139929cebf07] <==
I1002 07:01:02.136044 1 server.go:177] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I1002 07:01:02.138391 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
I1002 07:01:02.138431 1 shared_informer.go:349] "Waiting for caches to sync" controller="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
I1002 07:01:02.139278 1 secure_serving.go:211] Serving securely on 127.0.0.1:10259
I1002 07:01:02.139543 1 tlsconfig.go:243] "Starting DynamicServingCertificateController"
E1002 07:01:02.145643 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PodDisruptionBudget"
E1002 07:01:02.147822 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError" reflector="runtime/asm_arm64.s:1223" type="*v1.ConfigMap"
E1002 07:01:02.148936 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service"
E1002 07:01:02.149007 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicationController"
E1002 07:01:02.152132 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicaSet"
E1002 07:01:02.152289 1 reflector.go:205] "Failed to watch" err="failed to list *v1.VolumeAttachment: volumeattachments.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"volumeattachments\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.VolumeAttachment"
E1002 07:01:02.152430 1 reflector.go:205] "Failed to watch" err="failed to list *v1.DeviceClass: deviceclasses.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"deviceclasses\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.DeviceClass"
E1002 07:01:02.152626 1 reflector.go:205] "Failed to watch" err="failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.StorageClass"
E1002 07:01:02.152768 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolume"
E1002 07:01:02.152881 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver"
E1002 07:01:02.156046 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node"
E1002 07:01:02.156140 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ResourceClaim: resourceclaims.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"resourceclaims\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ResourceClaim"
E1002 07:01:02.156377 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolumeClaim"
E1002 07:01:02.156450 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ResourceSlice: resourceslices.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"resourceslices\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ResourceSlice"
E1002 07:01:02.156537 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSINode"
E1002 07:01:02.158536 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace"
E1002 07:01:02.158744 1 reflector.go:205] "Failed to watch" err="failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.StatefulSet"
E1002 07:01:02.158946 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod"
E1002 07:01:02.159070 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIStorageCapacity"
I1002 07:01:03.139273 1 shared_informer.go:356] "Caches are synced" controller="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
==> kubelet <==
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589490 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/9a4224e9bf45d23c8d42b6ed054a5be9-kubeconfig\") pod \"kube-scheduler-scheduled-stop-273808\" (UID: \"9a4224e9bf45d23c8d42b6ed054a5be9\") " pod="kube-system/kube-scheduler-scheduled-stop-273808"
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589523 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/00eaa4cd57314d0f43b27da561d2306e-k8s-certs\") pod \"kube-controller-manager-scheduled-stop-273808\" (UID: \"00eaa4cd57314d0f43b27da561d2306e\") " pod="kube-system/kube-controller-manager-scheduled-stop-273808"
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589544 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/00eaa4cd57314d0f43b27da561d2306e-usr-local-share-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-273808\" (UID: \"00eaa4cd57314d0f43b27da561d2306e\") " pod="kube-system/kube-controller-manager-scheduled-stop-273808"
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589564 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-certs\" (UniqueName: \"kubernetes.io/host-path/8c6bc071c734c037404cae344cf27e44-etcd-certs\") pod \"etcd-scheduled-stop-273808\" (UID: \"8c6bc071c734c037404cae344cf27e44\") " pod="kube-system/etcd-scheduled-stop-273808"
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589582 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-data\" (UniqueName: \"kubernetes.io/host-path/8c6bc071c734c037404cae344cf27e44-etcd-data\") pod \"etcd-scheduled-stop-273808\" (UID: \"8c6bc071c734c037404cae344cf27e44\") " pod="kube-system/etcd-scheduled-stop-273808"
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589598 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8bd6dcd5dbca2297dd68dc00514bb6f0-usr-share-ca-certificates\") pod \"kube-apiserver-scheduled-stop-273808\" (UID: \"8bd6dcd5dbca2297dd68dc00514bb6f0\") " pod="kube-system/kube-apiserver-scheduled-stop-273808"
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589613 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/00eaa4cd57314d0f43b27da561d2306e-ca-certs\") pod \"kube-controller-manager-scheduled-stop-273808\" (UID: \"00eaa4cd57314d0f43b27da561d2306e\") " pod="kube-system/kube-controller-manager-scheduled-stop-273808"
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589630 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/00eaa4cd57314d0f43b27da561d2306e-etc-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-273808\" (UID: \"00eaa4cd57314d0f43b27da561d2306e\") " pod="kube-system/kube-controller-manager-scheduled-stop-273808"
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589659 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8bd6dcd5dbca2297dd68dc00514bb6f0-etc-ca-certificates\") pod \"kube-apiserver-scheduled-stop-273808\" (UID: \"8bd6dcd5dbca2297dd68dc00514bb6f0\") " pod="kube-system/kube-apiserver-scheduled-stop-273808"
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589684 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8bd6dcd5dbca2297dd68dc00514bb6f0-usr-local-share-ca-certificates\") pod \"kube-apiserver-scheduled-stop-273808\" (UID: \"8bd6dcd5dbca2297dd68dc00514bb6f0\") " pod="kube-system/kube-apiserver-scheduled-stop-273808"
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589701 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/00eaa4cd57314d0f43b27da561d2306e-flexvolume-dir\") pod \"kube-controller-manager-scheduled-stop-273808\" (UID: \"00eaa4cd57314d0f43b27da561d2306e\") " pod="kube-system/kube-controller-manager-scheduled-stop-273808"
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589717 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/00eaa4cd57314d0f43b27da561d2306e-kubeconfig\") pod \"kube-controller-manager-scheduled-stop-273808\" (UID: \"00eaa4cd57314d0f43b27da561d2306e\") " pod="kube-system/kube-controller-manager-scheduled-stop-273808"
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589746 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8bd6dcd5dbca2297dd68dc00514bb6f0-ca-certs\") pod \"kube-apiserver-scheduled-stop-273808\" (UID: \"8bd6dcd5dbca2297dd68dc00514bb6f0\") " pod="kube-system/kube-apiserver-scheduled-stop-273808"
Oct 02 07:01:04 scheduled-stop-273808 kubelet[2299]: I1002 07:01:04.589761 2299 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8bd6dcd5dbca2297dd68dc00514bb6f0-k8s-certs\") pod \"kube-apiserver-scheduled-stop-273808\" (UID: \"8bd6dcd5dbca2297dd68dc00514bb6f0\") " pod="kube-system/kube-apiserver-scheduled-stop-273808"
Oct 02 07:01:05 scheduled-stop-273808 kubelet[2299]: I1002 07:01:05.138774 2299 apiserver.go:52] "Watching apiserver"
Oct 02 07:01:05 scheduled-stop-273808 kubelet[2299]: I1002 07:01:05.184510 2299 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world"
Oct 02 07:01:05 scheduled-stop-273808 kubelet[2299]: I1002 07:01:05.311663 2299 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-scheduled-stop-273808"
Oct 02 07:01:05 scheduled-stop-273808 kubelet[2299]: I1002 07:01:05.311909 2299 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/etcd-scheduled-stop-273808"
Oct 02 07:01:05 scheduled-stop-273808 kubelet[2299]: E1002 07:01:05.344957 2299 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-apiserver-scheduled-stop-273808\" already exists" pod="kube-system/kube-apiserver-scheduled-stop-273808"
Oct 02 07:01:05 scheduled-stop-273808 kubelet[2299]: E1002 07:01:05.345607 2299 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"etcd-scheduled-stop-273808\" already exists" pod="kube-system/etcd-scheduled-stop-273808"
Oct 02 07:01:05 scheduled-stop-273808 kubelet[2299]: I1002 07:01:05.375501 2299 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-scheduled-stop-273808" podStartSLOduration=3.375482902 podStartE2EDuration="3.375482902s" podCreationTimestamp="2025-10-02 07:01:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:01:05.355013058 +0000 UTC m=+1.294978952" watchObservedRunningTime="2025-10-02 07:01:05.375482902 +0000 UTC m=+1.315448812"
Oct 02 07:01:05 scheduled-stop-273808 kubelet[2299]: I1002 07:01:05.397158 2299 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-scheduled-stop-273808" podStartSLOduration=1.397129467 podStartE2EDuration="1.397129467s" podCreationTimestamp="2025-10-02 07:01:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:01:05.376735339 +0000 UTC m=+1.316701241" watchObservedRunningTime="2025-10-02 07:01:05.397129467 +0000 UTC m=+1.337095369"
Oct 02 07:01:05 scheduled-stop-273808 kubelet[2299]: I1002 07:01:05.418140 2299 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/etcd-scheduled-stop-273808" podStartSLOduration=1.418120826 podStartE2EDuration="1.418120826s" podCreationTimestamp="2025-10-02 07:01:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:01:05.397649448 +0000 UTC m=+1.337615342" watchObservedRunningTime="2025-10-02 07:01:05.418120826 +0000 UTC m=+1.358086720"
Oct 02 07:01:05 scheduled-stop-273808 kubelet[2299]: I1002 07:01:05.418238 2299 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-scheduled-stop-273808" podStartSLOduration=1.418231905 podStartE2EDuration="1.418231905s" podCreationTimestamp="2025-10-02 07:01:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:01:05.414630273 +0000 UTC m=+1.354596183" watchObservedRunningTime="2025-10-02 07:01:05.418231905 +0000 UTC m=+1.358197815"
Oct 02 07:01:07 scheduled-stop-273808 kubelet[2299]: I1002 07:01:07.482819 2299 kubelet_node_status.go:439] "Fast updating node status as it just became ready"
-- /stdout --
helpers_test.go:262: (dbg) Run: out/minikube-linux-arm64 status --format={{.APIServer}} -p scheduled-stop-273808 -n scheduled-stop-273808
helpers_test.go:269: (dbg) Run: kubectl --context scheduled-stop-273808 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:280: non-running pods: kube-proxy-hww7j storage-provisioner
helpers_test.go:282: ======> post-mortem[TestScheduledStopUnix]: describe non-running pods <======
helpers_test.go:285: (dbg) Run: kubectl --context scheduled-stop-273808 describe pod kube-proxy-hww7j storage-provisioner
helpers_test.go:285: (dbg) Non-zero exit: kubectl --context scheduled-stop-273808 describe pod kube-proxy-hww7j storage-provisioner: exit status 1 (99.92557ms)
** stderr **
Error from server (NotFound): pods "kube-proxy-hww7j" not found
Error from server (NotFound): pods "storage-provisioner" not found
** /stderr **
helpers_test.go:287: kubectl --context scheduled-stop-273808 describe pod kube-proxy-hww7j storage-provisioner: exit status 1
helpers_test.go:175: Cleaning up "scheduled-stop-273808" profile ...
helpers_test.go:178: (dbg) Run: out/minikube-linux-arm64 delete -p scheduled-stop-273808
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p scheduled-stop-273808: (2.192486165s)
--- FAIL: TestScheduledStopUnix (42.09s)