=== RUN TestScheduledStopUnix
scheduled_stop_test.go:128: (dbg) Run: out/minikube-linux-arm64 start -p scheduled-stop-160137 --memory=3072 --driver=docker --container-runtime=containerd
scheduled_stop_test.go:128: (dbg) Done: out/minikube-linux-arm64 start -p scheduled-stop-160137 --memory=3072 --driver=docker --container-runtime=containerd: (28.688364423s)
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-160137 --schedule 5m
scheduled_stop_test.go:191: (dbg) Run: out/minikube-linux-arm64 status --format={{.TimeToStop}} -p scheduled-stop-160137 -n scheduled-stop-160137
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-160137 --schedule 15s
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:98: process 155217 running but should have been killed on reschedule of stop
panic.go:636: *** TestScheduledStopUnix FAILED at 2025-09-08 14:02:21.391045925 +0000 UTC m=+2110.549390162
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:223: ======> post-mortem[TestScheduledStopUnix]: network settings <======
helpers_test.go:230: HOST ENV snapshots: PROXY env: HTTP_PROXY="<empty>" HTTPS_PROXY="<empty>" NO_PROXY="<empty>"
helpers_test.go:238: ======> post-mortem[TestScheduledStopUnix]: docker inspect <======
helpers_test.go:239: (dbg) Run: docker inspect scheduled-stop-160137
helpers_test.go:243: (dbg) docker inspect scheduled-stop-160137:
-- stdout --
[
{
"Id": "fad8a2923aa0ef49116e4863d35177de96cf07ada764280017c35b4ee50735ed",
"Created": "2025-09-08T14:01:57.45059943Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 153205,
"ExitCode": 0,
"Error": "",
"StartedAt": "2025-09-08T14:01:57.515622336Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:1a6e5b410fd9226cf2434621073598c7c01bccc994a53260ab0a0d834a0f1815",
"ResolvConfPath": "/var/lib/docker/containers/fad8a2923aa0ef49116e4863d35177de96cf07ada764280017c35b4ee50735ed/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/fad8a2923aa0ef49116e4863d35177de96cf07ada764280017c35b4ee50735ed/hostname",
"HostsPath": "/var/lib/docker/containers/fad8a2923aa0ef49116e4863d35177de96cf07ada764280017c35b4ee50735ed/hosts",
"LogPath": "/var/lib/docker/containers/fad8a2923aa0ef49116e4863d35177de96cf07ada764280017c35b4ee50735ed/fad8a2923aa0ef49116e4863d35177de96cf07ada764280017c35b4ee50735ed-json.log",
"Name": "/scheduled-stop-160137",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"scheduled-stop-160137:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {}
},
"NetworkMode": "scheduled-stop-160137",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 3221225472,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 6442450944,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"ID": "fad8a2923aa0ef49116e4863d35177de96cf07ada764280017c35b4ee50735ed",
"LowerDir": "/var/lib/docker/overlay2/51b0c657769dfb7da9eaf7681fc76880e551f44769ec054772818620c83f843d-init/diff:/var/lib/docker/overlay2/81b144fe83a3a806b065a20c9a28409512052a83c9af991906fac9b66cb41fc1/diff",
"MergedDir": "/var/lib/docker/overlay2/51b0c657769dfb7da9eaf7681fc76880e551f44769ec054772818620c83f843d/merged",
"UpperDir": "/var/lib/docker/overlay2/51b0c657769dfb7da9eaf7681fc76880e551f44769ec054772818620c83f843d/diff",
"WorkDir": "/var/lib/docker/overlay2/51b0c657769dfb7da9eaf7681fc76880e551f44769ec054772818620c83f843d/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "scheduled-stop-160137",
"Source": "/var/lib/docker/volumes/scheduled-stop-160137/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "scheduled-stop-160137",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "scheduled-stop-160137",
"name.minikube.sigs.k8s.io": "scheduled-stop-160137",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "4aafbcbe00806389039c829e296092fef372b111d4f2a60400275c3581822a5b",
"SandboxKey": "/var/run/docker/netns/4aafbcbe0080",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32968"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32969"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32972"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32970"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32971"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"scheduled-stop-160137": {
"IPAMConfig": {
"IPv4Address": "192.168.76.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "3a:ac:3e:36:bd:5a",
"DriverOpts": null,
"GwPriority": 0,
"NetworkID": "4ffe997009a8131c2e972d7bed8c324037b15ade543dadfcd261ae01c2938720",
"EndpointID": "2514fe1f7662623df3ef72e79d5bdd86bb4adb9fadc9129a07d4f3560e408485",
"Gateway": "192.168.76.1",
"IPAddress": "192.168.76.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"scheduled-stop-160137",
"fad8a2923aa0"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:247: (dbg) Run: out/minikube-linux-arm64 status --format={{.Host}} -p scheduled-stop-160137 -n scheduled-stop-160137
helpers_test.go:252: <<< TestScheduledStopUnix FAILED: start of post-mortem logs <<<
helpers_test.go:253: ======> post-mortem[TestScheduledStopUnix]: minikube logs <======
helpers_test.go:255: (dbg) Run: out/minikube-linux-arm64 -p scheduled-stop-160137 logs -n 25
helpers_test.go:255: (dbg) Done: out/minikube-linux-arm64 -p scheduled-stop-160137 logs -n 25: (1.259179582s)
helpers_test.go:260: TestScheduledStopUnix logs:
-- stdout --
==> Audit <==
┌─────────┬─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┬───────────────────────┬─────────┬─────────┬─────────────────────┬─────────────────────┐
│ COMMAND │ ARGS │ PROFILE │ USER │ VERSION │ START TIME │ END TIME │
├─────────┼─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┼───────────────────────┼─────────┼─────────┼─────────────────────┼─────────────────────┤
│ stop │ -p multinode-789083 │ multinode-789083 │ jenkins │ v1.36.0 │ 08 Sep 25 13:56 UTC │ 08 Sep 25 13:56 UTC │
│ start │ -p multinode-789083 --wait=true -v=5 --alsologtostderr │ multinode-789083 │ jenkins │ v1.36.0 │ 08 Sep 25 13:56 UTC │ 08 Sep 25 13:57 UTC │
│ node │ list -p multinode-789083 │ multinode-789083 │ jenkins │ v1.36.0 │ 08 Sep 25 13:57 UTC │ │
│ node │ multinode-789083 node delete m03 │ multinode-789083 │ jenkins │ v1.36.0 │ 08 Sep 25 13:57 UTC │ 08 Sep 25 13:57 UTC │
│ stop │ multinode-789083 stop │ multinode-789083 │ jenkins │ v1.36.0 │ 08 Sep 25 13:57 UTC │ 08 Sep 25 13:57 UTC │
│ start │ -p multinode-789083 --wait=true -v=5 --alsologtostderr --driver=docker --container-runtime=containerd │ multinode-789083 │ jenkins │ v1.36.0 │ 08 Sep 25 13:57 UTC │ 08 Sep 25 13:58 UTC │
│ node │ list -p multinode-789083 │ multinode-789083 │ jenkins │ v1.36.0 │ 08 Sep 25 13:58 UTC │ │
│ start │ -p multinode-789083-m02 --driver=docker --container-runtime=containerd │ multinode-789083-m02 │ jenkins │ v1.36.0 │ 08 Sep 25 13:58 UTC │ │
│ start │ -p multinode-789083-m03 --driver=docker --container-runtime=containerd │ multinode-789083-m03 │ jenkins │ v1.36.0 │ 08 Sep 25 13:58 UTC │ 08 Sep 25 13:59 UTC │
│ node │ add -p multinode-789083 │ multinode-789083 │ jenkins │ v1.36.0 │ 08 Sep 25 13:59 UTC │ │
│ delete │ -p multinode-789083-m03 │ multinode-789083-m03 │ jenkins │ v1.36.0 │ 08 Sep 25 13:59 UTC │ 08 Sep 25 13:59 UTC │
│ delete │ -p multinode-789083 │ multinode-789083 │ jenkins │ v1.36.0 │ 08 Sep 25 13:59 UTC │ 08 Sep 25 13:59 UTC │
│ start │ -p test-preload-053548 --memory=3072 --alsologtostderr --wait=true --preload=false --driver=docker --container-runtime=containerd --kubernetes-version=v1.32.0 │ test-preload-053548 │ jenkins │ v1.36.0 │ 08 Sep 25 13:59 UTC │ 08 Sep 25 14:00 UTC │
│ image │ test-preload-053548 image pull gcr.io/k8s-minikube/busybox │ test-preload-053548 │ jenkins │ v1.36.0 │ 08 Sep 25 14:00 UTC │ 08 Sep 25 14:00 UTC │
│ stop │ -p test-preload-053548 │ test-preload-053548 │ jenkins │ v1.36.0 │ 08 Sep 25 14:00 UTC │ 08 Sep 25 14:00 UTC │
│ start │ -p test-preload-053548 --memory=3072 --alsologtostderr -v=1 --wait=true --driver=docker --container-runtime=containerd │ test-preload-053548 │ jenkins │ v1.36.0 │ 08 Sep 25 14:00 UTC │ 08 Sep 25 14:01 UTC │
│ image │ test-preload-053548 image list │ test-preload-053548 │ jenkins │ v1.36.0 │ 08 Sep 25 14:01 UTC │ 08 Sep 25 14:01 UTC │
│ delete │ -p test-preload-053548 │ test-preload-053548 │ jenkins │ v1.36.0 │ 08 Sep 25 14:01 UTC │ 08 Sep 25 14:01 UTC │
│ start │ -p scheduled-stop-160137 --memory=3072 --driver=docker --container-runtime=containerd │ scheduled-stop-160137 │ jenkins │ v1.36.0 │ 08 Sep 25 14:01 UTC │ 08 Sep 25 14:02 UTC │
│ stop │ -p scheduled-stop-160137 --schedule 5m │ scheduled-stop-160137 │ jenkins │ v1.36.0 │ 08 Sep 25 14:02 UTC │ │
│ stop │ -p scheduled-stop-160137 --schedule 5m │ scheduled-stop-160137 │ jenkins │ v1.36.0 │ 08 Sep 25 14:02 UTC │ │
│ stop │ -p scheduled-stop-160137 --schedule 5m │ scheduled-stop-160137 │ jenkins │ v1.36.0 │ 08 Sep 25 14:02 UTC │ │
│ stop │ -p scheduled-stop-160137 --schedule 15s │ scheduled-stop-160137 │ jenkins │ v1.36.0 │ 08 Sep 25 14:02 UTC │ │
│ stop │ -p scheduled-stop-160137 --schedule 15s │ scheduled-stop-160137 │ jenkins │ v1.36.0 │ 08 Sep 25 14:02 UTC │ │
│ stop │ -p scheduled-stop-160137 --schedule 15s │ scheduled-stop-160137 │ jenkins │ v1.36.0 │ 08 Sep 25 14:02 UTC │ │
└─────────┴─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┴───────────────────────┴─────────┴─────────┴─────────────────────┴─────────────────────┘
==> Last Start <==
Log file created at: 2025/09/08 14:01:52
Running on machine: ip-172-31-30-239
Binary: Built with gc go1.24.6 for linux/arm64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0908 14:01:52.231146 152809 out.go:360] Setting OutFile to fd 1 ...
I0908 14:01:52.231250 152809 out.go:408] TERM=,COLORTERM=, which probably does not support color
I0908 14:01:52.231254 152809 out.go:374] Setting ErrFile to fd 2...
I0908 14:01:52.231257 152809 out.go:408] TERM=,COLORTERM=, which probably does not support color
I0908 14:01:52.231509 152809 root.go:338] Updating PATH: /home/jenkins/minikube-integration/21504-2314/.minikube/bin
I0908 14:01:52.231895 152809 out.go:368] Setting JSON to false
I0908 14:01:52.232712 152809 start.go:130] hostinfo: {"hostname":"ip-172-31-30-239","uptime":2663,"bootTime":1757337450,"procs":153,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1084-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"92f46a7d-c249-4c12-924a-77f64874c910"}
I0908 14:01:52.232786 152809 start.go:140] virtualization:
I0908 14:01:52.236576 152809 out.go:179] * [scheduled-stop-160137] minikube v1.36.0 on Ubuntu 20.04 (arm64)
I0908 14:01:52.241036 152809 out.go:179] - MINIKUBE_LOCATION=21504
I0908 14:01:52.241121 152809 notify.go:220] Checking for updates...
I0908 14:01:52.247632 152809 out.go:179] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0908 14:01:52.250841 152809 out.go:179] - KUBECONFIG=/home/jenkins/minikube-integration/21504-2314/kubeconfig
I0908 14:01:52.254041 152809 out.go:179] - MINIKUBE_HOME=/home/jenkins/minikube-integration/21504-2314/.minikube
I0908 14:01:52.257294 152809 out.go:179] - MINIKUBE_BIN=out/minikube-linux-arm64
I0908 14:01:52.260470 152809 out.go:179] - MINIKUBE_FORCE_SYSTEMD=
I0908 14:01:52.263622 152809 driver.go:421] Setting default libvirt URI to qemu:///system
I0908 14:01:52.296748 152809 docker.go:123] docker version: linux-28.1.1:Docker Engine - Community
I0908 14:01:52.296857 152809 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0908 14:01:52.351365 152809 info.go:266] docker info: {ID:6ZPO:QZND:VNGE:LUKL:4Y3K:XELL:AAX4:2GTK:E6LM:MPRN:3ZXR:TTMR Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:24 OomKillDisable:true NGoroutines:42 SystemTime:2025-09-08 14:01:52.341399291 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1084-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214831104 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-30-239 Labels:[] ExperimentalBuild:false ServerVersion:28.1.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:05044ec0a9a75232cad458027ca83437aae3f4da Expected:} RuncCommit:{ID:v1.2.5-0-g59923ef Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx P
ath:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.23.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.35.1]] Warnings:<nil>}}
I0908 14:01:52.351467 152809 docker.go:318] overlay module found
I0908 14:01:52.354699 152809 out.go:179] * Using the docker driver based on user configuration
I0908 14:01:52.357566 152809 start.go:304] selected driver: docker
I0908 14:01:52.357582 152809 start.go:918] validating driver "docker" against <nil>
I0908 14:01:52.357594 152809 start.go:929] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0908 14:01:52.358358 152809 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0908 14:01:52.413248 152809 info.go:266] docker info: {ID:6ZPO:QZND:VNGE:LUKL:4Y3K:XELL:AAX4:2GTK:E6LM:MPRN:3ZXR:TTMR Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:24 OomKillDisable:true NGoroutines:42 SystemTime:2025-09-08 14:01:52.403566955 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1084-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214831104 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-30-239 Labels:[] ExperimentalBuild:false ServerVersion:28.1.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:05044ec0a9a75232cad458027ca83437aae3f4da Expected:} RuncCommit:{ID:v1.2.5-0-g59923ef Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx P
ath:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.23.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.35.1]] Warnings:<nil>}}
I0908 14:01:52.413395 152809 start_flags.go:327] no existing cluster config was found, will generate one from the flags
I0908 14:01:52.413606 152809 start_flags.go:974] Wait components to verify : map[apiserver:true system_pods:true]
I0908 14:01:52.416768 152809 out.go:179] * Using Docker driver with root privileges
I0908 14:01:52.419819 152809 cni.go:84] Creating CNI manager for ""
I0908 14:01:52.419889 152809 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I0908 14:01:52.419896 152809 start_flags.go:336] Found "CNI" CNI - setting NetworkPlugin=cni
I0908 14:01:52.419984 152809 start.go:348] cluster config:
{Name:scheduled-stop-160137 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.0 ClusterName:scheduled-stop-160137 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local
ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.34.0 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0908 14:01:52.423180 152809 out.go:179] * Starting "scheduled-stop-160137" primary control-plane node in "scheduled-stop-160137" cluster
I0908 14:01:52.426087 152809 cache.go:123] Beginning downloading kic base image for docker with containerd
I0908 14:01:52.429026 152809 out.go:179] * Pulling base image v0.0.47-1756980985-21488 ...
I0908 14:01:52.431960 152809 preload.go:131] Checking if preload exists for k8s version v1.34.0 and runtime containerd
I0908 14:01:52.431997 152809 image.go:81] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 in local docker daemon
I0908 14:01:52.432024 152809 preload.go:146] Found local preload: /home/jenkins/minikube-integration/21504-2314/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.0-containerd-overlay2-arm64.tar.lz4
I0908 14:01:52.432032 152809 cache.go:58] Caching tarball of preloaded images
I0908 14:01:52.432113 152809 preload.go:172] Found /home/jenkins/minikube-integration/21504-2314/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.0-containerd-overlay2-arm64.tar.lz4 in cache, skipping download
I0908 14:01:52.432122 152809 cache.go:61] Finished verifying existence of preloaded tar for v1.34.0 on containerd
I0908 14:01:52.432516 152809 profile.go:143] Saving config to /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/config.json ...
I0908 14:01:52.432539 152809 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/config.json: {Name:mkf37b0caa8bf2a0d7488d00ebef3e592b60badd Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:01:52.451840 152809 image.go:100] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 in local docker daemon, skipping pull
I0908 14:01:52.451853 152809 cache.go:147] gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 exists in daemon, skipping load
I0908 14:01:52.451865 152809 cache.go:232] Successfully downloaded all kic artifacts
I0908 14:01:52.451895 152809 start.go:360] acquireMachinesLock for scheduled-stop-160137: {Name:mk2fd4b010769ec91004603a4e0ce667bfec7e77 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0908 14:01:52.452005 152809 start.go:364] duration metric: took 94.893µs to acquireMachinesLock for "scheduled-stop-160137"
I0908 14:01:52.452029 152809 start.go:93] Provisioning new machine with config: &{Name:scheduled-stop-160137 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.0 ClusterName:scheduled-stop-160137 Namespace:default APIServerHAVIP: A
PIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.34.0 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: Static
IP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.34.0 ContainerRuntime:containerd ControlPlane:true Worker:true}
I0908 14:01:52.452091 152809 start.go:125] createHost starting for "" (driver="docker")
I0908 14:01:52.455592 152809 out.go:252] * Creating docker container (CPUs=2, Memory=3072MB) ...
I0908 14:01:52.455883 152809 start.go:159] libmachine.API.Create for "scheduled-stop-160137" (driver="docker")
I0908 14:01:52.455923 152809 client.go:168] LocalClient.Create starting
I0908 14:01:52.456002 152809 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21504-2314/.minikube/certs/ca.pem
I0908 14:01:52.456040 152809 main.go:141] libmachine: Decoding PEM data...
I0908 14:01:52.456052 152809 main.go:141] libmachine: Parsing certificate...
I0908 14:01:52.456118 152809 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21504-2314/.minikube/certs/cert.pem
I0908 14:01:52.456137 152809 main.go:141] libmachine: Decoding PEM data...
I0908 14:01:52.456145 152809 main.go:141] libmachine: Parsing certificate...
I0908 14:01:52.456592 152809 cli_runner.go:164] Run: docker network inspect scheduled-stop-160137 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0908 14:01:52.473090 152809 cli_runner.go:211] docker network inspect scheduled-stop-160137 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0908 14:01:52.473161 152809 network_create.go:284] running [docker network inspect scheduled-stop-160137] to gather additional debugging logs...
I0908 14:01:52.473178 152809 cli_runner.go:164] Run: docker network inspect scheduled-stop-160137
W0908 14:01:52.492160 152809 cli_runner.go:211] docker network inspect scheduled-stop-160137 returned with exit code 1
I0908 14:01:52.492179 152809 network_create.go:287] error running [docker network inspect scheduled-stop-160137]: docker network inspect scheduled-stop-160137: exit status 1
stdout:
[]
stderr:
Error response from daemon: network scheduled-stop-160137 not found
I0908 14:01:52.492191 152809 network_create.go:289] output of [docker network inspect scheduled-stop-160137]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network scheduled-stop-160137 not found
** /stderr **
I0908 14:01:52.492304 152809 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0908 14:01:52.509292 152809 network.go:211] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-431c1a61966e IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:46:58:d5:96:47:2e} reservation:<nil>}
I0908 14:01:52.509536 152809 network.go:211] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-2cac6205be69 IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:d6:0a:55:ef:54:5c} reservation:<nil>}
I0908 14:01:52.509770 152809 network.go:211] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-e7f77c37dc8f IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:ea:d9:85:22:06:db} reservation:<nil>}
I0908 14:01:52.510095 152809 network.go:206] using free private subnet 192.168.76.0/24: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x400193bcf0}
I0908 14:01:52.510127 152809 network_create.go:124] attempt to create docker network scheduled-stop-160137 192.168.76.0/24 with gateway 192.168.76.1 and MTU of 1500 ...
I0908 14:01:52.510185 152809 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.76.0/24 --gateway=192.168.76.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=scheduled-stop-160137 scheduled-stop-160137
I0908 14:01:52.568329 152809 network_create.go:108] docker network scheduled-stop-160137 192.168.76.0/24 created
I0908 14:01:52.568528 152809 kic.go:121] calculated static IP "192.168.76.2" for the "scheduled-stop-160137" container
I0908 14:01:52.568628 152809 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0908 14:01:52.583778 152809 cli_runner.go:164] Run: docker volume create scheduled-stop-160137 --label name.minikube.sigs.k8s.io=scheduled-stop-160137 --label created_by.minikube.sigs.k8s.io=true
I0908 14:01:52.603002 152809 oci.go:103] Successfully created a docker volume scheduled-stop-160137
I0908 14:01:52.603085 152809 cli_runner.go:164] Run: docker run --rm --name scheduled-stop-160137-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-160137 --entrypoint /usr/bin/test -v scheduled-stop-160137:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 -d /var/lib
I0908 14:01:53.080698 152809 oci.go:107] Successfully prepared a docker volume scheduled-stop-160137
I0908 14:01:53.080750 152809 preload.go:131] Checking if preload exists for k8s version v1.34.0 and runtime containerd
I0908 14:01:53.080769 152809 kic.go:194] Starting extracting preloaded images to volume ...
I0908 14:01:53.080844 152809 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/21504-2314/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.0-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-160137:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 -I lz4 -xf /preloaded.tar -C /extractDir
I0908 14:01:57.372340 152809 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/21504-2314/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.0-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-160137:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 -I lz4 -xf /preloaded.tar -C /extractDir: (4.291455969s)
I0908 14:01:57.372373 152809 kic.go:203] duration metric: took 4.291601029s to extract preloaded images to volume ...
W0908 14:01:57.372509 152809 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0908 14:01:57.372607 152809 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0908 14:01:57.432944 152809 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname scheduled-stop-160137 --name scheduled-stop-160137 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-160137 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=scheduled-stop-160137 --network scheduled-stop-160137 --ip 192.168.76.2 --volume scheduled-stop-160137:/var --security-opt apparmor=unconfined --memory=3072mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79
I0908 14:01:57.763178 152809 cli_runner.go:164] Run: docker container inspect scheduled-stop-160137 --format={{.State.Running}}
I0908 14:01:57.784330 152809 cli_runner.go:164] Run: docker container inspect scheduled-stop-160137 --format={{.State.Status}}
I0908 14:01:57.814412 152809 cli_runner.go:164] Run: docker exec scheduled-stop-160137 stat /var/lib/dpkg/alternatives/iptables
I0908 14:01:57.871946 152809 oci.go:144] the created container "scheduled-stop-160137" has a running status.
I0908 14:01:57.871977 152809 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/21504-2314/.minikube/machines/scheduled-stop-160137/id_rsa...
I0908 14:01:58.211954 152809 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/21504-2314/.minikube/machines/scheduled-stop-160137/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0908 14:01:58.238111 152809 cli_runner.go:164] Run: docker container inspect scheduled-stop-160137 --format={{.State.Status}}
I0908 14:01:58.265116 152809 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0908 14:01:58.265127 152809 kic_runner.go:114] Args: [docker exec --privileged scheduled-stop-160137 chown docker:docker /home/docker/.ssh/authorized_keys]
I0908 14:01:58.325851 152809 cli_runner.go:164] Run: docker container inspect scheduled-stop-160137 --format={{.State.Status}}
I0908 14:01:58.353799 152809 machine.go:93] provisionDockerMachine start ...
I0908 14:01:58.353881 152809 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-160137
I0908 14:01:58.384629 152809 main.go:141] libmachine: Using SSH client type: native
I0908 14:01:58.384963 152809 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef840] 0x3f2000 <nil> [] 0s} 127.0.0.1 32968 <nil> <nil>}
I0908 14:01:58.384971 152809 main.go:141] libmachine: About to run SSH command:
hostname
I0908 14:01:58.562625 152809 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-160137
I0908 14:01:58.562643 152809 ubuntu.go:182] provisioning hostname "scheduled-stop-160137"
I0908 14:01:58.562722 152809 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-160137
I0908 14:01:58.593437 152809 main.go:141] libmachine: Using SSH client type: native
I0908 14:01:58.593803 152809 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef840] 0x3f2000 <nil> [] 0s} 127.0.0.1 32968 <nil> <nil>}
I0908 14:01:58.593814 152809 main.go:141] libmachine: About to run SSH command:
sudo hostname scheduled-stop-160137 && echo "scheduled-stop-160137" | sudo tee /etc/hostname
I0908 14:01:58.769812 152809 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-160137
I0908 14:01:58.769892 152809 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-160137
I0908 14:01:58.792207 152809 main.go:141] libmachine: Using SSH client type: native
I0908 14:01:58.792537 152809 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3ef840] 0x3f2000 <nil> [] 0s} 127.0.0.1 32968 <nil> <nil>}
I0908 14:01:58.792553 152809 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sscheduled-stop-160137' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 scheduled-stop-160137/g' /etc/hosts;
else
echo '127.0.1.1 scheduled-stop-160137' | sudo tee -a /etc/hosts;
fi
fi
I0908 14:01:58.924508 152809 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0908 14:01:58.924522 152809 ubuntu.go:188] set auth options {CertDir:/home/jenkins/minikube-integration/21504-2314/.minikube CaCertPath:/home/jenkins/minikube-integration/21504-2314/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/21504-2314/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/21504-2314/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/21504-2314/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/21504-2314/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/21504-2314/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/21504-2314/.minikube}
I0908 14:01:58.924551 152809 ubuntu.go:190] setting up certificates
I0908 14:01:58.924561 152809 provision.go:84] configureAuth start
I0908 14:01:58.924630 152809 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-160137
I0908 14:01:58.941779 152809 provision.go:143] copyHostCerts
I0908 14:01:58.941828 152809 exec_runner.go:144] found /home/jenkins/minikube-integration/21504-2314/.minikube/ca.pem, removing ...
I0908 14:01:58.941836 152809 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21504-2314/.minikube/ca.pem
I0908 14:01:58.941893 152809 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21504-2314/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/21504-2314/.minikube/ca.pem (1078 bytes)
I0908 14:01:58.941968 152809 exec_runner.go:144] found /home/jenkins/minikube-integration/21504-2314/.minikube/cert.pem, removing ...
I0908 14:01:58.941972 152809 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21504-2314/.minikube/cert.pem
I0908 14:01:58.941996 152809 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21504-2314/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/21504-2314/.minikube/cert.pem (1123 bytes)
I0908 14:01:58.942045 152809 exec_runner.go:144] found /home/jenkins/minikube-integration/21504-2314/.minikube/key.pem, removing ...
I0908 14:01:58.942048 152809 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21504-2314/.minikube/key.pem
I0908 14:01:58.942072 152809 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21504-2314/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/21504-2314/.minikube/key.pem (1679 bytes)
I0908 14:01:58.942115 152809 provision.go:117] generating server cert: /home/jenkins/minikube-integration/21504-2314/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/21504-2314/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/21504-2314/.minikube/certs/ca-key.pem org=jenkins.scheduled-stop-160137 san=[127.0.0.1 192.168.76.2 localhost minikube scheduled-stop-160137]
I0908 14:01:59.274235 152809 provision.go:177] copyRemoteCerts
I0908 14:01:59.274295 152809 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0908 14:01:59.274331 152809 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-160137
I0908 14:01:59.293937 152809 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32968 SSHKeyPath:/home/jenkins/minikube-integration/21504-2314/.minikube/machines/scheduled-stop-160137/id_rsa Username:docker}
I0908 14:01:59.385570 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0908 14:01:59.410850 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/machines/server.pem --> /etc/docker/server.pem (1229 bytes)
I0908 14:01:59.435721 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
I0908 14:01:59.461829 152809 provision.go:87] duration metric: took 537.245887ms to configureAuth
I0908 14:01:59.461858 152809 ubuntu.go:206] setting minikube options for container-runtime
I0908 14:01:59.462040 152809 config.go:182] Loaded profile config "scheduled-stop-160137": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.34.0
I0908 14:01:59.462046 152809 machine.go:96] duration metric: took 1.108237196s to provisionDockerMachine
I0908 14:01:59.462052 152809 client.go:171] duration metric: took 7.006123748s to LocalClient.Create
I0908 14:01:59.462082 152809 start.go:167] duration metric: took 7.006192836s to libmachine.API.Create "scheduled-stop-160137"
I0908 14:01:59.462089 152809 start.go:293] postStartSetup for "scheduled-stop-160137" (driver="docker")
I0908 14:01:59.462097 152809 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0908 14:01:59.462149 152809 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0908 14:01:59.462185 152809 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-160137
I0908 14:01:59.479418 152809 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32968 SSHKeyPath:/home/jenkins/minikube-integration/21504-2314/.minikube/machines/scheduled-stop-160137/id_rsa Username:docker}
I0908 14:01:59.569848 152809 ssh_runner.go:195] Run: cat /etc/os-release
I0908 14:01:59.573039 152809 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0908 14:01:59.573062 152809 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0908 14:01:59.573071 152809 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0908 14:01:59.573077 152809 info.go:137] Remote host: Ubuntu 22.04.5 LTS
I0908 14:01:59.573090 152809 filesync.go:126] Scanning /home/jenkins/minikube-integration/21504-2314/.minikube/addons for local assets ...
I0908 14:01:59.573159 152809 filesync.go:126] Scanning /home/jenkins/minikube-integration/21504-2314/.minikube/files for local assets ...
I0908 14:01:59.573238 152809 filesync.go:149] local asset: /home/jenkins/minikube-integration/21504-2314/.minikube/files/etc/ssl/certs/41182.pem -> 41182.pem in /etc/ssl/certs
I0908 14:01:59.573338 152809 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I0908 14:01:59.581926 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/files/etc/ssl/certs/41182.pem --> /etc/ssl/certs/41182.pem (1708 bytes)
I0908 14:01:59.606671 152809 start.go:296] duration metric: took 144.567521ms for postStartSetup
I0908 14:01:59.607071 152809 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-160137
I0908 14:01:59.624820 152809 profile.go:143] Saving config to /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/config.json ...
I0908 14:01:59.625113 152809 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0908 14:01:59.625154 152809 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-160137
I0908 14:01:59.642235 152809 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32968 SSHKeyPath:/home/jenkins/minikube-integration/21504-2314/.minikube/machines/scheduled-stop-160137/id_rsa Username:docker}
I0908 14:01:59.729137 152809 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0908 14:01:59.733593 152809 start.go:128] duration metric: took 7.28148785s to createHost
I0908 14:01:59.733607 152809 start.go:83] releasing machines lock for "scheduled-stop-160137", held for 7.281595289s
I0908 14:01:59.733675 152809 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-160137
I0908 14:01:59.754817 152809 ssh_runner.go:195] Run: cat /version.json
I0908 14:01:59.754860 152809 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-160137
I0908 14:01:59.755104 152809 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0908 14:01:59.755161 152809 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-160137
I0908 14:01:59.774691 152809 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32968 SSHKeyPath:/home/jenkins/minikube-integration/21504-2314/.minikube/machines/scheduled-stop-160137/id_rsa Username:docker}
I0908 14:01:59.776090 152809 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32968 SSHKeyPath:/home/jenkins/minikube-integration/21504-2314/.minikube/machines/scheduled-stop-160137/id_rsa Username:docker}
I0908 14:01:59.991598 152809 ssh_runner.go:195] Run: systemctl --version
I0908 14:01:59.995610 152809 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0908 14:01:59.999686 152809 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0908 14:02:00.088865 152809 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0908 14:02:00.088946 152809 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0908 14:02:00.177861 152809 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0908 14:02:00.177876 152809 start.go:495] detecting cgroup driver to use...
I0908 14:02:00.177914 152809 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0908 14:02:00.177978 152809 ssh_runner.go:195] Run: sudo systemctl stop -f crio
I0908 14:02:00.207455 152809 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0908 14:02:00.228552 152809 docker.go:218] disabling cri-docker service (if available) ...
I0908 14:02:00.228627 152809 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
I0908 14:02:00.248270 152809 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
I0908 14:02:00.275857 152809 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
I0908 14:02:00.392804 152809 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
I0908 14:02:00.491175 152809 docker.go:234] disabling docker service ...
I0908 14:02:00.491237 152809 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
I0908 14:02:00.515385 152809 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
I0908 14:02:00.529411 152809 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
I0908 14:02:00.618870 152809 ssh_runner.go:195] Run: sudo systemctl mask docker.service
I0908 14:02:00.709460 152809 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
I0908 14:02:00.721147 152809 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0908 14:02:00.738459 152809 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10.1"|' /etc/containerd/config.toml"
I0908 14:02:00.748818 152809 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0908 14:02:00.759282 152809 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0908 14:02:00.759344 152809 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0908 14:02:00.770032 152809 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0908 14:02:00.780126 152809 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0908 14:02:00.790252 152809 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0908 14:02:00.800223 152809 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0908 14:02:00.809501 152809 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0908 14:02:00.820116 152809 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0908 14:02:00.830761 152809 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0908 14:02:00.841005 152809 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0908 14:02:00.850071 152809 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0908 14:02:00.859141 152809 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0908 14:02:00.939630 152809 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0908 14:02:01.077077 152809 start.go:542] Will wait 60s for socket path /run/containerd/containerd.sock
I0908 14:02:01.077142 152809 ssh_runner.go:195] Run: stat /run/containerd/containerd.sock
I0908 14:02:01.081876 152809 start.go:563] Will wait 60s for crictl version
I0908 14:02:01.081936 152809 ssh_runner.go:195] Run: which crictl
I0908 14:02:01.085540 152809 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0908 14:02:01.126975 152809 start.go:579] Version: 0.1.0
RuntimeName: containerd
RuntimeVersion: 1.7.27
RuntimeApiVersion: v1
I0908 14:02:01.127035 152809 ssh_runner.go:195] Run: containerd --version
I0908 14:02:01.150026 152809 ssh_runner.go:195] Run: containerd --version
I0908 14:02:01.178469 152809 out.go:179] * Preparing Kubernetes v1.34.0 on containerd 1.7.27 ...
I0908 14:02:01.181728 152809 cli_runner.go:164] Run: docker network inspect scheduled-stop-160137 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0908 14:02:01.199632 152809 ssh_runner.go:195] Run: grep 192.168.76.1 host.minikube.internal$ /etc/hosts
I0908 14:02:01.203491 152809 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.76.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0908 14:02:01.215401 152809 kubeadm.go:875] updating cluster {Name:scheduled-stop-160137 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.0 ClusterName:scheduled-stop-160137 Namespace:default APIServerHAVIP: APIServerName:miniku
beCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.0 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSH
AuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0908 14:02:01.215506 152809 preload.go:131] Checking if preload exists for k8s version v1.34.0 and runtime containerd
I0908 14:02:01.215569 152809 ssh_runner.go:195] Run: sudo crictl images --output json
I0908 14:02:01.252572 152809 containerd.go:627] all images are preloaded for containerd runtime.
I0908 14:02:01.252585 152809 containerd.go:534] Images already preloaded, skipping extraction
I0908 14:02:01.252646 152809 ssh_runner.go:195] Run: sudo crictl images --output json
I0908 14:02:01.293629 152809 containerd.go:627] all images are preloaded for containerd runtime.
I0908 14:02:01.293643 152809 cache_images.go:85] Images are preloaded, skipping loading
I0908 14:02:01.293649 152809 kubeadm.go:926] updating node { 192.168.76.2 8443 v1.34.0 containerd true true} ...
I0908 14:02:01.293756 152809 kubeadm.go:938] kubelet [Unit]
Wants=containerd.service
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.34.0/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=scheduled-stop-160137 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.76.2
[Install]
config:
{KubernetesVersion:v1.34.0 ClusterName:scheduled-stop-160137 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0908 14:02:01.293823 152809 ssh_runner.go:195] Run: sudo crictl info
I0908 14:02:01.334039 152809 cni.go:84] Creating CNI manager for ""
I0908 14:02:01.334050 152809 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I0908 14:02:01.334059 152809 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0908 14:02:01.334079 152809 kubeadm.go:189] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.76.2 APIServerPort:8443 KubernetesVersion:v1.34.0 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:scheduled-stop-160137 NodeName:scheduled-stop-160137 DNSDomain:cluster.local CRISocket:/run/containerd/containerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.76.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.76.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt St
aticPodPath:/etc/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///run/containerd/containerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0908 14:02:01.334190 152809 kubeadm.go:195] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta4
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.76.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///run/containerd/containerd.sock
name: "scheduled-stop-160137"
kubeletExtraArgs:
- name: "node-ip"
value: "192.168.76.2"
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta4
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.76.2"]
extraArgs:
- name: "enable-admission-plugins"
value: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
- name: "allocate-node-cidrs"
value: "true"
- name: "leader-elect"
value: "false"
scheduler:
extraArgs:
- name: "leader-elect"
value: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
kubernetesVersion: v1.34.0
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///run/containerd/containerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0908 14:02:01.334259 152809 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.34.0
I0908 14:02:01.343572 152809 binaries.go:44] Found k8s binaries, skipping transfer
I0908 14:02:01.343636 152809 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0908 14:02:01.352634 152809 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (325 bytes)
I0908 14:02:01.372239 152809 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0908 14:02:01.392588 152809 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2234 bytes)
I0908 14:02:01.411623 152809 ssh_runner.go:195] Run: grep 192.168.76.2 control-plane.minikube.internal$ /etc/hosts
I0908 14:02:01.415282 152809 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.76.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0908 14:02:01.426242 152809 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0908 14:02:01.511401 152809 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0908 14:02:01.528198 152809 certs.go:68] Setting up /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137 for IP: 192.168.76.2
I0908 14:02:01.528210 152809 certs.go:194] generating shared ca certs ...
I0908 14:02:01.528227 152809 certs.go:226] acquiring lock for ca certs: {Name:mke132b78a39150f004355d03d18e99cfccd0efd Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:02:01.528441 152809 certs.go:235] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/21504-2314/.minikube/ca.key
I0908 14:02:01.528493 152809 certs.go:235] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/21504-2314/.minikube/proxy-client-ca.key
I0908 14:02:01.528499 152809 certs.go:256] generating profile certs ...
I0908 14:02:01.528557 152809 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/client.key
I0908 14:02:01.528575 152809 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/client.crt with IP's: []
I0908 14:02:02.235155 152809 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/client.crt ...
I0908 14:02:02.235173 152809 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/client.crt: {Name:mk0a51fbdf29af1b9109cef68eb177e7cf646d1a Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:02:02.235376 152809 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/client.key ...
I0908 14:02:02.235387 152809 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/client.key: {Name:mk6deb45f52ffce9f428d050b8baf515a2c1b72d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:02:02.235512 152809 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/apiserver.key.4bce23d1
I0908 14:02:02.235532 152809 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/apiserver.crt.4bce23d1 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.76.2]
I0908 14:02:02.497889 152809 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/apiserver.crt.4bce23d1 ...
I0908 14:02:02.497904 152809 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/apiserver.crt.4bce23d1: {Name:mkcc534aac4e22dbb4ec58dd3f28fee22ff23bba Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:02:02.498098 152809 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/apiserver.key.4bce23d1 ...
I0908 14:02:02.498106 152809 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/apiserver.key.4bce23d1: {Name:mk21f1f11ad586ebd907ae825687e3a88073eea7 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:02:02.498192 152809 certs.go:381] copying /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/apiserver.crt.4bce23d1 -> /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/apiserver.crt
I0908 14:02:02.498267 152809 certs.go:385] copying /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/apiserver.key.4bce23d1 -> /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/apiserver.key
I0908 14:02:02.498317 152809 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/proxy-client.key
I0908 14:02:02.498330 152809 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/proxy-client.crt with IP's: []
I0908 14:02:03.184389 152809 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/proxy-client.crt ...
I0908 14:02:03.184406 152809 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/proxy-client.crt: {Name:mk46af2331ff689498db959d70bb92c81e3d3490 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:02:03.184602 152809 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/proxy-client.key ...
I0908 14:02:03.184610 152809 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/proxy-client.key: {Name:mk6481bd02fd2e2c2b2a8bde8662d2fd4a1a4d68 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:02:03.184796 152809 certs.go:484] found cert: /home/jenkins/minikube-integration/21504-2314/.minikube/certs/4118.pem (1338 bytes)
W0908 14:02:03.184834 152809 certs.go:480] ignoring /home/jenkins/minikube-integration/21504-2314/.minikube/certs/4118_empty.pem, impossibly tiny 0 bytes
I0908 14:02:03.184849 152809 certs.go:484] found cert: /home/jenkins/minikube-integration/21504-2314/.minikube/certs/ca-key.pem (1675 bytes)
I0908 14:02:03.184876 152809 certs.go:484] found cert: /home/jenkins/minikube-integration/21504-2314/.minikube/certs/ca.pem (1078 bytes)
I0908 14:02:03.184898 152809 certs.go:484] found cert: /home/jenkins/minikube-integration/21504-2314/.minikube/certs/cert.pem (1123 bytes)
I0908 14:02:03.184918 152809 certs.go:484] found cert: /home/jenkins/minikube-integration/21504-2314/.minikube/certs/key.pem (1679 bytes)
I0908 14:02:03.184962 152809 certs.go:484] found cert: /home/jenkins/minikube-integration/21504-2314/.minikube/files/etc/ssl/certs/41182.pem (1708 bytes)
I0908 14:02:03.185655 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0908 14:02:03.210681 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I0908 14:02:03.235652 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0908 14:02:03.259667 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I0908 14:02:03.284544 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1432 bytes)
I0908 14:02:03.310655 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
I0908 14:02:03.335879 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0908 14:02:03.360560 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/profiles/scheduled-stop-160137/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
I0908 14:02:03.384823 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/files/etc/ssl/certs/41182.pem --> /usr/share/ca-certificates/41182.pem (1708 bytes)
I0908 14:02:03.409523 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0908 14:02:03.434381 152809 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21504-2314/.minikube/certs/4118.pem --> /usr/share/ca-certificates/4118.pem (1338 bytes)
I0908 14:02:03.459798 152809 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0908 14:02:03.479117 152809 ssh_runner.go:195] Run: openssl version
I0908 14:02:03.484797 152809 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/41182.pem && ln -fs /usr/share/ca-certificates/41182.pem /etc/ssl/certs/41182.pem"
I0908 14:02:03.494906 152809 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/41182.pem
I0908 14:02:03.498626 152809 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 Sep 8 13:35 /usr/share/ca-certificates/41182.pem
I0908 14:02:03.498693 152809 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/41182.pem
I0908 14:02:03.505782 152809 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/41182.pem /etc/ssl/certs/3ec20f2e.0"
I0908 14:02:03.515615 152809 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0908 14:02:03.525712 152809 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0908 14:02:03.529412 152809 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 8 13:27 /usr/share/ca-certificates/minikubeCA.pem
I0908 14:02:03.529474 152809 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0908 14:02:03.536718 152809 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0908 14:02:03.546369 152809 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/4118.pem && ln -fs /usr/share/ca-certificates/4118.pem /etc/ssl/certs/4118.pem"
I0908 14:02:03.555983 152809 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/4118.pem
I0908 14:02:03.559540 152809 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 Sep 8 13:35 /usr/share/ca-certificates/4118.pem
I0908 14:02:03.559613 152809 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/4118.pem
I0908 14:02:03.566758 152809 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/4118.pem /etc/ssl/certs/51391683.0"
I0908 14:02:03.577904 152809 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0908 14:02:03.583582 152809 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0908 14:02:03.583633 152809 kubeadm.go:392] StartCluster: {Name:scheduled-stop-160137 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.47-1756980985-21488@sha256:8004ef31c95f43ea4d909587f47b84b33af26368a459c00cd53d571affb59c79 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.0 ClusterName:scheduled-stop-160137 Namespace:default APIServerHAVIP: APIServerName:minikubeC
A APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.0 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAut
hSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0908 14:02:03.583701 152809 cri.go:54] listing CRI containers in root /run/containerd/runc/k8s.io: {State:paused Name: Namespaces:[kube-system]}
I0908 14:02:03.583764 152809 ssh_runner.go:195] Run: sudo -s eval "crictl ps -a --quiet --label io.kubernetes.pod.namespace=kube-system"
I0908 14:02:03.622126 152809 cri.go:89] found id: ""
I0908 14:02:03.622208 152809 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0908 14:02:03.631707 152809 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0908 14:02:03.641112 152809 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0908 14:02:03.641167 152809 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0908 14:02:03.650640 152809 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0908 14:02:03.650655 152809 kubeadm.go:157] found existing configuration files:
I0908 14:02:03.650706 152809 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0908 14:02:03.659535 152809 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0908 14:02:03.659591 152809 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0908 14:02:03.668214 152809 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0908 14:02:03.676886 152809 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0908 14:02:03.676954 152809 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0908 14:02:03.685758 152809 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0908 14:02:03.694390 152809 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0908 14:02:03.694493 152809 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0908 14:02:03.703179 152809 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0908 14:02:03.712270 152809 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0908 14:02:03.712326 152809 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0908 14:02:03.721503 152809 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.34.0:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0908 14:02:03.765201 152809 kubeadm.go:310] [init] Using Kubernetes version: v1.34.0
I0908 14:02:03.765273 152809 kubeadm.go:310] [preflight] Running pre-flight checks
I0908 14:02:03.782399 152809 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0908 14:02:03.782462 152809 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1084-aws[0m
I0908 14:02:03.782497 152809 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0908 14:02:03.782549 152809 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0908 14:02:03.782597 152809 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0908 14:02:03.782644 152809 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0908 14:02:03.782699 152809 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0908 14:02:03.782750 152809 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0908 14:02:03.782797 152809 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0908 14:02:03.782841 152809 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0908 14:02:03.782889 152809 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0908 14:02:03.782935 152809 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0908 14:02:03.850445 152809 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0908 14:02:03.850550 152809 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0908 14:02:03.850642 152809 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0908 14:02:03.860818 152809 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0908 14:02:03.867275 152809 out.go:252] - Generating certificates and keys ...
I0908 14:02:03.867359 152809 kubeadm.go:310] [certs] Using existing ca certificate authority
I0908 14:02:03.867426 152809 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0908 14:02:04.709086 152809 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0908 14:02:05.361276 152809 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0908 14:02:06.248429 152809 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0908 14:02:06.459735 152809 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0908 14:02:06.647889 152809 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0908 14:02:06.648233 152809 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [localhost scheduled-stop-160137] and IPs [192.168.76.2 127.0.0.1 ::1]
I0908 14:02:06.896236 152809 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0908 14:02:06.896833 152809 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [localhost scheduled-stop-160137] and IPs [192.168.76.2 127.0.0.1 ::1]
I0908 14:02:07.559094 152809 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0908 14:02:07.909212 152809 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0908 14:02:08.439521 152809 kubeadm.go:310] [certs] Generating "sa" key and public key
I0908 14:02:08.439746 152809 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0908 14:02:08.977135 152809 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0908 14:02:09.688590 152809 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0908 14:02:09.858486 152809 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0908 14:02:10.000264 152809 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0908 14:02:10.484386 152809 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0908 14:02:10.485253 152809 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0908 14:02:10.487959 152809 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0908 14:02:10.493674 152809 out.go:252] - Booting up control plane ...
I0908 14:02:10.493810 152809 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0908 14:02:10.493913 152809 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0908 14:02:10.493991 152809 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0908 14:02:10.504088 152809 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0908 14:02:10.504350 152809 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/instance-config.yaml"
I0908 14:02:10.511553 152809 kubeadm.go:310] [patches] Applied patch of type "application/strategic-merge-patch+json" to target "kubeletconfiguration"
I0908 14:02:10.512691 152809 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0908 14:02:10.512738 152809 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0908 14:02:10.614363 152809 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0908 14:02:10.614477 152809 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0908 14:02:11.625423 152809 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 1.011172178s
I0908 14:02:11.629153 152809 kubeadm.go:310] [control-plane-check] Waiting for healthy control plane components. This can take up to 4m0s
I0908 14:02:11.629356 152809 kubeadm.go:310] [control-plane-check] Checking kube-apiserver at https://192.168.76.2:8443/livez
I0908 14:02:11.629558 152809 kubeadm.go:310] [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz
I0908 14:02:11.629639 152809 kubeadm.go:310] [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez
I0908 14:02:14.469282 152809 kubeadm.go:310] [control-plane-check] kube-controller-manager is healthy after 2.839568769s
I0908 14:02:15.730613 152809 kubeadm.go:310] [control-plane-check] kube-scheduler is healthy after 4.101379283s
I0908 14:02:17.630750 152809 kubeadm.go:310] [control-plane-check] kube-apiserver is healthy after 6.00148528s
I0908 14:02:17.652619 152809 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0908 14:02:17.677324 152809 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0908 14:02:17.697028 152809 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0908 14:02:17.697227 152809 kubeadm.go:310] [mark-control-plane] Marking the node scheduled-stop-160137 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0908 14:02:17.709493 152809 kubeadm.go:310] [bootstrap-token] Using token: 7ew1dq.j6g33mwbz8hy5iji
I0908 14:02:17.712380 152809 out.go:252] - Configuring RBAC rules ...
I0908 14:02:17.712501 152809 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0908 14:02:17.720563 152809 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0908 14:02:17.734245 152809 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0908 14:02:17.741701 152809 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0908 14:02:17.747168 152809 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0908 14:02:17.754929 152809 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0908 14:02:18.038332 152809 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0908 14:02:18.475839 152809 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0908 14:02:19.038533 152809 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0908 14:02:19.039499 152809 kubeadm.go:310]
I0908 14:02:19.039566 152809 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0908 14:02:19.039570 152809 kubeadm.go:310]
I0908 14:02:19.039647 152809 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0908 14:02:19.039651 152809 kubeadm.go:310]
I0908 14:02:19.039677 152809 kubeadm.go:310] mkdir -p $HOME/.kube
I0908 14:02:19.039736 152809 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0908 14:02:19.039825 152809 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0908 14:02:19.039835 152809 kubeadm.go:310]
I0908 14:02:19.039890 152809 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0908 14:02:19.039893 152809 kubeadm.go:310]
I0908 14:02:19.039946 152809 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0908 14:02:19.039958 152809 kubeadm.go:310]
I0908 14:02:19.040011 152809 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0908 14:02:19.040091 152809 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0908 14:02:19.040170 152809 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0908 14:02:19.040174 152809 kubeadm.go:310]
I0908 14:02:19.040267 152809 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0908 14:02:19.040346 152809 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0908 14:02:19.040349 152809 kubeadm.go:310]
I0908 14:02:19.040462 152809 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 7ew1dq.j6g33mwbz8hy5iji \
I0908 14:02:19.040566 152809 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:95656961684b4abed2a441b60d9e5046bee63a59d43389631e28f6a3337554cd \
I0908 14:02:19.040587 152809 kubeadm.go:310] --control-plane
I0908 14:02:19.040590 152809 kubeadm.go:310]
I0908 14:02:19.040675 152809 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0908 14:02:19.040679 152809 kubeadm.go:310]
I0908 14:02:19.040760 152809 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 7ew1dq.j6g33mwbz8hy5iji \
I0908 14:02:19.040863 152809 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:95656961684b4abed2a441b60d9e5046bee63a59d43389631e28f6a3337554cd
I0908 14:02:19.044704 152809 kubeadm.go:310] [WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2
I0908 14:02:19.044925 152809 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1084-aws\n", err: exit status 1
I0908 14:02:19.045044 152809 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0908 14:02:19.045060 152809 cni.go:84] Creating CNI manager for ""
I0908 14:02:19.045066 152809 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I0908 14:02:19.048112 152809 out.go:179] * Configuring CNI (Container Networking Interface) ...
I0908 14:02:19.050916 152809 ssh_runner.go:195] Run: stat /opt/cni/bin/portmap
I0908 14:02:19.054547 152809 cni.go:182] applying CNI manifest using /var/lib/minikube/binaries/v1.34.0/kubectl ...
I0908 14:02:19.054556 152809 ssh_runner.go:362] scp memory --> /var/tmp/minikube/cni.yaml (2601 bytes)
I0908 14:02:19.074663 152809 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.0/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml
I0908 14:02:19.363729 152809 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0908 14:02:19.363894 152809 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.0/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0908 14:02:19.363978 152809 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes scheduled-stop-160137 minikube.k8s.io/updated_at=2025_09_08T14_02_19_0700 minikube.k8s.io/version=v1.36.0 minikube.k8s.io/commit=4237956cfce90d4ab760d817400bd4c89cad50d6 minikube.k8s.io/name=scheduled-stop-160137 minikube.k8s.io/primary=true
I0908 14:02:19.493330 152809 ops.go:34] apiserver oom_adj: -16
I0908 14:02:19.493349 152809 kubeadm.go:1105] duration metric: took 129.539712ms to wait for elevateKubeSystemPrivileges
I0908 14:02:19.493360 152809 kubeadm.go:394] duration metric: took 15.909730621s to StartCluster
I0908 14:02:19.493374 152809 settings.go:142] acquiring lock: {Name:mk4f8717708db28eef58408fb347a7d2170243fe Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:02:19.493437 152809 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/21504-2314/kubeconfig
I0908 14:02:19.494084 152809 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21504-2314/kubeconfig: {Name:mk59ae76c24dca3eb03e6fa665ed1169acb8310d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0908 14:02:19.494285 152809 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.0 ContainerRuntime:containerd ControlPlane:true Worker:true}
I0908 14:02:19.494398 152809 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.34.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0908 14:02:19.494624 152809 config.go:182] Loaded profile config "scheduled-stop-160137": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.34.0
I0908 14:02:19.494690 152809 addons.go:511] enable addons start: toEnable=map[ambassador:false amd-gpu-device-plugin:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubetail:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:false volcano:false volumesnapshots:false yakd:false]
I0908 14:02:19.494750 152809 addons.go:69] Setting storage-provisioner=true in profile "scheduled-stop-160137"
I0908 14:02:19.494763 152809 addons.go:238] Setting addon storage-provisioner=true in "scheduled-stop-160137"
I0908 14:02:19.494783 152809 host.go:66] Checking if "scheduled-stop-160137" exists ...
I0908 14:02:19.495279 152809 cli_runner.go:164] Run: docker container inspect scheduled-stop-160137 --format={{.State.Status}}
I0908 14:02:19.495691 152809 addons.go:69] Setting default-storageclass=true in profile "scheduled-stop-160137"
I0908 14:02:19.495708 152809 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "scheduled-stop-160137"
I0908 14:02:19.495997 152809 cli_runner.go:164] Run: docker container inspect scheduled-stop-160137 --format={{.State.Status}}
I0908 14:02:19.498471 152809 out.go:179] * Verifying Kubernetes components...
I0908 14:02:19.501470 152809 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0908 14:02:19.537028 152809 out.go:179] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0908 14:02:19.539167 152809 addons.go:238] Setting addon default-storageclass=true in "scheduled-stop-160137"
I0908 14:02:19.539194 152809 host.go:66] Checking if "scheduled-stop-160137" exists ...
I0908 14:02:19.539606 152809 cli_runner.go:164] Run: docker container inspect scheduled-stop-160137 --format={{.State.Status}}
I0908 14:02:19.541693 152809 addons.go:435] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0908 14:02:19.541704 152809 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0908 14:02:19.541761 152809 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-160137
I0908 14:02:19.584513 152809 addons.go:435] installing /etc/kubernetes/addons/storageclass.yaml
I0908 14:02:19.584525 152809 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0908 14:02:19.584616 152809 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-160137
I0908 14:02:19.592348 152809 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32968 SSHKeyPath:/home/jenkins/minikube-integration/21504-2314/.minikube/machines/scheduled-stop-160137/id_rsa Username:docker}
I0908 14:02:19.621704 152809 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32968 SSHKeyPath:/home/jenkins/minikube-integration/21504-2314/.minikube/machines/scheduled-stop-160137/id_rsa Username:docker}
I0908 14:02:19.801557 152809 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.34.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.76.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.34.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0908 14:02:19.801663 152809 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0908 14:02:19.805336 152809 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.34.0/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0908 14:02:19.827307 152809 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.34.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0908 14:02:20.281568 152809 api_server.go:52] waiting for apiserver process to appear ...
I0908 14:02:20.281637 152809 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0908 14:02:20.281722 152809 start.go:976] {"host.minikube.internal": 192.168.76.1} host record injected into CoreDNS's ConfigMap
I0908 14:02:20.393401 152809 api_server.go:72] duration metric: took 899.090682ms to wait for apiserver process to appear ...
I0908 14:02:20.393412 152809 api_server.go:88] waiting for apiserver healthz status ...
I0908 14:02:20.393429 152809 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
I0908 14:02:20.396220 152809 out.go:179] * Enabled addons: default-storageclass, storage-provisioner
I0908 14:02:20.399909 152809 addons.go:514] duration metric: took 905.224599ms for enable addons: enabled=[default-storageclass storage-provisioner]
I0908 14:02:20.404998 152809 api_server.go:279] https://192.168.76.2:8443/healthz returned 200:
ok
I0908 14:02:20.406071 152809 api_server.go:141] control plane version: v1.34.0
I0908 14:02:20.406093 152809 api_server.go:131] duration metric: took 12.675083ms to wait for apiserver health ...
I0908 14:02:20.406101 152809 system_pods.go:43] waiting for kube-system pods to appear ...
I0908 14:02:20.408914 152809 system_pods.go:59] 5 kube-system pods found
I0908 14:02:20.408938 152809 system_pods.go:61] "etcd-scheduled-stop-160137" [d0a800b1-1612-4cf1-81dd-6971c9de4c21] Running / Ready:ContainersNotReady (containers with unready status: [etcd]) / ContainersReady:ContainersNotReady (containers with unready status: [etcd])
I0908 14:02:20.408945 152809 system_pods.go:61] "kube-apiserver-scheduled-stop-160137" [96835325-3a90-4104-a50d-68d34a53e068] Running / Ready:ContainersNotReady (containers with unready status: [kube-apiserver]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-apiserver])
I0908 14:02:20.408956 152809 system_pods.go:61] "kube-controller-manager-scheduled-stop-160137" [38d52c79-bf9e-4186-a82f-228c055b5847] Running / Ready:ContainersNotReady (containers with unready status: [kube-controller-manager]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-controller-manager])
I0908 14:02:20.408963 152809 system_pods.go:61] "kube-scheduler-scheduled-stop-160137" [fbcd5c37-9cb8-44ba-a744-96af01623051] Running / Ready:ContainersNotReady (containers with unready status: [kube-scheduler]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-scheduler])
I0908 14:02:20.408967 152809 system_pods.go:61] "storage-provisioner" [d8ec5534-5260-45ae-a980-e0c94bbb3244] Pending: PodScheduled:Unschedulable (0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. no new claims to deallocate, preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.)
I0908 14:02:20.408971 152809 system_pods.go:74] duration metric: took 2.866845ms to wait for pod list to return data ...
I0908 14:02:20.408980 152809 kubeadm.go:578] duration metric: took 914.677633ms to wait for: map[apiserver:true system_pods:true]
I0908 14:02:20.408993 152809 node_conditions.go:102] verifying NodePressure condition ...
I0908 14:02:20.411537 152809 node_conditions.go:122] node storage ephemeral capacity is 203034800Ki
I0908 14:02:20.411554 152809 node_conditions.go:123] node cpu capacity is 2
I0908 14:02:20.411565 152809 node_conditions.go:105] duration metric: took 2.568479ms to run NodePressure ...
I0908 14:02:20.411577 152809 start.go:241] waiting for startup goroutines ...
I0908 14:02:20.786416 152809 kapi.go:214] "coredns" deployment in "kube-system" namespace and "scheduled-stop-160137" context rescaled to 1 replicas
I0908 14:02:20.786438 152809 start.go:246] waiting for cluster config update ...
I0908 14:02:20.786447 152809 start.go:255] writing updated cluster config ...
I0908 14:02:20.786752 152809 ssh_runner.go:195] Run: rm -f paused
I0908 14:02:20.846924 152809 start.go:617] kubectl: 1.33.2, cluster: 1.34.0 (minor skew: 1)
I0908 14:02:20.850235 152809 out.go:179] * Done! kubectl is now configured to use "scheduled-stop-160137" cluster and "default" namespace by default
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
0184e8be01d87 996be7e86d9b3 10 seconds ago Running kube-controller-manager 0 1f6d67a0cb9e5 kube-controller-manager-scheduled-stop-160137
04b7c58417b31 a1894772a478e 10 seconds ago Running etcd 0 b84f4cd2b52fa etcd-scheduled-stop-160137
c9982c44bd861 d291939e99406 10 seconds ago Running kube-apiserver 0 37e87d323e020 kube-apiserver-scheduled-stop-160137
0a8bfcfa70eed a25f5ef9c34c3 10 seconds ago Running kube-scheduler 0 0d49b4ce5b00c kube-scheduler-scheduled-stop-160137
==> containerd <==
Sep 08 14:02:01 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:01.075735295Z" level=info msg="containerd successfully booted in 0.088388s"
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.033058938Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-scheduled-stop-160137,Uid:ffbb50371569b5ad6ebfc4161ae551a5,Namespace:kube-system,Attempt:0,}"
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.039575792Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:etcd-scheduled-stop-160137,Uid:c7181bb829f59210ffc51b2fd67e0217,Namespace:kube-system,Attempt:0,}"
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.043664838Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-scheduled-stop-160137,Uid:c457d52c24dbccf78b0af8f0ab04285c,Namespace:kube-system,Attempt:0,}"
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.051832959Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-scheduled-stop-160137,Uid:e23c7c4cd1f10c12a726a5f3f5563375,Namespace:kube-system,Attempt:0,}"
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.150488260Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-scheduled-stop-160137,Uid:ffbb50371569b5ad6ebfc4161ae551a5,Namespace:kube-system,Attempt:0,} returns sandbox id \"0d49b4ce5b00c28e8becba27d856dc0dd8ef0265440a7e9d64d9b57a71bce775\""
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.159595190Z" level=info msg="CreateContainer within sandbox \"0d49b4ce5b00c28e8becba27d856dc0dd8ef0265440a7e9d64d9b57a71bce775\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}"
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.185289295Z" level=info msg="CreateContainer within sandbox \"0d49b4ce5b00c28e8becba27d856dc0dd8ef0265440a7e9d64d9b57a71bce775\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"0a8bfcfa70eed98313b87318bdfa580f89a9d4908479637f3827d093cf7a1133\""
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.186075421Z" level=info msg="StartContainer for \"0a8bfcfa70eed98313b87318bdfa580f89a9d4908479637f3827d093cf7a1133\""
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.272685676Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-scheduled-stop-160137,Uid:c457d52c24dbccf78b0af8f0ab04285c,Namespace:kube-system,Attempt:0,} returns sandbox id \"37e87d323e020414e1ef262b75ea29a0b15e5ee553405b8a03222719bceff241\""
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.291412665Z" level=info msg="CreateContainer within sandbox \"37e87d323e020414e1ef262b75ea29a0b15e5ee553405b8a03222719bceff241\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}"
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.323552444Z" level=info msg="StartContainer for \"0a8bfcfa70eed98313b87318bdfa580f89a9d4908479637f3827d093cf7a1133\" returns successfully"
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.326078946Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:etcd-scheduled-stop-160137,Uid:c7181bb829f59210ffc51b2fd67e0217,Namespace:kube-system,Attempt:0,} returns sandbox id \"b84f4cd2b52fae060d4384fa3464138b6bb950027315245b72f76ce402e95f27\""
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.326778719Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-scheduled-stop-160137,Uid:e23c7c4cd1f10c12a726a5f3f5563375,Namespace:kube-system,Attempt:0,} returns sandbox id \"1f6d67a0cb9e5a777b653db7280331390659ed4c26433a08f26f19849d916c4e\""
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.334547879Z" level=info msg="CreateContainer within sandbox \"b84f4cd2b52fae060d4384fa3464138b6bb950027315245b72f76ce402e95f27\" for container &ContainerMetadata{Name:etcd,Attempt:0,}"
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.337496571Z" level=info msg="CreateContainer within sandbox \"1f6d67a0cb9e5a777b653db7280331390659ed4c26433a08f26f19849d916c4e\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}"
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.346473547Z" level=info msg="CreateContainer within sandbox \"37e87d323e020414e1ef262b75ea29a0b15e5ee553405b8a03222719bceff241\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"c9982c44bd861361732aa2f07498f8ab2a24919eabe2bda48a6e2eb4f786ce56\""
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.347333995Z" level=info msg="StartContainer for \"c9982c44bd861361732aa2f07498f8ab2a24919eabe2bda48a6e2eb4f786ce56\""
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.394748698Z" level=info msg="CreateContainer within sandbox \"b84f4cd2b52fae060d4384fa3464138b6bb950027315245b72f76ce402e95f27\" for &ContainerMetadata{Name:etcd,Attempt:0,} returns container id \"04b7c58417b31f7905ee0228718c91461c6989db69e2870831f4c3788b704541\""
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.395530507Z" level=info msg="StartContainer for \"04b7c58417b31f7905ee0228718c91461c6989db69e2870831f4c3788b704541\""
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.412649413Z" level=info msg="CreateContainer within sandbox \"1f6d67a0cb9e5a777b653db7280331390659ed4c26433a08f26f19849d916c4e\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"0184e8be01d87b1545a60040e0a2b8d62e841025be306a2f411af3c0562e4b7a\""
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.413340711Z" level=info msg="StartContainer for \"0184e8be01d87b1545a60040e0a2b8d62e841025be306a2f411af3c0562e4b7a\""
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.474282486Z" level=info msg="StartContainer for \"c9982c44bd861361732aa2f07498f8ab2a24919eabe2bda48a6e2eb4f786ce56\" returns successfully"
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.557963421Z" level=info msg="StartContainer for \"04b7c58417b31f7905ee0228718c91461c6989db69e2870831f4c3788b704541\" returns successfully"
Sep 08 14:02:12 scheduled-stop-160137 containerd[837]: time="2025-09-08T14:02:12.630706817Z" level=info msg="StartContainer for \"0184e8be01d87b1545a60040e0a2b8d62e841025be306a2f411af3c0562e4b7a\" returns successfully"
==> describe nodes <==
Name: scheduled-stop-160137
Roles: control-plane
Labels: beta.kubernetes.io/arch=arm64
beta.kubernetes.io/os=linux
kubernetes.io/arch=arm64
kubernetes.io/hostname=scheduled-stop-160137
kubernetes.io/os=linux
minikube.k8s.io/commit=4237956cfce90d4ab760d817400bd4c89cad50d6
minikube.k8s.io/name=scheduled-stop-160137
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2025_09_08T14_02_19_0700
minikube.k8s.io/version=v1.36.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
Annotations: volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Mon, 08 Sep 2025 14:02:15 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: scheduled-stop-160137
AcquireTime: <unset>
RenewTime: Mon, 08 Sep 2025 14:02:18 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Mon, 08 Sep 2025 14:02:18 +0000 Mon, 08 Sep 2025 14:02:13 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Mon, 08 Sep 2025 14:02:18 +0000 Mon, 08 Sep 2025 14:02:13 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Mon, 08 Sep 2025 14:02:18 +0000 Mon, 08 Sep 2025 14:02:13 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Mon, 08 Sep 2025 14:02:18 +0000 Mon, 08 Sep 2025 14:02:15 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.76.2
Hostname: scheduled-stop-160137
Capacity:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022296Ki
pods: 110
Allocatable:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022296Ki
pods: 110
System Info:
Machine ID: 0f20ba50c6bf4a5b8283cbb26fa967ed
System UUID: 82836647-7b69-47b6-bf86-0a58c58dc3ed
Boot ID: e9996d3c-7ca0-44f4-a0bc-36bb577e6736
Kernel Version: 5.15.0-1084-aws
OS Image: Ubuntu 22.04.5 LTS
Operating System: linux
Architecture: arm64
Container Runtime Version: containerd://1.7.27
Kubelet Version: v1.34.0
Kube-Proxy Version:
Non-terminated Pods: (5 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
kube-system etcd-scheduled-stop-160137 100m (5%) 0 (0%) 100Mi (1%) 0 (0%) 4s
kube-system kube-apiserver-scheduled-stop-160137 250m (12%) 0 (0%) 0 (0%) 0 (0%) 5s
kube-system kube-controller-manager-scheduled-stop-160137 200m (10%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-scheduler-scheduled-stop-160137 100m (5%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 2s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 650m (32%) 0 (0%)
memory 100Mi (1%) 0 (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
hugepages-32Mi 0 (0%) 0 (0%)
hugepages-64Ki 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 4s kubelet Starting kubelet.
Warning CgroupV1 4s kubelet cgroup v1 support is in maintenance mode, please migrate to cgroup v2
Normal NodeAllocatableEnforced 4s kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 4s kubelet Node scheduled-stop-160137 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 4s kubelet Node scheduled-stop-160137 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 4s kubelet Node scheduled-stop-160137 status is now: NodeHasSufficientPID
Normal RegisteredNode 0s node-controller Node scheduled-stop-160137 event: Registered Node scheduled-stop-160137 in Controller
==> dmesg <==
[Sep 8 13:17] ACPI: SRAT not present
[ +0.000000] ACPI: SRAT not present
[ +0.000000] SPI driver altr_a10sr has no spi_device_id for altr,a10sr
[ +0.014416] device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log.
[ +0.488283] systemd[1]: Configuration file /run/systemd/system/netplan-ovs-cleanup.service is marked world-inaccessible. This has no effect as configuration data is accessible via APIs without restrictions. Proceeding anyway.
[ +0.036945] systemd[1]: /lib/systemd/system/snapd.service:23: Unknown key name 'RestartMode' in section 'Service', ignoring.
[ +0.751194] ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy.
[ +6.289622] kauditd_printk_skb: 36 callbacks suppressed
[Sep 8 13:57] hrtimer: interrupt took 7896607 ns
==> etcd [04b7c58417b31f7905ee0228718c91461c6989db69e2870831f4c3788b704541] <==
{"level":"warn","ts":"2025-09-08T14:02:14.380445Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54698","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.420447Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54714","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.456013Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54720","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.472432Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54744","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.489887Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54756","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.505752Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54774","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.523176Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54788","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.541706Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54810","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.559709Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54826","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.577361Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54840","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.595427Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54864","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.623349Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54888","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.637932Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54896","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.655829Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54916","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.676531Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54944","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.693175Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54956","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.712922Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54976","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.725015Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:54982","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.741134Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:55002","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.752728Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:55026","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.774022Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:55044","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.795903Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:55058","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.812859Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:55080","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.829346Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:55096","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-09-08T14:02:14.908455Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:55112","server-name":"","error":"EOF"}
==> kernel <==
14:02:22 up 44 min, 0 users, load average: 1.84, 1.92, 2.06
Linux scheduled-stop-160137 5.15.0-1084-aws #91~20.04.1-Ubuntu SMP Fri May 2 07:00:04 UTC 2025 aarch64 aarch64 aarch64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.5 LTS"
==> kube-apiserver [c9982c44bd861361732aa2f07498f8ab2a24919eabe2bda48a6e2eb4f786ce56] <==
I0908 14:02:15.745622 1 cache.go:39] Caches are synced for autoregister controller
I0908 14:02:15.751767 1 controller.go:667] quota admission added evaluator for: leases.coordination.k8s.io
I0908 14:02:15.756895 1 default_servicecidr_controller.go:228] Setting default ServiceCIDR condition Ready to True
I0908 14:02:15.760822 1 cache.go:39] Caches are synced for APIServiceRegistrationController controller
I0908 14:02:15.760881 1 cache.go:39] Caches are synced for RemoteAvailability controller
I0908 14:02:15.761125 1 handler_discovery.go:451] Starting ResourceDiscoveryManager
I0908 14:02:15.761332 1 apf_controller.go:382] Running API Priority and Fairness config worker
I0908 14:02:15.761349 1 apf_controller.go:385] Running API Priority and Fairness periodic rebalancing process
I0908 14:02:15.761862 1 cache.go:39] Caches are synced for LocalAvailability controller
I0908 14:02:15.776663 1 cidrallocator.go:301] created ClusterIP allocator for Service CIDR 10.96.0.0/12
I0908 14:02:15.792550 1 cidrallocator.go:277] updated ClusterIP allocator for Service CIDR 10.96.0.0/12
I0908 14:02:15.795361 1 default_servicecidr_controller.go:137] Shutting down kubernetes-service-cidr-controller
I0908 14:02:16.438520 1 storage_scheduling.go:95] created PriorityClass system-node-critical with value 2000001000
I0908 14:02:16.443635 1 storage_scheduling.go:95] created PriorityClass system-cluster-critical with value 2000000000
I0908 14:02:16.443658 1 storage_scheduling.go:111] all system priority classes are created successfully or already exist.
I0908 14:02:17.252056 1 controller.go:667] quota admission added evaluator for: roles.rbac.authorization.k8s.io
I0908 14:02:17.310479 1 controller.go:667] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
I0908 14:02:17.447930 1 alloc.go:328] "allocated clusterIPs" service="default/kubernetes" clusterIPs={"IPv4":"10.96.0.1"}
W0908 14:02:17.460693 1 lease.go:265] Resetting endpoints for master service "kubernetes" to [192.168.76.2]
I0908 14:02:17.461938 1 controller.go:667] quota admission added evaluator for: endpoints
I0908 14:02:17.470565 1 controller.go:667] quota admission added evaluator for: endpointslices.discovery.k8s.io
I0908 14:02:17.623234 1 controller.go:667] quota admission added evaluator for: serviceaccounts
I0908 14:02:18.449816 1 controller.go:667] quota admission added evaluator for: deployments.apps
I0908 14:02:18.473454 1 alloc.go:328] "allocated clusterIPs" service="kube-system/kube-dns" clusterIPs={"IPv4":"10.96.0.10"}
I0908 14:02:18.486933 1 controller.go:667] quota admission added evaluator for: daemonsets.apps
==> kube-controller-manager [0184e8be01d87b1545a60040e0a2b8d62e841025be306a2f411af3c0562e4b7a] <==
I0908 14:02:22.644536 1 shared_informer.go:356] "Caches are synced" controller="VAC protection"
I0908 14:02:22.660845 1 shared_informer.go:356] "Caches are synced" controller="ephemeral"
I0908 14:02:22.666102 1 shared_informer.go:356] "Caches are synced" controller="disruption"
I0908 14:02:22.669656 1 shared_informer.go:356] "Caches are synced" controller="stateful set"
I0908 14:02:22.669861 1 shared_informer.go:356] "Caches are synced" controller="ReplicaSet"
I0908 14:02:22.669882 1 shared_informer.go:356] "Caches are synced" controller="PVC protection"
I0908 14:02:22.669909 1 shared_informer.go:356] "Caches are synced" controller="ReplicationController"
I0908 14:02:22.669926 1 shared_informer.go:356] "Caches are synced" controller="endpoint_slice"
I0908 14:02:22.669962 1 shared_informer.go:356] "Caches are synced" controller="HPA"
I0908 14:02:22.669980 1 shared_informer.go:356] "Caches are synced" controller="TTL"
I0908 14:02:22.669998 1 shared_informer.go:356] "Caches are synced" controller="ClusterRoleAggregator"
I0908 14:02:22.671006 1 shared_informer.go:356] "Caches are synced" controller="garbage collector"
I0908 14:02:22.671035 1 shared_informer.go:356] "Caches are synced" controller="legacy-service-account-token-cleaner"
I0908 14:02:22.671159 1 shared_informer.go:356] "Caches are synced" controller="crt configmap"
I0908 14:02:22.671610 1 range_allocator.go:428] "Set node PodCIDR" logger="node-ipam-controller" node="scheduled-stop-160137" podCIDRs=["10.244.0.0/24"]
I0908 14:02:22.673300 1 shared_informer.go:356] "Caches are synced" controller="service account"
I0908 14:02:22.679579 1 shared_informer.go:356] "Caches are synced" controller="namespace"
I0908 14:02:22.679726 1 shared_informer.go:356] "Caches are synced" controller="resource quota"
I0908 14:02:22.685521 1 shared_informer.go:356] "Caches are synced" controller="daemon sets"
I0908 14:02:22.695936 1 shared_informer.go:356] "Caches are synced" controller="certificate-csrsigning-kubelet-serving"
I0908 14:02:22.697133 1 shared_informer.go:356] "Caches are synced" controller="certificate-csrsigning-legacy-unknown"
I0908 14:02:22.698380 1 shared_informer.go:356] "Caches are synced" controller="certificate-csrsigning-kube-apiserver-client"
I0908 14:02:22.715950 1 shared_informer.go:356] "Caches are synced" controller="garbage collector"
I0908 14:02:22.715978 1 garbagecollector.go:154] "Garbage collector: all resource monitors have synced" logger="garbage-collector-controller"
I0908 14:02:22.715985 1 garbagecollector.go:157] "Proceeding to collect garbage" logger="garbage-collector-controller"
==> kube-scheduler [0a8bfcfa70eed98313b87318bdfa580f89a9d4908479637f3827d093cf7a1133] <==
E0908 14:02:15.736584 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ResourceSlice: resourceslices.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"resourceslices\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ResourceSlice"
E0908 14:02:15.736758 1 reflector.go:205] "Failed to watch" err="failed to list *v1.DeviceClass: deviceclasses.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"deviceclasses\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.DeviceClass"
E0908 14:02:15.736949 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace"
E0908 14:02:15.737015 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node"
E0908 14:02:15.737076 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicationController"
E0908 14:02:15.737138 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicaSet"
E0908 14:02:15.737195 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod"
E0908 14:02:15.737244 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ResourceClaim: resourceclaims.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"resourceclaims\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ResourceClaim"
E0908 14:02:15.737313 1 reflector.go:205] "Failed to watch" err="failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.StorageClass"
E0908 14:02:15.737369 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PodDisruptionBudget"
E0908 14:02:15.740278 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIStorageCapacity"
E0908 14:02:15.740431 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSINode"
E0908 14:02:15.740498 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolumeClaim"
E0908 14:02:15.740546 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service"
E0908 14:02:15.740591 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolume"
E0908 14:02:16.646305 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolumeClaim"
E0908 14:02:16.755286 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicationController"
E0908 14:02:16.764003 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PodDisruptionBudget"
E0908 14:02:16.789242 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSINode"
E0908 14:02:16.808941 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ResourceSlice: resourceslices.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"resourceslices\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ResourceSlice"
E0908 14:02:16.853329 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIStorageCapacity"
E0908 14:02:16.858543 1 reflector.go:205] "Failed to watch" err="failed to list *v1.DeviceClass: deviceclasses.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"deviceclasses\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.DeviceClass"
E0908 14:02:16.904573 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolume"
E0908 14:02:16.999892 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError" reflector="runtime/asm_arm64.s:1223" type="*v1.ConfigMap"
I0908 14:02:18.618667 1 shared_informer.go:356] "Caches are synced" controller="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
==> kubelet <==
Sep 08 14:02:18 scheduled-stop-160137 kubelet[1549]: I0908 14:02:18.797757 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ffbb50371569b5ad6ebfc4161ae551a5-kubeconfig\") pod \"kube-scheduler-scheduled-stop-160137\" (UID: \"ffbb50371569b5ad6ebfc4161ae551a5\") " pod="kube-system/kube-scheduler-scheduled-stop-160137"
Sep 08 14:02:18 scheduled-stop-160137 kubelet[1549]: I0908 14:02:18.797781 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-certs\" (UniqueName: \"kubernetes.io/host-path/c7181bb829f59210ffc51b2fd67e0217-etcd-certs\") pod \"etcd-scheduled-stop-160137\" (UID: \"c7181bb829f59210ffc51b2fd67e0217\") " pod="kube-system/etcd-scheduled-stop-160137"
Sep 08 14:02:18 scheduled-stop-160137 kubelet[1549]: I0908 14:02:18.797801 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/c457d52c24dbccf78b0af8f0ab04285c-ca-certs\") pod \"kube-apiserver-scheduled-stop-160137\" (UID: \"c457d52c24dbccf78b0af8f0ab04285c\") " pod="kube-system/kube-apiserver-scheduled-stop-160137"
Sep 08 14:02:18 scheduled-stop-160137 kubelet[1549]: I0908 14:02:18.797819 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/e23c7c4cd1f10c12a726a5f3f5563375-flexvolume-dir\") pod \"kube-controller-manager-scheduled-stop-160137\" (UID: \"e23c7c4cd1f10c12a726a5f3f5563375\") " pod="kube-system/kube-controller-manager-scheduled-stop-160137"
Sep 08 14:02:18 scheduled-stop-160137 kubelet[1549]: I0908 14:02:18.797839 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-data\" (UniqueName: \"kubernetes.io/host-path/c7181bb829f59210ffc51b2fd67e0217-etcd-data\") pod \"etcd-scheduled-stop-160137\" (UID: \"c7181bb829f59210ffc51b2fd67e0217\") " pod="kube-system/etcd-scheduled-stop-160137"
Sep 08 14:02:18 scheduled-stop-160137 kubelet[1549]: I0908 14:02:18.797860 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/c457d52c24dbccf78b0af8f0ab04285c-k8s-certs\") pod \"kube-apiserver-scheduled-stop-160137\" (UID: \"c457d52c24dbccf78b0af8f0ab04285c\") " pod="kube-system/kube-apiserver-scheduled-stop-160137"
Sep 08 14:02:18 scheduled-stop-160137 kubelet[1549]: I0908 14:02:18.797876 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e23c7c4cd1f10c12a726a5f3f5563375-ca-certs\") pod \"kube-controller-manager-scheduled-stop-160137\" (UID: \"e23c7c4cd1f10c12a726a5f3f5563375\") " pod="kube-system/kube-controller-manager-scheduled-stop-160137"
Sep 08 14:02:18 scheduled-stop-160137 kubelet[1549]: I0908 14:02:18.797895 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e23c7c4cd1f10c12a726a5f3f5563375-usr-share-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-160137\" (UID: \"e23c7c4cd1f10c12a726a5f3f5563375\") " pod="kube-system/kube-controller-manager-scheduled-stop-160137"
Sep 08 14:02:18 scheduled-stop-160137 kubelet[1549]: I0908 14:02:18.797913 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/c457d52c24dbccf78b0af8f0ab04285c-etc-ca-certificates\") pod \"kube-apiserver-scheduled-stop-160137\" (UID: \"c457d52c24dbccf78b0af8f0ab04285c\") " pod="kube-system/kube-apiserver-scheduled-stop-160137"
Sep 08 14:02:18 scheduled-stop-160137 kubelet[1549]: I0908 14:02:18.797936 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/c457d52c24dbccf78b0af8f0ab04285c-usr-local-share-ca-certificates\") pod \"kube-apiserver-scheduled-stop-160137\" (UID: \"c457d52c24dbccf78b0af8f0ab04285c\") " pod="kube-system/kube-apiserver-scheduled-stop-160137"
Sep 08 14:02:18 scheduled-stop-160137 kubelet[1549]: I0908 14:02:18.797952 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e23c7c4cd1f10c12a726a5f3f5563375-k8s-certs\") pod \"kube-controller-manager-scheduled-stop-160137\" (UID: \"e23c7c4cd1f10c12a726a5f3f5563375\") " pod="kube-system/kube-controller-manager-scheduled-stop-160137"
Sep 08 14:02:18 scheduled-stop-160137 kubelet[1549]: I0908 14:02:18.797973 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e23c7c4cd1f10c12a726a5f3f5563375-kubeconfig\") pod \"kube-controller-manager-scheduled-stop-160137\" (UID: \"e23c7c4cd1f10c12a726a5f3f5563375\") " pod="kube-system/kube-controller-manager-scheduled-stop-160137"
Sep 08 14:02:19 scheduled-stop-160137 kubelet[1549]: I0908 14:02:19.341939 1549 apiserver.go:52] "Watching apiserver"
Sep 08 14:02:19 scheduled-stop-160137 kubelet[1549]: I0908 14:02:19.395395 1549 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world"
Sep 08 14:02:19 scheduled-stop-160137 kubelet[1549]: I0908 14:02:19.514230 1549 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-scheduled-stop-160137"
Sep 08 14:02:19 scheduled-stop-160137 kubelet[1549]: E0908 14:02:19.531723 1549 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-apiserver-scheduled-stop-160137\" already exists" pod="kube-system/kube-apiserver-scheduled-stop-160137"
Sep 08 14:02:19 scheduled-stop-160137 kubelet[1549]: I0908 14:02:19.609320 1549 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-scheduled-stop-160137" podStartSLOduration=1.60929785 podStartE2EDuration="1.60929785s" podCreationTimestamp="2025-09-08 14:02:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 14:02:19.553276844 +0000 UTC m=+1.300418480" watchObservedRunningTime="2025-09-08 14:02:19.60929785 +0000 UTC m=+1.356439486"
Sep 08 14:02:19 scheduled-stop-160137 kubelet[1549]: I0908 14:02:19.645284 1549 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-scheduled-stop-160137" podStartSLOduration=1.6452655919999999 podStartE2EDuration="1.645265592s" podCreationTimestamp="2025-09-08 14:02:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 14:02:19.645060971 +0000 UTC m=+1.392202606" watchObservedRunningTime="2025-09-08 14:02:19.645265592 +0000 UTC m=+1.392407228"
Sep 08 14:02:19 scheduled-stop-160137 kubelet[1549]: I0908 14:02:19.645464 1549 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-scheduled-stop-160137" podStartSLOduration=2.645458981 podStartE2EDuration="2.645458981s" podCreationTimestamp="2025-09-08 14:02:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 14:02:19.609643561 +0000 UTC m=+1.356785205" watchObservedRunningTime="2025-09-08 14:02:19.645458981 +0000 UTC m=+1.392600617"
Sep 08 14:02:19 scheduled-stop-160137 kubelet[1549]: I0908 14:02:19.690632 1549 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/etcd-scheduled-stop-160137" podStartSLOduration=1.6906128169999999 podStartE2EDuration="1.690612817s" podCreationTimestamp="2025-09-08 14:02:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 14:02:19.667190015 +0000 UTC m=+1.414331716" watchObservedRunningTime="2025-09-08 14:02:19.690612817 +0000 UTC m=+1.437754461"
Sep 08 14:02:22 scheduled-stop-160137 kubelet[1549]: I0908 14:02:22.750924 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/host-path/d8ec5534-5260-45ae-a980-e0c94bbb3244-tmp\") pod \"storage-provisioner\" (UID: \"d8ec5534-5260-45ae-a980-e0c94bbb3244\") " pod="kube-system/storage-provisioner"
Sep 08 14:02:22 scheduled-stop-160137 kubelet[1549]: I0908 14:02:22.750993 1549 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgzfq\" (UniqueName: \"kubernetes.io/projected/d8ec5534-5260-45ae-a980-e0c94bbb3244-kube-api-access-kgzfq\") pod \"storage-provisioner\" (UID: \"d8ec5534-5260-45ae-a980-e0c94bbb3244\") " pod="kube-system/storage-provisioner"
Sep 08 14:02:22 scheduled-stop-160137 kubelet[1549]: E0908 14:02:22.864447 1549 projected.go:291] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found
Sep 08 14:02:22 scheduled-stop-160137 kubelet[1549]: E0908 14:02:22.864490 1549 projected.go:196] Error preparing data for projected volume kube-api-access-kgzfq for pod kube-system/storage-provisioner: configmap "kube-root-ca.crt" not found
Sep 08 14:02:22 scheduled-stop-160137 kubelet[1549]: E0908 14:02:22.867752 1549 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d8ec5534-5260-45ae-a980-e0c94bbb3244-kube-api-access-kgzfq podName:d8ec5534-5260-45ae-a980-e0c94bbb3244 nodeName:}" failed. No retries permitted until 2025-09-08 14:02:23.36740589 +0000 UTC m=+5.114547526 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-kgzfq" (UniqueName: "kubernetes.io/projected/d8ec5534-5260-45ae-a980-e0c94bbb3244-kube-api-access-kgzfq") pod "storage-provisioner" (UID: "d8ec5534-5260-45ae-a980-e0c94bbb3244") : configmap "kube-root-ca.crt" not found
-- /stdout --
helpers_test.go:262: (dbg) Run: out/minikube-linux-arm64 status --format={{.APIServer}} -p scheduled-stop-160137 -n scheduled-stop-160137
helpers_test.go:269: (dbg) Run: kubectl --context scheduled-stop-160137 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:280: non-running pods: storage-provisioner
helpers_test.go:282: ======> post-mortem[TestScheduledStopUnix]: describe non-running pods <======
helpers_test.go:285: (dbg) Run: kubectl --context scheduled-stop-160137 describe pod storage-provisioner
helpers_test.go:285: (dbg) Non-zero exit: kubectl --context scheduled-stop-160137 describe pod storage-provisioner: exit status 1 (132.241572ms)
** stderr **
Error from server (NotFound): pods "storage-provisioner" not found
** /stderr **
helpers_test.go:287: kubectl --context scheduled-stop-160137 describe pod storage-provisioner: exit status 1
helpers_test.go:175: Cleaning up "scheduled-stop-160137" profile ...
helpers_test.go:178: (dbg) Run: out/minikube-linux-arm64 delete -p scheduled-stop-160137
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p scheduled-stop-160137: (2.009927527s)
--- FAIL: TestScheduledStopUnix (33.66s)