=== RUN TestScheduledStopUnix
scheduled_stop_test.go:128: (dbg) Run: out/minikube-linux-amd64 start -p scheduled-stop-120087 --memory=3072 --driver=docker --container-runtime=containerd
E1025 09:47:37.952892 347030 cert_rotation.go:172] "Loading client cert failed" err="open /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/addons-338308/client.crt: no such file or directory" logger="tls-transport-cache.UnhandledError" key="key"
E1025 09:47:54.257812 347030 cert_rotation.go:172] "Loading client cert failed" err="open /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/functional-600067/client.crt: no such file or directory" logger="tls-transport-cache.UnhandledError" key="key"
scheduled_stop_test.go:128: (dbg) Done: out/minikube-linux-amd64 start -p scheduled-stop-120087 --memory=3072 --driver=docker --container-runtime=containerd: (21.274385272s)
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-amd64 stop -p scheduled-stop-120087 --schedule 5m
scheduled_stop_test.go:191: (dbg) Run: out/minikube-linux-amd64 status --format={{.TimeToStop}} -p scheduled-stop-120087 -n scheduled-stop-120087
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-amd64 stop -p scheduled-stop-120087 --schedule 15s
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:98: process 526910 running but should have been killed on reschedule of stop
panic.go:636: *** TestScheduledStopUnix FAILED at 2025-10-25 09:47:57.528592204 +0000 UTC m=+1690.260738220
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:223: ======> post-mortem[TestScheduledStopUnix]: network settings <======
helpers_test.go:230: HOST ENV snapshots: PROXY env: HTTP_PROXY="<empty>" HTTPS_PROXY="<empty>" NO_PROXY="<empty>"
helpers_test.go:238: ======> post-mortem[TestScheduledStopUnix]: docker inspect <======
helpers_test.go:239: (dbg) Run: docker inspect scheduled-stop-120087
helpers_test.go:243: (dbg) docker inspect scheduled-stop-120087:
-- stdout --
[
{
"Id": "720e0ff94720ddf70b93ce19d613237aeec02bfd6b5a8b5270c257d9639102c2",
"Created": "2025-10-25T09:47:40.969633662Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 524745,
"ExitCode": 0,
"Error": "",
"StartedAt": "2025-10-25T09:47:41.008339382Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:a1caeebaf98ed0136731e905a1e086f77985a42c2ebb5a7e0b3d0bd7fcbe10cc",
"ResolvConfPath": "/var/lib/docker/containers/720e0ff94720ddf70b93ce19d613237aeec02bfd6b5a8b5270c257d9639102c2/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/720e0ff94720ddf70b93ce19d613237aeec02bfd6b5a8b5270c257d9639102c2/hostname",
"HostsPath": "/var/lib/docker/containers/720e0ff94720ddf70b93ce19d613237aeec02bfd6b5a8b5270c257d9639102c2/hosts",
"LogPath": "/var/lib/docker/containers/720e0ff94720ddf70b93ce19d613237aeec02bfd6b5a8b5270c257d9639102c2/720e0ff94720ddf70b93ce19d613237aeec02bfd6b5a8b5270c257d9639102c2-json.log",
"Name": "/scheduled-stop-120087",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"scheduled-stop-120087:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "scheduled-stop-120087",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "private",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 3221225472,
"NanoCpus": 0,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 6442450944,
"MemorySwappiness": null,
"OomKillDisable": null,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"ID": "720e0ff94720ddf70b93ce19d613237aeec02bfd6b5a8b5270c257d9639102c2",
"LowerDir": "/var/lib/docker/overlay2/587f4963548ff424d19e71ceac6149233604a1ff44a83e76cf5daae466853ddd-init/diff:/var/lib/docker/overlay2/8a512527d833ed26fe5ee55317b8f867be3e2f0f5ba207cfe5a883ec8f70fe9a/diff",
"MergedDir": "/var/lib/docker/overlay2/587f4963548ff424d19e71ceac6149233604a1ff44a83e76cf5daae466853ddd/merged",
"UpperDir": "/var/lib/docker/overlay2/587f4963548ff424d19e71ceac6149233604a1ff44a83e76cf5daae466853ddd/diff",
"WorkDir": "/var/lib/docker/overlay2/587f4963548ff424d19e71ceac6149233604a1ff44a83e76cf5daae466853ddd/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "scheduled-stop-120087",
"Source": "/var/lib/docker/volumes/scheduled-stop-120087/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "scheduled-stop-120087",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "scheduled-stop-120087",
"name.minikube.sigs.k8s.io": "scheduled-stop-120087",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "6a3363930ae97cbf7165c60a8eed8b3095c1d86d89f2f1b4dbcfddcc583570c8",
"SandboxKey": "/var/run/docker/netns/6a3363930ae9",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33342"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33343"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33346"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33344"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33345"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"scheduled-stop-120087": {
"IPAMConfig": {
"IPv4Address": "192.168.76.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "d6:a1:bf:5a:55:67",
"DriverOpts": null,
"GwPriority": 0,
"NetworkID": "2dcbc9672af276fd5948b77d11ae23d89c3c523edc25503fd6d32f881ec29171",
"EndpointID": "8a82b158155b6407cbda9eebb9e4561677aa276a117da97bdf5f9f9f338408b3",
"Gateway": "192.168.76.1",
"IPAddress": "192.168.76.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"scheduled-stop-120087",
"720e0ff94720"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p scheduled-stop-120087 -n scheduled-stop-120087
helpers_test.go:252: <<< TestScheduledStopUnix FAILED: start of post-mortem logs <<<
helpers_test.go:253: ======> post-mortem[TestScheduledStopUnix]: minikube logs <======
helpers_test.go:255: (dbg) Run: out/minikube-linux-amd64 -p scheduled-stop-120087 logs -n 25
helpers_test.go:260: TestScheduledStopUnix logs:
-- stdout --
==> Audit <==
┌─────────┬─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┬───────────────────────┬─────────┬─────────┬─────────────────────┬─────────────────────┐
│ COMMAND │ ARGS │ PROFILE │ USER │ VERSION │ START TIME │ END TIME │
├─────────┼─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┼───────────────────────┼─────────┼─────────┼─────────────────────┼─────────────────────┤
│ stop │ -p multinode-817144 │ multinode-817144 │ jenkins │ v1.37.0 │ 25 Oct 25 09:42 UTC │ 25 Oct 25 09:43 UTC │
│ start │ -p multinode-817144 --wait=true -v=5 --alsologtostderr │ multinode-817144 │ jenkins │ v1.37.0 │ 25 Oct 25 09:43 UTC │ 25 Oct 25 09:44 UTC │
│ node │ list -p multinode-817144 │ multinode-817144 │ jenkins │ v1.37.0 │ 25 Oct 25 09:44 UTC │ │
│ node │ multinode-817144 node delete m03 │ multinode-817144 │ jenkins │ v1.37.0 │ 25 Oct 25 09:44 UTC │ 25 Oct 25 09:44 UTC │
│ stop │ multinode-817144 stop │ multinode-817144 │ jenkins │ v1.37.0 │ 25 Oct 25 09:44 UTC │ 25 Oct 25 09:44 UTC │
│ start │ -p multinode-817144 --wait=true -v=5 --alsologtostderr --driver=docker --container-runtime=containerd │ multinode-817144 │ jenkins │ v1.37.0 │ 25 Oct 25 09:44 UTC │ 25 Oct 25 09:45 UTC │
│ node │ list -p multinode-817144 │ multinode-817144 │ jenkins │ v1.37.0 │ 25 Oct 25 09:45 UTC │ │
│ start │ -p multinode-817144-m02 --driver=docker --container-runtime=containerd │ multinode-817144-m02 │ jenkins │ v1.37.0 │ 25 Oct 25 09:45 UTC │ │
│ start │ -p multinode-817144-m03 --driver=docker --container-runtime=containerd │ multinode-817144-m03 │ jenkins │ v1.37.0 │ 25 Oct 25 09:45 UTC │ 25 Oct 25 09:45 UTC │
│ node │ add -p multinode-817144 │ multinode-817144 │ jenkins │ v1.37.0 │ 25 Oct 25 09:45 UTC │ │
│ delete │ -p multinode-817144-m03 │ multinode-817144-m03 │ jenkins │ v1.37.0 │ 25 Oct 25 09:45 UTC │ 25 Oct 25 09:45 UTC │
│ delete │ -p multinode-817144 │ multinode-817144 │ jenkins │ v1.37.0 │ 25 Oct 25 09:45 UTC │ 25 Oct 25 09:45 UTC │
│ start │ -p test-preload-740858 --memory=3072 --alsologtostderr --wait=true --preload=false --driver=docker --container-runtime=containerd --kubernetes-version=v1.32.0 │ test-preload-740858 │ jenkins │ v1.37.0 │ 25 Oct 25 09:45 UTC │ 25 Oct 25 09:46 UTC │
│ image │ test-preload-740858 image pull gcr.io/k8s-minikube/busybox │ test-preload-740858 │ jenkins │ v1.37.0 │ 25 Oct 25 09:46 UTC │ 25 Oct 25 09:46 UTC │
│ stop │ -p test-preload-740858 │ test-preload-740858 │ jenkins │ v1.37.0 │ 25 Oct 25 09:46 UTC │ 25 Oct 25 09:46 UTC │
│ start │ -p test-preload-740858 --memory=3072 --alsologtostderr -v=1 --wait=true --driver=docker --container-runtime=containerd │ test-preload-740858 │ jenkins │ v1.37.0 │ 25 Oct 25 09:46 UTC │ 25 Oct 25 09:47 UTC │
│ image │ test-preload-740858 image list │ test-preload-740858 │ jenkins │ v1.37.0 │ 25 Oct 25 09:47 UTC │ 25 Oct 25 09:47 UTC │
│ delete │ -p test-preload-740858 │ test-preload-740858 │ jenkins │ v1.37.0 │ 25 Oct 25 09:47 UTC │ 25 Oct 25 09:47 UTC │
│ start │ -p scheduled-stop-120087 --memory=3072 --driver=docker --container-runtime=containerd │ scheduled-stop-120087 │ jenkins │ v1.37.0 │ 25 Oct 25 09:47 UTC │ 25 Oct 25 09:47 UTC │
│ stop │ -p scheduled-stop-120087 --schedule 5m │ scheduled-stop-120087 │ jenkins │ v1.37.0 │ 25 Oct 25 09:47 UTC │ │
│ stop │ -p scheduled-stop-120087 --schedule 5m │ scheduled-stop-120087 │ jenkins │ v1.37.0 │ 25 Oct 25 09:47 UTC │ │
│ stop │ -p scheduled-stop-120087 --schedule 5m │ scheduled-stop-120087 │ jenkins │ v1.37.0 │ 25 Oct 25 09:47 UTC │ │
│ stop │ -p scheduled-stop-120087 --schedule 15s │ scheduled-stop-120087 │ jenkins │ v1.37.0 │ 25 Oct 25 09:47 UTC │ │
│ stop │ -p scheduled-stop-120087 --schedule 15s │ scheduled-stop-120087 │ jenkins │ v1.37.0 │ 25 Oct 25 09:47 UTC │ │
│ stop │ -p scheduled-stop-120087 --schedule 15s │ scheduled-stop-120087 │ jenkins │ v1.37.0 │ 25 Oct 25 09:47 UTC │ │
└─────────┴─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┴───────────────────────┴─────────┴─────────┴─────────────────────┴─────────────────────┘
==> Last Start <==
Log file created at: 2025/10/25 09:47:35
Running on machine: ubuntu-20-agent-15
Binary: Built with gc go1.24.6 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I1025 09:47:35.852769 524174 out.go:360] Setting OutFile to fd 1 ...
I1025 09:47:35.853030 524174 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1025 09:47:35.853033 524174 out.go:374] Setting ErrFile to fd 2...
I1025 09:47:35.853036 524174 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1025 09:47:35.853240 524174 root.go:338] Updating PATH: /home/jenkins/minikube-integration/21767-343484/.minikube/bin
I1025 09:47:35.853708 524174 out.go:368] Setting JSON to false
I1025 09:47:35.854618 524174 start.go:131] hostinfo: {"hostname":"ubuntu-20-agent-15","uptime":5404,"bootTime":1761380252,"procs":203,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"22.04","kernelVersion":"6.8.0-1042-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I1025 09:47:35.854695 524174 start.go:141] virtualization: kvm guest
I1025 09:47:35.856778 524174 out.go:179] * [scheduled-stop-120087] minikube v1.37.0 on Ubuntu 22.04 (kvm/amd64)
I1025 09:47:35.858102 524174 notify.go:220] Checking for updates...
I1025 09:47:35.858424 524174 out.go:179] - MINIKUBE_LOCATION=21767
I1025 09:47:35.859919 524174 out.go:179] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I1025 09:47:35.861064 524174 out.go:179] - KUBECONFIG=/home/jenkins/minikube-integration/21767-343484/kubeconfig
I1025 09:47:35.862257 524174 out.go:179] - MINIKUBE_HOME=/home/jenkins/minikube-integration/21767-343484/.minikube
I1025 09:47:35.863385 524174 out.go:179] - MINIKUBE_BIN=out/minikube-linux-amd64
I1025 09:47:35.864412 524174 out.go:179] - MINIKUBE_FORCE_SYSTEMD=
I1025 09:47:35.865714 524174 driver.go:421] Setting default libvirt URI to qemu:///system
I1025 09:47:35.891774 524174 docker.go:123] docker version: linux-28.5.1:Docker Engine - Community
I1025 09:47:35.891878 524174 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1025 09:47:35.948192 524174 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:false CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:25 OomKillDisable:false NGoroutines:45 SystemTime:2025-10-25 09:47:35.937931695 +0000 UTC LoggingDriver:json-file CgroupDriver:systemd NEventsListener:0 KernelVersion:6.8.0-1042-gcp OperatingSystem:Ubuntu 22.04.5 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33652080640 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-15 Labels:[] ExperimentalBuild:false ServerVersion:28.5.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:b98a3aace656320842a23f4a392a33f46af97866 Expected:} RuncCommit:{ID:v1.3.0-0-g4ca628d1 Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin name=cgroupns] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[
map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.29.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.40.2] map[Name:model Path:/usr/libexec/docker/cli-plugins/docker-model SchemaVersion:0.1.0 ShortDescription:Docker Model Runner Vendor:Docker Inc. Version:v0.1.44] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I1025 09:47:35.948291 524174 docker.go:318] overlay module found
I1025 09:47:35.949922 524174 out.go:179] * Using the docker driver based on user configuration
I1025 09:47:35.951059 524174 start.go:305] selected driver: docker
I1025 09:47:35.951067 524174 start.go:925] validating driver "docker" against <nil>
I1025 09:47:35.951076 524174 start.go:936] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I1025 09:47:35.951669 524174 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1025 09:47:36.009977 524174 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:false CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:25 OomKillDisable:false NGoroutines:45 SystemTime:2025-10-25 09:47:35.999420338 +0000 UTC LoggingDriver:json-file CgroupDriver:systemd NEventsListener:0 KernelVersion:6.8.0-1042-gcp OperatingSystem:Ubuntu 22.04.5 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33652080640 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-15 Labels:[] ExperimentalBuild:false ServerVersion:28.5.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:b98a3aace656320842a23f4a392a33f46af97866 Expected:} RuncCommit:{ID:v1.3.0-0-g4ca628d1 Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin name=cgroupns] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[
map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.29.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.40.2] map[Name:model Path:/usr/libexec/docker/cli-plugins/docker-model SchemaVersion:0.1.0 ShortDescription:Docker Model Runner Vendor:Docker Inc. Version:v0.1.44] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I1025 09:47:36.010134 524174 start_flags.go:327] no existing cluster config was found, will generate one from the flags
I1025 09:47:36.010330 524174 start_flags.go:974] Wait components to verify : map[apiserver:true system_pods:true]
I1025 09:47:36.011904 524174 out.go:179] * Using Docker driver with root privileges
I1025 09:47:36.013060 524174 cni.go:84] Creating CNI manager for ""
I1025 09:47:36.013115 524174 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I1025 09:47:36.013122 524174 start_flags.go:336] Found "CNI" CNI - setting NetworkPlugin=cni
I1025 09:47:36.013184 524174 start.go:349] cluster config:
{Name:scheduled-stop-120087 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-120087 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local
ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I1025 09:47:36.014433 524174 out.go:179] * Starting "scheduled-stop-120087" primary control-plane node in "scheduled-stop-120087" cluster
I1025 09:47:36.015550 524174 cache.go:123] Beginning downloading kic base image for docker with containerd
I1025 09:47:36.016856 524174 out.go:179] * Pulling base image v0.0.48-1760939008-21773 ...
I1025 09:47:36.018017 524174 preload.go:183] Checking if preload exists for k8s version v1.34.1 and runtime containerd
I1025 09:47:36.018063 524174 preload.go:198] Found local preload: /home/jenkins/minikube-integration/21767-343484/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-containerd-overlay2-amd64.tar.lz4
I1025 09:47:36.018070 524174 cache.go:58] Caching tarball of preloaded images
I1025 09:47:36.018125 524174 image.go:81] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 in local docker daemon
I1025 09:47:36.018177 524174 preload.go:233] Found /home/jenkins/minikube-integration/21767-343484/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-containerd-overlay2-amd64.tar.lz4 in cache, skipping download
I1025 09:47:36.018184 524174 cache.go:61] Finished verifying existence of preloaded tar for v1.34.1 on containerd
I1025 09:47:36.018556 524174 profile.go:143] Saving config to /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/config.json ...
I1025 09:47:36.018579 524174 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/config.json: {Name:mk51be43efe67624dc6eb672aecac4248d993507 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1025 09:47:36.038800 524174 image.go:100] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 in local docker daemon, skipping pull
I1025 09:47:36.038811 524174 cache.go:147] gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 exists in daemon, skipping load
I1025 09:47:36.038829 524174 cache.go:232] Successfully downloaded all kic artifacts
I1025 09:47:36.038868 524174 start.go:360] acquireMachinesLock for scheduled-stop-120087: {Name:mk7af0f91cf4442f10d8d7622a6a14a77d3793f1 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I1025 09:47:36.038978 524174 start.go:364] duration metric: took 94.46µs to acquireMachinesLock for "scheduled-stop-120087"
I1025 09:47:36.039005 524174 start.go:93] Provisioning new machine with config: &{Name:scheduled-stop-120087 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-120087 Namespace:default APIServerHAVIP: A
PIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: Static
IP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
I1025 09:47:36.039082 524174 start.go:125] createHost starting for "" (driver="docker")
I1025 09:47:36.041554 524174 out.go:252] * Creating docker container (CPUs=2, Memory=3072MB) ...
I1025 09:47:36.041782 524174 start.go:159] libmachine.API.Create for "scheduled-stop-120087" (driver="docker")
I1025 09:47:36.041812 524174 client.go:168] LocalClient.Create starting
I1025 09:47:36.041894 524174 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21767-343484/.minikube/certs/ca.pem
I1025 09:47:36.041926 524174 main.go:141] libmachine: Decoding PEM data...
I1025 09:47:36.041939 524174 main.go:141] libmachine: Parsing certificate...
I1025 09:47:36.042000 524174 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21767-343484/.minikube/certs/cert.pem
I1025 09:47:36.042017 524174 main.go:141] libmachine: Decoding PEM data...
I1025 09:47:36.042033 524174 main.go:141] libmachine: Parsing certificate...
I1025 09:47:36.042371 524174 cli_runner.go:164] Run: docker network inspect scheduled-stop-120087 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W1025 09:47:36.060044 524174 cli_runner.go:211] docker network inspect scheduled-stop-120087 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I1025 09:47:36.060121 524174 network_create.go:284] running [docker network inspect scheduled-stop-120087] to gather additional debugging logs...
I1025 09:47:36.060136 524174 cli_runner.go:164] Run: docker network inspect scheduled-stop-120087
W1025 09:47:36.077840 524174 cli_runner.go:211] docker network inspect scheduled-stop-120087 returned with exit code 1
I1025 09:47:36.077867 524174 network_create.go:287] error running [docker network inspect scheduled-stop-120087]: docker network inspect scheduled-stop-120087: exit status 1
stdout:
[]
stderr:
Error response from daemon: network scheduled-stop-120087 not found
I1025 09:47:36.077881 524174 network_create.go:289] output of [docker network inspect scheduled-stop-120087]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network scheduled-stop-120087 not found
** /stderr **
I1025 09:47:36.078017 524174 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I1025 09:47:36.096250 524174 network.go:211] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-0432ebc9605a IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:2a:fa:02:6a:77:bd} reservation:<nil>}
I1025 09:47:36.096630 524174 network.go:211] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-d6b2dc48f0a8 IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:e6:09:0e:07:9c:1a} reservation:<nil>}
I1025 09:47:36.096963 524174 network.go:211] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-8b46e3854b9d IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:46:d3:0b:ec:ec:2e} reservation:<nil>}
I1025 09:47:36.097392 524174 network.go:206] using free private subnet 192.168.76.0/24: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc001dc15a0}
I1025 09:47:36.097414 524174 network_create.go:124] attempt to create docker network scheduled-stop-120087 192.168.76.0/24 with gateway 192.168.76.1 and MTU of 1500 ...
I1025 09:47:36.097469 524174 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.76.0/24 --gateway=192.168.76.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=scheduled-stop-120087 scheduled-stop-120087
I1025 09:47:36.156714 524174 network_create.go:108] docker network scheduled-stop-120087 192.168.76.0/24 created
I1025 09:47:36.156751 524174 kic.go:121] calculated static IP "192.168.76.2" for the "scheduled-stop-120087" container
I1025 09:47:36.156844 524174 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I1025 09:47:36.174247 524174 cli_runner.go:164] Run: docker volume create scheduled-stop-120087 --label name.minikube.sigs.k8s.io=scheduled-stop-120087 --label created_by.minikube.sigs.k8s.io=true
I1025 09:47:36.192854 524174 oci.go:103] Successfully created a docker volume scheduled-stop-120087
I1025 09:47:36.192951 524174 cli_runner.go:164] Run: docker run --rm --name scheduled-stop-120087-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-120087 --entrypoint /usr/bin/test -v scheduled-stop-120087:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 -d /var/lib
I1025 09:47:36.547436 524174 oci.go:107] Successfully prepared a docker volume scheduled-stop-120087
I1025 09:47:36.547486 524174 preload.go:183] Checking if preload exists for k8s version v1.34.1 and runtime containerd
I1025 09:47:36.547509 524174 kic.go:194] Starting extracting preloaded images to volume ...
I1025 09:47:36.547574 524174 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/21767-343484/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-containerd-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-120087:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 -I lz4 -xf /preloaded.tar -C /extractDir
I1025 09:47:40.900105 524174 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/21767-343484/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.34.1-containerd-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-120087:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 -I lz4 -xf /preloaded.tar -C /extractDir: (4.352475967s)
I1025 09:47:40.900129 524174 kic.go:203] duration metric: took 4.352616341s to extract preloaded images to volume ...
W1025 09:47:40.900223 524174 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
W1025 09:47:40.900248 524174 oci.go:252] Your kernel does not support CPU cfs period/quota or the cgroup is not mounted.
I1025 09:47:40.900280 524174 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I1025 09:47:40.953505 524174 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname scheduled-stop-120087 --name scheduled-stop-120087 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-120087 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=scheduled-stop-120087 --network scheduled-stop-120087 --ip 192.168.76.2 --volume scheduled-stop-120087:/var --security-opt apparmor=unconfined --memory=3072mb -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8
I1025 09:47:41.225222 524174 cli_runner.go:164] Run: docker container inspect scheduled-stop-120087 --format={{.State.Running}}
I1025 09:47:41.244554 524174 cli_runner.go:164] Run: docker container inspect scheduled-stop-120087 --format={{.State.Status}}
I1025 09:47:41.263945 524174 cli_runner.go:164] Run: docker exec scheduled-stop-120087 stat /var/lib/dpkg/alternatives/iptables
I1025 09:47:41.312048 524174 oci.go:144] the created container "scheduled-stop-120087" has a running status.
I1025 09:47:41.312070 524174 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/21767-343484/.minikube/machines/scheduled-stop-120087/id_rsa...
I1025 09:47:41.522877 524174 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/21767-343484/.minikube/machines/scheduled-stop-120087/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I1025 09:47:41.551705 524174 cli_runner.go:164] Run: docker container inspect scheduled-stop-120087 --format={{.State.Status}}
I1025 09:47:41.574522 524174 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I1025 09:47:41.574532 524174 kic_runner.go:114] Args: [docker exec --privileged scheduled-stop-120087 chown docker:docker /home/docker/.ssh/authorized_keys]
I1025 09:47:41.620624 524174 cli_runner.go:164] Run: docker container inspect scheduled-stop-120087 --format={{.State.Status}}
I1025 09:47:41.640588 524174 machine.go:93] provisionDockerMachine start ...
I1025 09:47:41.640713 524174 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-120087
I1025 09:47:41.660715 524174 main.go:141] libmachine: Using SSH client type: native
I1025 09:47:41.661027 524174 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x841760] 0x844460 <nil> [] 0s} 127.0.0.1 33342 <nil> <nil>}
I1025 09:47:41.661039 524174 main.go:141] libmachine: About to run SSH command:
hostname
I1025 09:47:41.801910 524174 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-120087
I1025 09:47:41.801936 524174 ubuntu.go:182] provisioning hostname "scheduled-stop-120087"
I1025 09:47:41.802018 524174 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-120087
I1025 09:47:41.821625 524174 main.go:141] libmachine: Using SSH client type: native
I1025 09:47:41.821877 524174 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x841760] 0x844460 <nil> [] 0s} 127.0.0.1 33342 <nil> <nil>}
I1025 09:47:41.821890 524174 main.go:141] libmachine: About to run SSH command:
sudo hostname scheduled-stop-120087 && echo "scheduled-stop-120087" | sudo tee /etc/hostname
I1025 09:47:41.974438 524174 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-120087
I1025 09:47:41.974500 524174 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-120087
I1025 09:47:41.993129 524174 main.go:141] libmachine: Using SSH client type: native
I1025 09:47:41.993365 524174 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x841760] 0x844460 <nil> [] 0s} 127.0.0.1 33342 <nil> <nil>}
I1025 09:47:41.993390 524174 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sscheduled-stop-120087' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 scheduled-stop-120087/g' /etc/hosts;
else
echo '127.0.1.1 scheduled-stop-120087' | sudo tee -a /etc/hosts;
fi
fi
I1025 09:47:42.133620 524174 main.go:141] libmachine: SSH cmd err, output: <nil>:
I1025 09:47:42.133663 524174 ubuntu.go:188] set auth options {CertDir:/home/jenkins/minikube-integration/21767-343484/.minikube CaCertPath:/home/jenkins/minikube-integration/21767-343484/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/21767-343484/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/21767-343484/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/21767-343484/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/21767-343484/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/21767-343484/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/21767-343484/.minikube}
I1025 09:47:42.133715 524174 ubuntu.go:190] setting up certificates
I1025 09:47:42.133730 524174 provision.go:84] configureAuth start
I1025 09:47:42.133816 524174 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-120087
I1025 09:47:42.151754 524174 provision.go:143] copyHostCerts
I1025 09:47:42.151824 524174 exec_runner.go:144] found /home/jenkins/minikube-integration/21767-343484/.minikube/ca.pem, removing ...
I1025 09:47:42.151832 524174 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21767-343484/.minikube/ca.pem
I1025 09:47:42.151897 524174 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21767-343484/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/21767-343484/.minikube/ca.pem (1078 bytes)
I1025 09:47:42.152009 524174 exec_runner.go:144] found /home/jenkins/minikube-integration/21767-343484/.minikube/cert.pem, removing ...
I1025 09:47:42.152012 524174 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21767-343484/.minikube/cert.pem
I1025 09:47:42.152037 524174 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21767-343484/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/21767-343484/.minikube/cert.pem (1123 bytes)
I1025 09:47:42.152101 524174 exec_runner.go:144] found /home/jenkins/minikube-integration/21767-343484/.minikube/key.pem, removing ...
I1025 09:47:42.152104 524174 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21767-343484/.minikube/key.pem
I1025 09:47:42.152125 524174 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21767-343484/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/21767-343484/.minikube/key.pem (1675 bytes)
I1025 09:47:42.152196 524174 provision.go:117] generating server cert: /home/jenkins/minikube-integration/21767-343484/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/21767-343484/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/21767-343484/.minikube/certs/ca-key.pem org=jenkins.scheduled-stop-120087 san=[127.0.0.1 192.168.76.2 localhost minikube scheduled-stop-120087]
I1025 09:47:42.340098 524174 provision.go:177] copyRemoteCerts
I1025 09:47:42.340147 524174 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I1025 09:47:42.340183 524174 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-120087
I1025 09:47:42.358682 524174 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33342 SSHKeyPath:/home/jenkins/minikube-integration/21767-343484/.minikube/machines/scheduled-stop-120087/id_rsa Username:docker}
I1025 09:47:42.459739 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/machines/server.pem --> /etc/docker/server.pem (1233 bytes)
I1025 09:47:42.480388 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I1025 09:47:42.498414 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I1025 09:47:42.516348 524174 provision.go:87] duration metric: took 382.604654ms to configureAuth
I1025 09:47:42.516369 524174 ubuntu.go:206] setting minikube options for container-runtime
I1025 09:47:42.516534 524174 config.go:182] Loaded profile config "scheduled-stop-120087": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.34.1
I1025 09:47:42.516539 524174 machine.go:96] duration metric: took 875.938019ms to provisionDockerMachine
I1025 09:47:42.516545 524174 client.go:171] duration metric: took 6.474728986s to LocalClient.Create
I1025 09:47:42.516565 524174 start.go:167] duration metric: took 6.474785994s to libmachine.API.Create "scheduled-stop-120087"
I1025 09:47:42.516571 524174 start.go:293] postStartSetup for "scheduled-stop-120087" (driver="docker")
I1025 09:47:42.516594 524174 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I1025 09:47:42.516650 524174 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I1025 09:47:42.516698 524174 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-120087
I1025 09:47:42.534656 524174 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33342 SSHKeyPath:/home/jenkins/minikube-integration/21767-343484/.minikube/machines/scheduled-stop-120087/id_rsa Username:docker}
I1025 09:47:42.636698 524174 ssh_runner.go:195] Run: cat /etc/os-release
I1025 09:47:42.640298 524174 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I1025 09:47:42.640319 524174 info.go:137] Remote host: Debian GNU/Linux 12 (bookworm)
I1025 09:47:42.640330 524174 filesync.go:126] Scanning /home/jenkins/minikube-integration/21767-343484/.minikube/addons for local assets ...
I1025 09:47:42.640376 524174 filesync.go:126] Scanning /home/jenkins/minikube-integration/21767-343484/.minikube/files for local assets ...
I1025 09:47:42.640450 524174 filesync.go:149] local asset: /home/jenkins/minikube-integration/21767-343484/.minikube/files/etc/ssl/certs/3470302.pem -> 3470302.pem in /etc/ssl/certs
I1025 09:47:42.640539 524174 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I1025 09:47:42.648351 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/files/etc/ssl/certs/3470302.pem --> /etc/ssl/certs/3470302.pem (1708 bytes)
I1025 09:47:42.669431 524174 start.go:296] duration metric: took 152.843002ms for postStartSetup
I1025 09:47:42.669866 524174 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-120087
I1025 09:47:42.687978 524174 profile.go:143] Saving config to /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/config.json ...
I1025 09:47:42.688248 524174 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I1025 09:47:42.688289 524174 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-120087
I1025 09:47:42.706807 524174 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33342 SSHKeyPath:/home/jenkins/minikube-integration/21767-343484/.minikube/machines/scheduled-stop-120087/id_rsa Username:docker}
I1025 09:47:42.805326 524174 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I1025 09:47:42.810024 524174 start.go:128] duration metric: took 6.770924686s to createHost
I1025 09:47:42.810044 524174 start.go:83] releasing machines lock for "scheduled-stop-120087", held for 6.771056823s
I1025 09:47:42.810136 524174 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-120087
I1025 09:47:42.827776 524174 ssh_runner.go:195] Run: cat /version.json
I1025 09:47:42.827824 524174 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-120087
I1025 09:47:42.827872 524174 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I1025 09:47:42.827939 524174 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-120087
I1025 09:47:42.846338 524174 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33342 SSHKeyPath:/home/jenkins/minikube-integration/21767-343484/.minikube/machines/scheduled-stop-120087/id_rsa Username:docker}
I1025 09:47:42.847129 524174 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33342 SSHKeyPath:/home/jenkins/minikube-integration/21767-343484/.minikube/machines/scheduled-stop-120087/id_rsa Username:docker}
I1025 09:47:42.996676 524174 ssh_runner.go:195] Run: systemctl --version
I1025 09:47:43.003521 524174 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
W1025 09:47:43.008312 524174 cni.go:209] loopback cni configuration skipped: "/etc/cni/net.d/*loopback.conf*" not found
I1025 09:47:43.008368 524174 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I1025 09:47:43.033389 524174 cni.go:262] disabled [/etc/cni/net.d/10-crio-bridge.conflist.disabled, /etc/cni/net.d/87-podman-bridge.conflist] bridge cni config(s)
I1025 09:47:43.033416 524174 start.go:495] detecting cgroup driver to use...
I1025 09:47:43.033457 524174 detect.go:190] detected "systemd" cgroup driver on host os
I1025 09:47:43.033507 524174 ssh_runner.go:195] Run: sudo systemctl stop -f crio
I1025 09:47:43.048025 524174 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I1025 09:47:43.060799 524174 docker.go:218] disabling cri-docker service (if available) ...
I1025 09:47:43.060852 524174 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
I1025 09:47:43.078175 524174 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
I1025 09:47:43.095792 524174 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
I1025 09:47:43.176094 524174 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
I1025 09:47:43.261197 524174 docker.go:234] disabling docker service ...
I1025 09:47:43.261262 524174 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
I1025 09:47:43.281128 524174 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
I1025 09:47:43.294714 524174 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
I1025 09:47:43.375068 524174 ssh_runner.go:195] Run: sudo systemctl mask docker.service
I1025 09:47:43.454547 524174 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
I1025 09:47:43.467794 524174 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I1025 09:47:43.482444 524174 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10.1"|' /etc/containerd/config.toml"
I1025 09:47:43.492807 524174 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I1025 09:47:43.502259 524174 containerd.go:146] configuring containerd to use "systemd" as cgroup driver...
I1025 09:47:43.502321 524174 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = true|g' /etc/containerd/config.toml"
I1025 09:47:43.512034 524174 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I1025 09:47:43.520993 524174 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I1025 09:47:43.529972 524174 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I1025 09:47:43.538774 524174 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I1025 09:47:43.546868 524174 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I1025 09:47:43.555680 524174 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I1025 09:47:43.564625 524174 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I1025 09:47:43.574967 524174 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I1025 09:47:43.583123 524174 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I1025 09:47:43.591170 524174 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1025 09:47:43.669813 524174 ssh_runner.go:195] Run: sudo systemctl restart containerd
I1025 09:47:43.775097 524174 start.go:542] Will wait 60s for socket path /run/containerd/containerd.sock
I1025 09:47:43.775168 524174 ssh_runner.go:195] Run: stat /run/containerd/containerd.sock
I1025 09:47:43.779567 524174 start.go:563] Will wait 60s for crictl version
I1025 09:47:43.779624 524174 ssh_runner.go:195] Run: which crictl
I1025 09:47:43.783455 524174 ssh_runner.go:195] Run: sudo /usr/local/bin/crictl version
I1025 09:47:43.809715 524174 start.go:579] Version: 0.1.0
RuntimeName: containerd
RuntimeVersion: v1.7.28
RuntimeApiVersion: v1
I1025 09:47:43.809763 524174 ssh_runner.go:195] Run: containerd --version
I1025 09:47:43.833154 524174 ssh_runner.go:195] Run: containerd --version
I1025 09:47:43.860182 524174 out.go:179] * Preparing Kubernetes v1.34.1 on containerd 1.7.28 ...
I1025 09:47:43.861515 524174 cli_runner.go:164] Run: docker network inspect scheduled-stop-120087 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I1025 09:47:43.879343 524174 ssh_runner.go:195] Run: grep 192.168.76.1 host.minikube.internal$ /etc/hosts
I1025 09:47:43.883634 524174 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.76.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I1025 09:47:43.894427 524174 kubeadm.go:883] updating cluster {Name:scheduled-stop-120087 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-120087 Namespace:default APIServerHAVIP: APIServerName:miniku
beCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSH
AuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I1025 09:47:43.894544 524174 preload.go:183] Checking if preload exists for k8s version v1.34.1 and runtime containerd
I1025 09:47:43.894605 524174 ssh_runner.go:195] Run: sudo crictl images --output json
I1025 09:47:43.919651 524174 containerd.go:627] all images are preloaded for containerd runtime.
I1025 09:47:43.919673 524174 containerd.go:534] Images already preloaded, skipping extraction
I1025 09:47:43.919731 524174 ssh_runner.go:195] Run: sudo crictl images --output json
I1025 09:47:43.945939 524174 containerd.go:627] all images are preloaded for containerd runtime.
I1025 09:47:43.945952 524174 cache_images.go:85] Images are preloaded, skipping loading
I1025 09:47:43.945959 524174 kubeadm.go:934] updating node { 192.168.76.2 8443 v1.34.1 containerd true true} ...
I1025 09:47:43.946423 524174 kubeadm.go:946] kubelet [Unit]
Wants=containerd.service
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.34.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=scheduled-stop-120087 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.76.2
[Install]
config:
{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-120087 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I1025 09:47:43.946520 524174 ssh_runner.go:195] Run: sudo crictl info
I1025 09:47:43.972720 524174 cni.go:84] Creating CNI manager for ""
I1025 09:47:43.972735 524174 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I1025 09:47:43.972753 524174 kubeadm.go:85] Using pod CIDR: 10.244.0.0/16
I1025 09:47:43.972810 524174 kubeadm.go:190] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.76.2 APIServerPort:8443 KubernetesVersion:v1.34.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:scheduled-stop-120087 NodeName:scheduled-stop-120087 DNSDomain:cluster.local CRISocket:/run/containerd/containerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.76.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.76.2 CgroupDriver:systemd ClientCAFile:/var/lib/minikube/certs/ca.crt Sta
ticPodPath:/etc/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///run/containerd/containerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I1025 09:47:43.972918 524174 kubeadm.go:196] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta4
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.76.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///run/containerd/containerd.sock
name: "scheduled-stop-120087"
kubeletExtraArgs:
- name: "node-ip"
value: "192.168.76.2"
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta4
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.76.2"]
extraArgs:
- name: "enable-admission-plugins"
value: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
- name: "allocate-node-cidrs"
value: "true"
- name: "leader-elect"
value: "false"
scheduler:
extraArgs:
- name: "leader-elect"
value: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
kubernetesVersion: v1.34.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: systemd
containerRuntimeEndpoint: unix:///run/containerd/containerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I1025 09:47:43.972974 524174 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.34.1
I1025 09:47:43.981131 524174 binaries.go:44] Found k8s binaries, skipping transfer
I1025 09:47:43.981215 524174 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I1025 09:47:43.988948 524174 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (325 bytes)
I1025 09:47:44.001553 524174 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I1025 09:47:44.017037 524174 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2233 bytes)
I1025 09:47:44.029962 524174 ssh_runner.go:195] Run: grep 192.168.76.2 control-plane.minikube.internal$ /etc/hosts
I1025 09:47:44.033736 524174 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.76.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I1025 09:47:44.043794 524174 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1025 09:47:44.126374 524174 ssh_runner.go:195] Run: sudo systemctl start kubelet
I1025 09:47:44.149119 524174 certs.go:69] Setting up /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087 for IP: 192.168.76.2
I1025 09:47:44.149134 524174 certs.go:195] generating shared ca certs ...
I1025 09:47:44.149151 524174 certs.go:227] acquiring lock for ca certs: {Name:mkdaf0ad503443626f2647e8329a304ba0044a17 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1025 09:47:44.149302 524174 certs.go:236] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/21767-343484/.minikube/ca.key
I1025 09:47:44.149356 524174 certs.go:236] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/21767-343484/.minikube/proxy-client-ca.key
I1025 09:47:44.149364 524174 certs.go:257] generating profile certs ...
I1025 09:47:44.149417 524174 certs.go:364] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/client.key
I1025 09:47:44.149433 524174 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/client.crt with IP's: []
I1025 09:47:44.308915 524174 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/client.crt ...
I1025 09:47:44.308935 524174 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/client.crt: {Name:mk54200cbaa058223e0a0a27303c8c3bebedda9b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1025 09:47:44.309126 524174 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/client.key ...
I1025 09:47:44.309135 524174 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/client.key: {Name:mka64c7dd47839cf8736edf4631ebf654f9234ff Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1025 09:47:44.309214 524174 certs.go:364] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/apiserver.key.4ec89f29
I1025 09:47:44.309226 524174 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/apiserver.crt.4ec89f29 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.76.2]
I1025 09:47:44.483107 524174 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/apiserver.crt.4ec89f29 ...
I1025 09:47:44.483124 524174 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/apiserver.crt.4ec89f29: {Name:mk3d900dd4b4ae18a7631da34a17e0d97fc36140 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1025 09:47:44.483303 524174 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/apiserver.key.4ec89f29 ...
I1025 09:47:44.483311 524174 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/apiserver.key.4ec89f29: {Name:mk3e88e917f2fc3754f17907235ddc6f8444756c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1025 09:47:44.483396 524174 certs.go:382] copying /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/apiserver.crt.4ec89f29 -> /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/apiserver.crt
I1025 09:47:44.483471 524174 certs.go:386] copying /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/apiserver.key.4ec89f29 -> /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/apiserver.key
I1025 09:47:44.483532 524174 certs.go:364] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/proxy-client.key
I1025 09:47:44.483544 524174 crypto.go:68] Generating cert /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/proxy-client.crt with IP's: []
I1025 09:47:44.865047 524174 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/proxy-client.crt ...
I1025 09:47:44.865067 524174 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/proxy-client.crt: {Name:mkd1729cafa2112115836f7c29dc1615fcc6af4f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1025 09:47:44.865243 524174 crypto.go:164] Writing key to /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/proxy-client.key ...
I1025 09:47:44.865254 524174 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/proxy-client.key: {Name:mk7bf91e41ee5e0f870a74ab47c79d1b3a42d9a1 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1025 09:47:44.865442 524174 certs.go:484] found cert: /home/jenkins/minikube-integration/21767-343484/.minikube/certs/347030.pem (1338 bytes)
W1025 09:47:44.865472 524174 certs.go:480] ignoring /home/jenkins/minikube-integration/21767-343484/.minikube/certs/347030_empty.pem, impossibly tiny 0 bytes
I1025 09:47:44.865479 524174 certs.go:484] found cert: /home/jenkins/minikube-integration/21767-343484/.minikube/certs/ca-key.pem (1675 bytes)
I1025 09:47:44.865498 524174 certs.go:484] found cert: /home/jenkins/minikube-integration/21767-343484/.minikube/certs/ca.pem (1078 bytes)
I1025 09:47:44.865516 524174 certs.go:484] found cert: /home/jenkins/minikube-integration/21767-343484/.minikube/certs/cert.pem (1123 bytes)
I1025 09:47:44.865532 524174 certs.go:484] found cert: /home/jenkins/minikube-integration/21767-343484/.minikube/certs/key.pem (1675 bytes)
I1025 09:47:44.865570 524174 certs.go:484] found cert: /home/jenkins/minikube-integration/21767-343484/.minikube/files/etc/ssl/certs/3470302.pem (1708 bytes)
I1025 09:47:44.866164 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I1025 09:47:44.885294 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I1025 09:47:44.903371 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I1025 09:47:44.921043 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I1025 09:47:44.939202 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1432 bytes)
I1025 09:47:44.957521 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I1025 09:47:44.975445 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I1025 09:47:44.993255 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/profiles/scheduled-stop-120087/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I1025 09:47:45.012155 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I1025 09:47:45.034022 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/certs/347030.pem --> /usr/share/ca-certificates/347030.pem (1338 bytes)
I1025 09:47:45.053849 524174 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21767-343484/.minikube/files/etc/ssl/certs/3470302.pem --> /usr/share/ca-certificates/3470302.pem (1708 bytes)
I1025 09:47:45.071400 524174 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I1025 09:47:45.084165 524174 ssh_runner.go:195] Run: openssl version
I1025 09:47:45.090828 524174 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/347030.pem && ln -fs /usr/share/ca-certificates/347030.pem /etc/ssl/certs/347030.pem"
I1025 09:47:45.100212 524174 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/347030.pem
I1025 09:47:45.104188 524174 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 Oct 25 09:26 /usr/share/ca-certificates/347030.pem
I1025 09:47:45.104227 524174 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/347030.pem
I1025 09:47:45.138568 524174 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/347030.pem /etc/ssl/certs/51391683.0"
I1025 09:47:45.147553 524174 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/3470302.pem && ln -fs /usr/share/ca-certificates/3470302.pem /etc/ssl/certs/3470302.pem"
I1025 09:47:45.156051 524174 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/3470302.pem
I1025 09:47:45.159653 524174 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 Oct 25 09:26 /usr/share/ca-certificates/3470302.pem
I1025 09:47:45.159714 524174 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/3470302.pem
I1025 09:47:45.193975 524174 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/3470302.pem /etc/ssl/certs/3ec20f2e.0"
I1025 09:47:45.202810 524174 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I1025 09:47:45.211369 524174 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I1025 09:47:45.215057 524174 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Oct 25 09:20 /usr/share/ca-certificates/minikubeCA.pem
I1025 09:47:45.215098 524174 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I1025 09:47:45.248932 524174 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I1025 09:47:45.257895 524174 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I1025 09:47:45.261814 524174 certs.go:400] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I1025 09:47:45.261864 524174 kubeadm.go:400] StartCluster: {Name:scheduled-stop-120087 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1760939008-21773@sha256:d8d8a3f29f027433bea12764bddd1aa26c7ad9bb912e016c1bc51278db1343d8 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.34.1 ClusterName:scheduled-stop-120087 Namespace:default APIServerHAVIP: APIServerName:minikubeC
A APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAut
hSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I1025 09:47:45.261926 524174 cri.go:54] listing CRI containers in root /run/containerd/runc/k8s.io: {State:paused Name: Namespaces:[kube-system]}
I1025 09:47:45.261967 524174 ssh_runner.go:195] Run: sudo -s eval "crictl ps -a --quiet --label io.kubernetes.pod.namespace=kube-system"
I1025 09:47:45.289473 524174 cri.go:89] found id: ""
I1025 09:47:45.289534 524174 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I1025 09:47:45.297551 524174 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I1025 09:47:45.305728 524174 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I1025 09:47:45.305776 524174 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I1025 09:47:45.314219 524174 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I1025 09:47:45.314230 524174 kubeadm.go:157] found existing configuration files:
I1025 09:47:45.314275 524174 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I1025 09:47:45.322021 524174 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I1025 09:47:45.322080 524174 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I1025 09:47:45.329364 524174 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I1025 09:47:45.336767 524174 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I1025 09:47:45.336811 524174 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I1025 09:47:45.344020 524174 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I1025 09:47:45.351389 524174 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I1025 09:47:45.351435 524174 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I1025 09:47:45.358585 524174 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I1025 09:47:45.366130 524174 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I1025 09:47:45.366173 524174 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I1025 09:47:45.373551 524174 ssh_runner.go:286] Start: sudo /bin/bash -c "env PATH="/var/lib/minikube/binaries/v1.34.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I1025 09:47:45.431098 524174 kubeadm.go:318] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/6.8.0-1042-gcp\n", err: exit status 1
I1025 09:47:45.487873 524174 kubeadm.go:318] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I1025 09:47:55.926973 524174 kubeadm.go:318] [init] Using Kubernetes version: v1.34.1
I1025 09:47:55.927038 524174 kubeadm.go:318] [preflight] Running pre-flight checks
I1025 09:47:55.927154 524174 kubeadm.go:318] [preflight] The system verification failed. Printing the output from the verification:
I1025 09:47:55.927227 524174 kubeadm.go:318] [0;37mKERNEL_VERSION[0m: [0;32m6.8.0-1042-gcp[0m
I1025 09:47:55.927266 524174 kubeadm.go:318] [0;37mOS[0m: [0;32mLinux[0m
I1025 09:47:55.927334 524174 kubeadm.go:318] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I1025 09:47:55.927386 524174 kubeadm.go:318] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I1025 09:47:55.927455 524174 kubeadm.go:318] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I1025 09:47:55.927498 524174 kubeadm.go:318] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I1025 09:47:55.927534 524174 kubeadm.go:318] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I1025 09:47:55.927599 524174 kubeadm.go:318] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I1025 09:47:55.927706 524174 kubeadm.go:318] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I1025 09:47:55.927743 524174 kubeadm.go:318] [0;37mCGROUPS_IO[0m: [0;32menabled[0m
I1025 09:47:55.927803 524174 kubeadm.go:318] [preflight] Pulling images required for setting up a Kubernetes cluster
I1025 09:47:55.927876 524174 kubeadm.go:318] [preflight] This might take a minute or two, depending on the speed of your internet connection
I1025 09:47:55.927955 524174 kubeadm.go:318] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I1025 09:47:55.928003 524174 kubeadm.go:318] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I1025 09:47:55.929668 524174 out.go:252] - Generating certificates and keys ...
I1025 09:47:55.929755 524174 kubeadm.go:318] [certs] Using existing ca certificate authority
I1025 09:47:55.929844 524174 kubeadm.go:318] [certs] Using existing apiserver certificate and key on disk
I1025 09:47:55.929947 524174 kubeadm.go:318] [certs] Generating "apiserver-kubelet-client" certificate and key
I1025 09:47:55.930031 524174 kubeadm.go:318] [certs] Generating "front-proxy-ca" certificate and key
I1025 09:47:55.930116 524174 kubeadm.go:318] [certs] Generating "front-proxy-client" certificate and key
I1025 09:47:55.930188 524174 kubeadm.go:318] [certs] Generating "etcd/ca" certificate and key
I1025 09:47:55.930236 524174 kubeadm.go:318] [certs] Generating "etcd/server" certificate and key
I1025 09:47:55.930352 524174 kubeadm.go:318] [certs] etcd/server serving cert is signed for DNS names [localhost scheduled-stop-120087] and IPs [192.168.76.2 127.0.0.1 ::1]
I1025 09:47:55.930406 524174 kubeadm.go:318] [certs] Generating "etcd/peer" certificate and key
I1025 09:47:55.930511 524174 kubeadm.go:318] [certs] etcd/peer serving cert is signed for DNS names [localhost scheduled-stop-120087] and IPs [192.168.76.2 127.0.0.1 ::1]
I1025 09:47:55.930577 524174 kubeadm.go:318] [certs] Generating "etcd/healthcheck-client" certificate and key
I1025 09:47:55.930651 524174 kubeadm.go:318] [certs] Generating "apiserver-etcd-client" certificate and key
I1025 09:47:55.930720 524174 kubeadm.go:318] [certs] Generating "sa" key and public key
I1025 09:47:55.930800 524174 kubeadm.go:318] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I1025 09:47:55.930848 524174 kubeadm.go:318] [kubeconfig] Writing "admin.conf" kubeconfig file
I1025 09:47:55.930908 524174 kubeadm.go:318] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I1025 09:47:55.930971 524174 kubeadm.go:318] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I1025 09:47:55.931063 524174 kubeadm.go:318] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I1025 09:47:55.931132 524174 kubeadm.go:318] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I1025 09:47:55.931231 524174 kubeadm.go:318] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I1025 09:47:55.931291 524174 kubeadm.go:318] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I1025 09:47:55.932781 524174 out.go:252] - Booting up control plane ...
I1025 09:47:55.932851 524174 kubeadm.go:318] [control-plane] Creating static Pod manifest for "kube-apiserver"
I1025 09:47:55.932935 524174 kubeadm.go:318] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I1025 09:47:55.932995 524174 kubeadm.go:318] [control-plane] Creating static Pod manifest for "kube-scheduler"
I1025 09:47:55.933081 524174 kubeadm.go:318] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I1025 09:47:55.933158 524174 kubeadm.go:318] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/instance-config.yaml"
I1025 09:47:55.933242 524174 kubeadm.go:318] [patches] Applied patch of type "application/strategic-merge-patch+json" to target "kubeletconfiguration"
I1025 09:47:55.933309 524174 kubeadm.go:318] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I1025 09:47:55.933353 524174 kubeadm.go:318] [kubelet-start] Starting the kubelet
I1025 09:47:55.933458 524174 kubeadm.go:318] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I1025 09:47:55.933559 524174 kubeadm.go:318] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I1025 09:47:55.933610 524174 kubeadm.go:318] [kubelet-check] The kubelet is healthy after 1.001037295s
I1025 09:47:55.933726 524174 kubeadm.go:318] [control-plane-check] Waiting for healthy control plane components. This can take up to 4m0s
I1025 09:47:55.933793 524174 kubeadm.go:318] [control-plane-check] Checking kube-apiserver at https://192.168.76.2:8443/livez
I1025 09:47:55.933866 524174 kubeadm.go:318] [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz
I1025 09:47:55.933935 524174 kubeadm.go:318] [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez
I1025 09:47:55.933987 524174 kubeadm.go:318] [control-plane-check] kube-scheduler is healthy after 1.729465398s
I1025 09:47:55.934045 524174 kubeadm.go:318] [control-plane-check] kube-controller-manager is healthy after 2.154618386s
I1025 09:47:55.934096 524174 kubeadm.go:318] [control-plane-check] kube-apiserver is healthy after 3.501657466s
I1025 09:47:55.934179 524174 kubeadm.go:318] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I1025 09:47:55.934293 524174 kubeadm.go:318] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I1025 09:47:55.934343 524174 kubeadm.go:318] [upload-certs] Skipping phase. Please see --upload-certs
I1025 09:47:55.934503 524174 kubeadm.go:318] [mark-control-plane] Marking the node scheduled-stop-120087 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I1025 09:47:55.934582 524174 kubeadm.go:318] [bootstrap-token] Using token: wcr1ag.syxfh3pfdttfy06l
I1025 09:47:55.935853 524174 out.go:252] - Configuring RBAC rules ...
I1025 09:47:55.935928 524174 kubeadm.go:318] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I1025 09:47:55.935991 524174 kubeadm.go:318] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I1025 09:47:55.936105 524174 kubeadm.go:318] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I1025 09:47:55.936200 524174 kubeadm.go:318] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I1025 09:47:55.936303 524174 kubeadm.go:318] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I1025 09:47:55.936415 524174 kubeadm.go:318] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I1025 09:47:55.936510 524174 kubeadm.go:318] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I1025 09:47:55.936557 524174 kubeadm.go:318] [addons] Applied essential addon: CoreDNS
I1025 09:47:55.936592 524174 kubeadm.go:318] [addons] Applied essential addon: kube-proxy
I1025 09:47:55.936595 524174 kubeadm.go:318]
I1025 09:47:55.936650 524174 kubeadm.go:318] Your Kubernetes control-plane has initialized successfully!
I1025 09:47:55.936653 524174 kubeadm.go:318]
I1025 09:47:55.936714 524174 kubeadm.go:318] To start using your cluster, you need to run the following as a regular user:
I1025 09:47:55.936716 524174 kubeadm.go:318]
I1025 09:47:55.936734 524174 kubeadm.go:318] mkdir -p $HOME/.kube
I1025 09:47:55.936782 524174 kubeadm.go:318] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I1025 09:47:55.936820 524174 kubeadm.go:318] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I1025 09:47:55.936825 524174 kubeadm.go:318]
I1025 09:47:55.936866 524174 kubeadm.go:318] Alternatively, if you are the root user, you can run:
I1025 09:47:55.936868 524174 kubeadm.go:318]
I1025 09:47:55.936911 524174 kubeadm.go:318] export KUBECONFIG=/etc/kubernetes/admin.conf
I1025 09:47:55.936914 524174 kubeadm.go:318]
I1025 09:47:55.936954 524174 kubeadm.go:318] You should now deploy a pod network to the cluster.
I1025 09:47:55.937025 524174 kubeadm.go:318] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I1025 09:47:55.937092 524174 kubeadm.go:318] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I1025 09:47:55.937094 524174 kubeadm.go:318]
I1025 09:47:55.937171 524174 kubeadm.go:318] You can now join any number of control-plane nodes by copying certificate authorities
I1025 09:47:55.937232 524174 kubeadm.go:318] and service account keys on each node and then running the following as root:
I1025 09:47:55.937234 524174 kubeadm.go:318]
I1025 09:47:55.937298 524174 kubeadm.go:318] kubeadm join control-plane.minikube.internal:8443 --token wcr1ag.syxfh3pfdttfy06l \
I1025 09:47:55.937388 524174 kubeadm.go:318] --discovery-token-ca-cert-hash sha256:29384c97d7e223ded7031e70f1d83bb89a1d61c063a7053804f0c4ff5b55a7e1 \
I1025 09:47:55.937405 524174 kubeadm.go:318] --control-plane
I1025 09:47:55.937415 524174 kubeadm.go:318]
I1025 09:47:55.937484 524174 kubeadm.go:318] Then you can join any number of worker nodes by running the following on each as root:
I1025 09:47:55.937486 524174 kubeadm.go:318]
I1025 09:47:55.937552 524174 kubeadm.go:318] kubeadm join control-plane.minikube.internal:8443 --token wcr1ag.syxfh3pfdttfy06l \
I1025 09:47:55.937655 524174 kubeadm.go:318] --discovery-token-ca-cert-hash sha256:29384c97d7e223ded7031e70f1d83bb89a1d61c063a7053804f0c4ff5b55a7e1
I1025 09:47:55.937691 524174 cni.go:84] Creating CNI manager for ""
I1025 09:47:55.937698 524174 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I1025 09:47:55.939098 524174 out.go:179] * Configuring CNI (Container Networking Interface) ...
I1025 09:47:55.940193 524174 ssh_runner.go:195] Run: stat /opt/cni/bin/portmap
I1025 09:47:55.944744 524174 cni.go:182] applying CNI manifest using /var/lib/minikube/binaries/v1.34.1/kubectl ...
I1025 09:47:55.944752 524174 ssh_runner.go:362] scp memory --> /var/tmp/minikube/cni.yaml (2601 bytes)
I1025 09:47:55.957886 524174 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.1/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml
I1025 09:47:56.166718 524174 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I1025 09:47:56.166780 524174 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I1025 09:47:56.166826 524174 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes scheduled-stop-120087 minikube.k8s.io/updated_at=2025_10_25T09_47_56_0700 minikube.k8s.io/version=v1.37.0 minikube.k8s.io/commit=6017293569ff48e99407bb5ade8e9ba1a7a0c689 minikube.k8s.io/name=scheduled-stop-120087 minikube.k8s.io/primary=true
I1025 09:47:56.252136 524174 kubeadm.go:1113] duration metric: took 85.417415ms to wait for elevateKubeSystemPrivileges
I1025 09:47:56.252220 524174 ops.go:34] apiserver oom_adj: -16
I1025 09:47:56.252237 524174 kubeadm.go:402] duration metric: took 10.990376256s to StartCluster
I1025 09:47:56.252261 524174 settings.go:142] acquiring lock: {Name:mk8537c08ddf1c1990512eb704707568b29eef37 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1025 09:47:56.252343 524174 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/21767-343484/kubeconfig
I1025 09:47:56.253232 524174 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21767-343484/kubeconfig: {Name:mk017bd58ff76dd9eada6ffa30752f95af6119b1 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1025 09:47:56.253458 524174 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I1025 09:47:56.253476 524174 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.34.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
I1025 09:47:56.253542 524174 addons.go:511] enable addons start: toEnable=map[ambassador:false amd-gpu-device-plugin:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubetail:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-rancher:false volcano:false volumesnapshots:false yakd:false]
I1025 09:47:56.253630 524174 addons.go:69] Setting storage-provisioner=true in profile "scheduled-stop-120087"
I1025 09:47:56.253659 524174 addons.go:238] Setting addon storage-provisioner=true in "scheduled-stop-120087"
I1025 09:47:56.253669 524174 addons.go:69] Setting default-storageclass=true in profile "scheduled-stop-120087"
I1025 09:47:56.253683 524174 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "scheduled-stop-120087"
I1025 09:47:56.253690 524174 host.go:66] Checking if "scheduled-stop-120087" exists ...
I1025 09:47:56.253737 524174 config.go:182] Loaded profile config "scheduled-stop-120087": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.34.1
I1025 09:47:56.254081 524174 cli_runner.go:164] Run: docker container inspect scheduled-stop-120087 --format={{.State.Status}}
I1025 09:47:56.254225 524174 cli_runner.go:164] Run: docker container inspect scheduled-stop-120087 --format={{.State.Status}}
I1025 09:47:56.254934 524174 out.go:179] * Verifying Kubernetes components...
I1025 09:47:56.256235 524174 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1025 09:47:56.276133 524174 addons.go:238] Setting addon default-storageclass=true in "scheduled-stop-120087"
I1025 09:47:56.276171 524174 host.go:66] Checking if "scheduled-stop-120087" exists ...
I1025 09:47:56.276825 524174 cli_runner.go:164] Run: docker container inspect scheduled-stop-120087 --format={{.State.Status}}
I1025 09:47:56.277080 524174 out.go:179] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I1025 09:47:56.280829 524174 addons.go:435] installing /etc/kubernetes/addons/storage-provisioner.yaml
I1025 09:47:56.280840 524174 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I1025 09:47:56.280905 524174 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-120087
I1025 09:47:56.305813 524174 addons.go:435] installing /etc/kubernetes/addons/storageclass.yaml
I1025 09:47:56.305828 524174 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I1025 09:47:56.305893 524174 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-120087
I1025 09:47:56.307828 524174 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33342 SSHKeyPath:/home/jenkins/minikube-integration/21767-343484/.minikube/machines/scheduled-stop-120087/id_rsa Username:docker}
I1025 09:47:56.327102 524174 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33342 SSHKeyPath:/home/jenkins/minikube-integration/21767-343484/.minikube/machines/scheduled-stop-120087/id_rsa Username:docker}
I1025 09:47:56.346196 524174 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.76.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.34.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I1025 09:47:56.392454 524174 ssh_runner.go:195] Run: sudo systemctl start kubelet
I1025 09:47:56.423581 524174 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.34.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I1025 09:47:56.441541 524174 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.34.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I1025 09:47:56.503312 524174 start.go:976] {"host.minikube.internal": 192.168.76.1} host record injected into CoreDNS's ConfigMap
I1025 09:47:56.504238 524174 api_server.go:52] waiting for apiserver process to appear ...
I1025 09:47:56.504293 524174 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I1025 09:47:56.716547 524174 api_server.go:72] duration metric: took 463.039314ms to wait for apiserver process to appear ...
I1025 09:47:56.716562 524174 api_server.go:88] waiting for apiserver healthz status ...
I1025 09:47:56.716583 524174 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
I1025 09:47:56.722729 524174 api_server.go:279] https://192.168.76.2:8443/healthz returned 200:
ok
I1025 09:47:56.723457 524174 out.go:179] * Enabled addons: storage-provisioner, default-storageclass
I1025 09:47:56.723761 524174 api_server.go:141] control plane version: v1.34.1
I1025 09:47:56.723778 524174 api_server.go:131] duration metric: took 7.210236ms to wait for apiserver health ...
I1025 09:47:56.723787 524174 system_pods.go:43] waiting for kube-system pods to appear ...
I1025 09:47:56.724888 524174 addons.go:514] duration metric: took 471.344246ms for enable addons: enabled=[storage-provisioner default-storageclass]
I1025 09:47:56.726397 524174 system_pods.go:59] 5 kube-system pods found
I1025 09:47:56.726420 524174 system_pods.go:61] "etcd-scheduled-stop-120087" [a4a9b59c-dae0-4779-abf4-9ec7d554e76c] Running / Ready:ContainersNotReady (containers with unready status: [etcd]) / ContainersReady:ContainersNotReady (containers with unready status: [etcd])
I1025 09:47:56.726430 524174 system_pods.go:61] "kube-apiserver-scheduled-stop-120087" [87e42cdc-9a7f-4562-a137-25c3e6dc6804] Running / Ready:ContainersNotReady (containers with unready status: [kube-apiserver]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-apiserver])
I1025 09:47:56.726439 524174 system_pods.go:61] "kube-controller-manager-scheduled-stop-120087" [5fad7aeb-a9e8-4da6-a194-e93572ec5709] Running / Ready:ContainersNotReady (containers with unready status: [kube-controller-manager]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-controller-manager])
I1025 09:47:56.726454 524174 system_pods.go:61] "kube-scheduler-scheduled-stop-120087" [11a58581-76ca-45cc-b02e-b2cb8e3454a4] Running / Ready:ContainersNotReady (containers with unready status: [kube-scheduler]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-scheduler])
I1025 09:47:56.726459 524174 system_pods.go:61] "storage-provisioner" [f43dc1bb-6e0c-4b8e-8568-8163376b2723] Pending
I1025 09:47:56.726467 524174 system_pods.go:74] duration metric: took 2.67389ms to wait for pod list to return data ...
I1025 09:47:56.726478 524174 kubeadm.go:586] duration metric: took 472.97656ms to wait for: map[apiserver:true system_pods:true]
I1025 09:47:56.726491 524174 node_conditions.go:102] verifying NodePressure condition ...
I1025 09:47:56.728612 524174 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I1025 09:47:56.728624 524174 node_conditions.go:123] node cpu capacity is 8
I1025 09:47:56.728635 524174 node_conditions.go:105] duration metric: took 2.14026ms to run NodePressure ...
I1025 09:47:56.728665 524174 start.go:241] waiting for startup goroutines ...
I1025 09:47:57.007344 524174 kapi.go:214] "coredns" deployment in "kube-system" namespace and "scheduled-stop-120087" context rescaled to 1 replicas
I1025 09:47:57.007371 524174 start.go:246] waiting for cluster config update ...
I1025 09:47:57.007381 524174 start.go:255] writing updated cluster config ...
I1025 09:47:57.007684 524174 ssh_runner.go:195] Run: rm -f paused
I1025 09:47:57.057011 524174 start.go:624] kubectl: 1.34.1, cluster: 1.34.1 (minor skew: 0)
I1025 09:47:57.058725 524174 out.go:179] * Done! kubectl is now configured to use "scheduled-stop-120087" cluster and "default" namespace by default
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD NAMESPACE
d0ba4a6458049 5f1f5298c888d 7 seconds ago Running etcd 0 19f182a3f9533 etcd-scheduled-stop-120087 kube-system
694ed725a6cf5 7dd6aaa1717ab 7 seconds ago Running kube-scheduler 0 9b0d2e5822ca8 kube-scheduler-scheduled-stop-120087 kube-system
25223f701eea7 c3994bc696102 7 seconds ago Running kube-apiserver 0 5612296d479b9 kube-apiserver-scheduled-stop-120087 kube-system
d58e1419868e0 c80c8dbafe7dd 7 seconds ago Running kube-controller-manager 0 da474ed027412 kube-controller-manager-scheduled-stop-120087 kube-system
==> containerd <==
Oct 25 09:47:43 scheduled-stop-120087 systemd[1]: Started containerd.service - containerd container runtime.
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.052405776Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-scheduled-stop-120087,Uid:efcedea45f933d5b0d8312714f2ac611,Namespace:kube-system,Attempt:0,}"
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.053545090Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-scheduled-stop-120087,Uid:5f8254f424014ac94cf68614848c604c,Namespace:kube-system,Attempt:0,}"
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.067242784Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-scheduled-stop-120087,Uid:c5c25dc7dc33c5716c30ee4a4da1e1bf,Namespace:kube-system,Attempt:0,}"
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.069866645Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:etcd-scheduled-stop-120087,Uid:42393771443bf7e01fd7a1f384763bc9,Namespace:kube-system,Attempt:0,}"
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.151006787Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-scheduled-stop-120087,Uid:5f8254f424014ac94cf68614848c604c,Namespace:kube-system,Attempt:0,} returns sandbox id \"da474ed027412a7e32cd0a10415c14010082a5d5b16c1d2319df5cbf541bcfc2\""
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.154599386Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-scheduled-stop-120087,Uid:efcedea45f933d5b0d8312714f2ac611,Namespace:kube-system,Attempt:0,} returns sandbox id \"5612296d479b9087db7a4423cbcaade2a8e02aa57b0db2bccd617c47cdf3e41d\""
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.154923664Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-scheduled-stop-120087,Uid:c5c25dc7dc33c5716c30ee4a4da1e1bf,Namespace:kube-system,Attempt:0,} returns sandbox id \"9b0d2e5822ca82b6158b570e3c216e9abfbea129935b0b7f95663656a9298ad7\""
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.158038415Z" level=info msg="CreateContainer within sandbox \"da474ed027412a7e32cd0a10415c14010082a5d5b16c1d2319df5cbf541bcfc2\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}"
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.159252664Z" level=info msg="CreateContainer within sandbox \"5612296d479b9087db7a4423cbcaade2a8e02aa57b0db2bccd617c47cdf3e41d\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}"
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.159615723Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:etcd-scheduled-stop-120087,Uid:42393771443bf7e01fd7a1f384763bc9,Namespace:kube-system,Attempt:0,} returns sandbox id \"19f182a3f9533eefbfeba491810d8deebf9c87c81b2714c186d8ce85dab43dee\""
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.161377246Z" level=info msg="CreateContainer within sandbox \"9b0d2e5822ca82b6158b570e3c216e9abfbea129935b0b7f95663656a9298ad7\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}"
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.163514134Z" level=info msg="CreateContainer within sandbox \"19f182a3f9533eefbfeba491810d8deebf9c87c81b2714c186d8ce85dab43dee\" for container &ContainerMetadata{Name:etcd,Attempt:0,}"
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.176208156Z" level=info msg="CreateContainer within sandbox \"da474ed027412a7e32cd0a10415c14010082a5d5b16c1d2319df5cbf541bcfc2\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"d58e1419868e0a837d3318dcf6e91804a4e9025f3562bc311ce1efe6c3907f79\""
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.177438845Z" level=info msg="StartContainer for \"d58e1419868e0a837d3318dcf6e91804a4e9025f3562bc311ce1efe6c3907f79\""
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.180118083Z" level=info msg="CreateContainer within sandbox \"5612296d479b9087db7a4423cbcaade2a8e02aa57b0db2bccd617c47cdf3e41d\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"25223f701eea7df174b2ba1af625a6aac7d9b7e5ba9982b5de7a4d76a6a31ebc\""
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.180683161Z" level=info msg="StartContainer for \"25223f701eea7df174b2ba1af625a6aac7d9b7e5ba9982b5de7a4d76a6a31ebc\""
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.192775179Z" level=info msg="CreateContainer within sandbox \"19f182a3f9533eefbfeba491810d8deebf9c87c81b2714c186d8ce85dab43dee\" for &ContainerMetadata{Name:etcd,Attempt:0,} returns container id \"d0ba4a64580495daa8fee690bfff70f188eb4f1f24e01480b2949a8c86cbf0a4\""
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.192844943Z" level=info msg="CreateContainer within sandbox \"9b0d2e5822ca82b6158b570e3c216e9abfbea129935b0b7f95663656a9298ad7\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"694ed725a6cf5d351aee170316fe7b89a81dae03cc239694ff1e3515ea5586fe\""
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.193414469Z" level=info msg="StartContainer for \"d0ba4a64580495daa8fee690bfff70f188eb4f1f24e01480b2949a8c86cbf0a4\""
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.193456652Z" level=info msg="StartContainer for \"694ed725a6cf5d351aee170316fe7b89a81dae03cc239694ff1e3515ea5586fe\""
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.277456127Z" level=info msg="StartContainer for \"694ed725a6cf5d351aee170316fe7b89a81dae03cc239694ff1e3515ea5586fe\" returns successfully"
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.277580955Z" level=info msg="StartContainer for \"d58e1419868e0a837d3318dcf6e91804a4e9025f3562bc311ce1efe6c3907f79\" returns successfully"
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.277785487Z" level=info msg="StartContainer for \"25223f701eea7df174b2ba1af625a6aac7d9b7e5ba9982b5de7a4d76a6a31ebc\" returns successfully"
Oct 25 09:47:51 scheduled-stop-120087 containerd[670]: time="2025-10-25T09:47:51.277955959Z" level=info msg="StartContainer for \"d0ba4a64580495daa8fee690bfff70f188eb4f1f24e01480b2949a8c86cbf0a4\" returns successfully"
==> describe nodes <==
Name: scheduled-stop-120087
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=scheduled-stop-120087
kubernetes.io/os=linux
minikube.k8s.io/commit=6017293569ff48e99407bb5ade8e9ba1a7a0c689
minikube.k8s.io/name=scheduled-stop-120087
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2025_10_25T09_47_56_0700
minikube.k8s.io/version=v1.37.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
Annotations: volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Sat, 25 Oct 2025 09:47:52 +0000
Taints: node.kubernetes.io/not-ready:NoSchedule
Unschedulable: false
Lease:
HolderIdentity: scheduled-stop-120087
AcquireTime: <unset>
RenewTime: Sat, 25 Oct 2025 09:47:55 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Sat, 25 Oct 2025 09:47:55 +0000 Sat, 25 Oct 2025 09:47:51 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Sat, 25 Oct 2025 09:47:55 +0000 Sat, 25 Oct 2025 09:47:51 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Sat, 25 Oct 2025 09:47:55 +0000 Sat, 25 Oct 2025 09:47:51 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready False Sat, 25 Oct 2025 09:47:55 +0000 Sat, 25 Oct 2025 09:47:51 +0000 KubeletNotReady container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized
Addresses:
InternalIP: 192.168.76.2
Hostname: scheduled-stop-120087
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32863360Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32863360Ki
pods: 110
System Info:
Machine ID: 98aac72b9abe9f06f1b9b38568f5cc96
System UUID: 1a299b4c-4bb2-441a-98ea-b292513540ae
Boot ID: 590a8a07-3e37-4e62-94d6-23acfbec29af
Kernel Version: 6.8.0-1042-gcp
OS Image: Debian GNU/Linux 12 (bookworm)
Operating System: linux
Architecture: amd64
Container Runtime Version: containerd://1.7.28
Kubelet Version: v1.34.1
Kube-Proxy Version:
Non-terminated Pods: (4 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
kube-system etcd-scheduled-stop-120087 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 4s
kube-system kube-apiserver-scheduled-stop-120087 250m (3%) 0 (0%) 0 (0%) 0 (0%) 3s
kube-system kube-controller-manager-scheduled-stop-120087 200m (2%) 0 (0%) 0 (0%) 0 (0%) 3s
kube-system kube-scheduler-scheduled-stop-120087 100m (1%) 0 (0%) 0 (0%) 0 (0%) 3s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 650m (8%) 0 (0%)
memory 100Mi (0%) 0 (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 3s kubelet Starting kubelet.
Normal NodeAllocatableEnforced 3s kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 3s kubelet Node scheduled-stop-120087 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 3s kubelet Node scheduled-stop-120087 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 3s kubelet Node scheduled-stop-120087 status is now: NodeHasSufficientPID
==> dmesg <==
[ +0.000030] ll header: 00000000: 9e ba e1 7e 23 84 86 8a 5e fa 59 10 08 00
[Oct25 09:16] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000009] ll header: 00000000: ff ff ff ff ff ff 32 4b fd 37 cb 3c 08 06
[ +14.229168] IPv4: martian source 10.244.0.1 from 10.244.0.3, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 9e 62 05 f8 45 f0 08 06
[ +0.000636] IPv4: martian source 10.244.0.3 from 10.244.0.2, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 32 4b fd 37 cb 3c 08 06
[Oct25 09:17] IPv4: martian source 10.244.0.1 from 10.244.0.3, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 9a c6 2e c2 ac 96 08 06
[ +0.000393] IPv4: martian source 10.244.0.3 from 10.244.0.2, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff d6 3c 3f f7 09 73 08 06
[ +1.767747] IPv4: martian source 10.244.0.1 from 10.244.0.3, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff e6 1e cc c6 d5 d5 08 06
[ +18.586103] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 2a 14 eb 62 47 ce 08 06
[ +8.652729] IPv4: martian source 10.244.0.1 from 10.244.0.4, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 7e c7 2e ec bf c3 08 06
[ +0.000358] IPv4: martian source 10.244.0.4 from 10.244.0.2, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff f2 1d cb f7 8e 07 08 06
[ +0.001107] IPv4: martian source 10.244.0.4 from 10.244.0.3, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff e6 1e cc c6 d5 d5 08 06
[Oct25 09:18] IPv4: martian source 10.244.0.1 from 10.244.0.3, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 72 4d f0 7f c3 44 08 06
[ +0.000338] IPv4: martian source 10.244.0.3 from 10.244.0.2, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 2a 14 eb 62 47 ce 08 06
==> etcd [d0ba4a64580495daa8fee690bfff70f188eb4f1f24e01480b2949a8c86cbf0a4] <==
{"level":"warn","ts":"2025-10-25T09:47:52.090324Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48198","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.099198Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48230","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.105491Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48246","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.111924Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48264","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.118066Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48282","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.124806Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48308","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.131897Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48330","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.137917Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48350","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.145536Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48358","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.152751Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48384","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.158856Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48400","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.165157Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48424","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.173146Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48436","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.179410Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48446","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.186053Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48470","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.196506Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48482","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.202482Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48492","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.209139Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48512","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.216477Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48528","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.222436Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48546","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.228555Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48564","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.241893Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48574","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.248633Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48588","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.254709Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48614","server-name":"","error":"EOF"}
{"level":"warn","ts":"2025-10-25T09:47:52.299296Z","caller":"embed/config_logging.go:188","msg":"rejected connection on client endpoint","remote-addr":"127.0.0.1:48626","server-name":"","error":"EOF"}
==> kernel <==
09:47:58 up 1:30, 0 user, load average: 1.01, 1.04, 1.45
Linux scheduled-stop-120087 6.8.0-1042-gcp #45~22.04.1-Ubuntu SMP Tue Oct 7 19:06:40 UTC 2025 x86_64 GNU/Linux
PRETTY_NAME="Debian GNU/Linux 12 (bookworm)"
==> kube-apiserver [25223f701eea7df174b2ba1af625a6aac7d9b7e5ba9982b5de7a4d76a6a31ebc] <==
I1025 09:47:52.788235 1 cache.go:32] Waiting for caches to sync for autoregister controller
I1025 09:47:52.788241 1 cache.go:39] Caches are synced for autoregister controller
I1025 09:47:52.790228 1 shared_informer.go:356] "Caches are synced" controller="*generic.policySource[*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicy,*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicyBinding,k8s.io/apiserver/pkg/admission/plugin/policy/validating.Validator]"
I1025 09:47:52.790255 1 policy_source.go:240] refreshing policies
I1025 09:47:52.792657 1 shared_informer.go:356] "Caches are synced" controller="configmaps"
I1025 09:47:52.798894 1 controller.go:667] quota admission added evaluator for: namespaces
I1025 09:47:52.803851 1 shared_informer.go:356] "Caches are synced" controller="cluster_authentication_trust_controller"
I1025 09:47:52.893765 1 cidrallocator.go:301] created ClusterIP allocator for Service CIDR 10.96.0.0/12
I1025 09:47:52.894125 1 default_servicecidr_controller.go:228] Setting default ServiceCIDR condition Ready to True
I1025 09:47:52.899691 1 cidrallocator.go:277] updated ClusterIP allocator for Service CIDR 10.96.0.0/12
I1025 09:47:52.900021 1 default_servicecidr_controller.go:137] Shutting down kubernetes-service-cidr-controller
I1025 09:47:52.967879 1 controller.go:667] quota admission added evaluator for: leases.coordination.k8s.io
I1025 09:47:53.692881 1 storage_scheduling.go:95] created PriorityClass system-node-critical with value 2000001000
I1025 09:47:53.697833 1 storage_scheduling.go:95] created PriorityClass system-cluster-critical with value 2000000000
I1025 09:47:53.697853 1 storage_scheduling.go:111] all system priority classes are created successfully or already exist.
I1025 09:47:54.158794 1 controller.go:667] quota admission added evaluator for: roles.rbac.authorization.k8s.io
I1025 09:47:54.194079 1 controller.go:667] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
I1025 09:47:54.297147 1 alloc.go:328] "allocated clusterIPs" service="default/kubernetes" clusterIPs={"IPv4":"10.96.0.1"}
W1025 09:47:54.303720 1 lease.go:265] Resetting endpoints for master service "kubernetes" to [192.168.76.2]
I1025 09:47:54.304876 1 controller.go:667] quota admission added evaluator for: endpoints
I1025 09:47:54.309297 1 controller.go:667] quota admission added evaluator for: endpointslices.discovery.k8s.io
I1025 09:47:55.181584 1 controller.go:667] quota admission added evaluator for: serviceaccounts
I1025 09:47:55.328823 1 controller.go:667] quota admission added evaluator for: deployments.apps
I1025 09:47:55.339196 1 alloc.go:328] "allocated clusterIPs" service="kube-system/kube-dns" clusterIPs={"IPv4":"10.96.0.10"}
I1025 09:47:55.347521 1 controller.go:667] quota admission added evaluator for: daemonsets.apps
==> kube-controller-manager [d58e1419868e0a837d3318dcf6e91804a4e9025f3562bc311ce1efe6c3907f79] <==
I1025 09:47:57.479270 1 serviceaccounts_controller.go:114] "Starting service account controller" logger="serviceaccount-controller"
I1025 09:47:57.479286 1 shared_informer.go:349] "Waiting for caches to sync" controller="service account"
I1025 09:47:57.630221 1 controllermanager.go:781] "Started controller" controller="job-controller"
I1025 09:47:57.630398 1 job_controller.go:257] "Starting job controller" logger="job-controller"
I1025 09:47:57.630421 1 shared_informer.go:349] "Waiting for caches to sync" controller="job"
I1025 09:47:57.779247 1 controllermanager.go:781] "Started controller" controller="cronjob-controller"
I1025 09:47:57.779375 1 cronjob_controllerv2.go:145] "Starting cronjob controller v2" logger="cronjob-controller"
I1025 09:47:57.779404 1 shared_informer.go:349] "Waiting for caches to sync" controller="cronjob"
I1025 09:47:57.929681 1 node_lifecycle_controller.go:419] "Controller will reconcile labels" logger="node-lifecycle-controller"
I1025 09:47:57.929734 1 controllermanager.go:781] "Started controller" controller="node-lifecycle-controller"
I1025 09:47:57.929845 1 node_lifecycle_controller.go:453] "Sending events to api server" logger="node-lifecycle-controller"
I1025 09:47:57.929889 1 node_lifecycle_controller.go:464] "Starting node controller" logger="node-lifecycle-controller"
I1025 09:47:57.929907 1 shared_informer.go:349] "Waiting for caches to sync" controller="taint"
I1025 09:47:58.079887 1 controllermanager.go:781] "Started controller" controller="persistentvolume-attach-detach-controller"
I1025 09:47:58.080051 1 attach_detach_controller.go:336] "Starting attach detach controller" logger="persistentvolume-attach-detach-controller"
I1025 09:47:58.080071 1 shared_informer.go:349] "Waiting for caches to sync" controller="attach detach"
I1025 09:47:58.229355 1 controllermanager.go:781] "Started controller" controller="clusterrole-aggregation-controller"
I1025 09:47:58.229420 1 clusterroleaggregation_controller.go:194] "Starting ClusterRoleAggregator controller" logger="clusterrole-aggregation-controller"
I1025 09:47:58.229458 1 shared_informer.go:349] "Waiting for caches to sync" controller="ClusterRoleAggregator"
I1025 09:47:58.379672 1 controllermanager.go:781] "Started controller" controller="resourceclaim-controller"
I1025 09:47:58.379740 1 controller.go:397] "Starting resource claim controller" logger="resourceclaim-controller"
I1025 09:47:58.379824 1 shared_informer.go:349] "Waiting for caches to sync" controller="resource_claim"
I1025 09:47:58.529486 1 controllermanager.go:781] "Started controller" controller="legacy-serviceaccount-token-cleaner-controller"
I1025 09:47:58.529535 1 legacy_serviceaccount_token_cleaner.go:103] "Starting legacy service account token cleaner controller" logger="legacy-serviceaccount-token-cleaner-controller"
I1025 09:47:58.529544 1 shared_informer.go:349] "Waiting for caches to sync" controller="legacy-service-account-token-cleaner"
==> kube-scheduler [694ed725a6cf5d351aee170316fe7b89a81dae03cc239694ff1e3515ea5586fe] <==
E1025 09:47:52.741931 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicationController"
E1025 09:47:52.742169 1 reflector.go:205] "Failed to watch" err="failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.StatefulSet"
E1025 09:47:52.742563 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIStorageCapacity"
E1025 09:47:52.742635 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PodDisruptionBudget"
E1025 09:47:52.742697 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSINode"
E1025 09:47:52.742776 1 reflector.go:205] "Failed to watch" err="failed to list *v1.VolumeAttachment: volumeattachments.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"volumeattachments\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.VolumeAttachment"
E1025 09:47:52.742791 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod"
E1025 09:47:52.742827 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ResourceSlice: resourceslices.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"resourceslices\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ResourceSlice"
E1025 09:47:52.743012 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node"
E1025 09:47:52.743037 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ResourceClaim: resourceclaims.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"resourceclaims\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ResourceClaim"
E1025 09:47:52.743045 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolumeClaim"
E1025 09:47:52.743077 1 reflector.go:205] "Failed to watch" err="failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.StorageClass"
E1025 09:47:52.743133 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicaSet"
E1025 09:47:52.743157 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver"
E1025 09:47:52.743193 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace"
E1025 09:47:53.605303 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ResourceSlice: resourceslices.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"resourceslices\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ResourceSlice"
E1025 09:47:53.611751 1 reflector.go:205] "Failed to watch" err="failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSINode"
E1025 09:47:53.686513 1 reflector.go:205] "Failed to watch" err="failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.StorageClass"
E1025 09:47:53.762325 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace"
E1025 09:47:53.889682 1 reflector.go:205] "Failed to watch" err="failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolume"
E1025 09:47:53.905821 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service"
E1025 09:47:53.942203 1 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node"
E1025 09:47:53.958972 1 reflector.go:205] "Failed to watch" err="failed to list *v1.DeviceClass: deviceclasses.resource.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"deviceclasses\" in API group \"resource.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.DeviceClass"
E1025 09:47:53.972125 1 reflector.go:205] "Failed to watch" err="failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicaSet"
I1025 09:47:54.339759 1 shared_informer.go:356] "Caches are synced" controller="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
==> kubelet <==
Oct 25 09:47:55 scheduled-stop-120087 kubelet[1482]: I1025 09:47:55.430566 1482 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5f8254f424014ac94cf68614848c604c-k8s-certs\") pod \"kube-controller-manager-scheduled-stop-120087\" (UID: \"5f8254f424014ac94cf68614848c604c\") " pod="kube-system/kube-controller-manager-scheduled-stop-120087"
Oct 25 09:47:55 scheduled-stop-120087 kubelet[1482]: I1025 09:47:55.430596 1482 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/efcedea45f933d5b0d8312714f2ac611-ca-certs\") pod \"kube-apiserver-scheduled-stop-120087\" (UID: \"efcedea45f933d5b0d8312714f2ac611\") " pod="kube-system/kube-apiserver-scheduled-stop-120087"
Oct 25 09:47:55 scheduled-stop-120087 kubelet[1482]: I1025 09:47:55.430691 1482 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/efcedea45f933d5b0d8312714f2ac611-usr-local-share-ca-certificates\") pod \"kube-apiserver-scheduled-stop-120087\" (UID: \"efcedea45f933d5b0d8312714f2ac611\") " pod="kube-system/kube-apiserver-scheduled-stop-120087"
Oct 25 09:47:55 scheduled-stop-120087 kubelet[1482]: I1025 09:47:55.430731 1482 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5f8254f424014ac94cf68614848c604c-kubeconfig\") pod \"kube-controller-manager-scheduled-stop-120087\" (UID: \"5f8254f424014ac94cf68614848c604c\") " pod="kube-system/kube-controller-manager-scheduled-stop-120087"
Oct 25 09:47:55 scheduled-stop-120087 kubelet[1482]: I1025 09:47:55.430751 1482 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-data\" (UniqueName: \"kubernetes.io/host-path/42393771443bf7e01fd7a1f384763bc9-etcd-data\") pod \"etcd-scheduled-stop-120087\" (UID: \"42393771443bf7e01fd7a1f384763bc9\") " pod="kube-system/etcd-scheduled-stop-120087"
Oct 25 09:47:55 scheduled-stop-120087 kubelet[1482]: I1025 09:47:55.430770 1482 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/efcedea45f933d5b0d8312714f2ac611-etc-ca-certificates\") pod \"kube-apiserver-scheduled-stop-120087\" (UID: \"efcedea45f933d5b0d8312714f2ac611\") " pod="kube-system/kube-apiserver-scheduled-stop-120087"
Oct 25 09:47:55 scheduled-stop-120087 kubelet[1482]: I1025 09:47:55.430797 1482 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/efcedea45f933d5b0d8312714f2ac611-usr-share-ca-certificates\") pod \"kube-apiserver-scheduled-stop-120087\" (UID: \"efcedea45f933d5b0d8312714f2ac611\") " pod="kube-system/kube-apiserver-scheduled-stop-120087"
Oct 25 09:47:55 scheduled-stop-120087 kubelet[1482]: I1025 09:47:55.430867 1482 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5f8254f424014ac94cf68614848c604c-etc-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-120087\" (UID: \"5f8254f424014ac94cf68614848c604c\") " pod="kube-system/kube-controller-manager-scheduled-stop-120087"
Oct 25 09:47:55 scheduled-stop-120087 kubelet[1482]: I1025 09:47:55.430922 1482 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5f8254f424014ac94cf68614848c604c-usr-share-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-120087\" (UID: \"5f8254f424014ac94cf68614848c604c\") " pod="kube-system/kube-controller-manager-scheduled-stop-120087"
Oct 25 09:47:55 scheduled-stop-120087 kubelet[1482]: I1025 09:47:55.430963 1482 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/c5c25dc7dc33c5716c30ee4a4da1e1bf-kubeconfig\") pod \"kube-scheduler-scheduled-stop-120087\" (UID: \"c5c25dc7dc33c5716c30ee4a4da1e1bf\") " pod="kube-system/kube-scheduler-scheduled-stop-120087"
Oct 25 09:47:55 scheduled-stop-120087 kubelet[1482]: I1025 09:47:55.430986 1482 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/efcedea45f933d5b0d8312714f2ac611-k8s-certs\") pod \"kube-apiserver-scheduled-stop-120087\" (UID: \"efcedea45f933d5b0d8312714f2ac611\") " pod="kube-system/kube-apiserver-scheduled-stop-120087"
Oct 25 09:47:55 scheduled-stop-120087 kubelet[1482]: I1025 09:47:55.431002 1482 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5f8254f424014ac94cf68614848c604c-ca-certs\") pod \"kube-controller-manager-scheduled-stop-120087\" (UID: \"5f8254f424014ac94cf68614848c604c\") " pod="kube-system/kube-controller-manager-scheduled-stop-120087"
Oct 25 09:47:55 scheduled-stop-120087 kubelet[1482]: I1025 09:47:55.431022 1482 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5f8254f424014ac94cf68614848c604c-flexvolume-dir\") pod \"kube-controller-manager-scheduled-stop-120087\" (UID: \"5f8254f424014ac94cf68614848c604c\") " pod="kube-system/kube-controller-manager-scheduled-stop-120087"
Oct 25 09:47:56 scheduled-stop-120087 kubelet[1482]: I1025 09:47:56.124368 1482 apiserver.go:52] "Watching apiserver"
Oct 25 09:47:56 scheduled-stop-120087 kubelet[1482]: I1025 09:47:56.130154 1482 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world"
Oct 25 09:47:56 scheduled-stop-120087 kubelet[1482]: I1025 09:47:56.164423 1482 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-scheduled-stop-120087"
Oct 25 09:47:56 scheduled-stop-120087 kubelet[1482]: I1025 09:47:56.164539 1482 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/etcd-scheduled-stop-120087"
Oct 25 09:47:56 scheduled-stop-120087 kubelet[1482]: I1025 09:47:56.164572 1482 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-scheduled-stop-120087"
Oct 25 09:47:56 scheduled-stop-120087 kubelet[1482]: E1025 09:47:56.171823 1482 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-apiserver-scheduled-stop-120087\" already exists" pod="kube-system/kube-apiserver-scheduled-stop-120087"
Oct 25 09:47:56 scheduled-stop-120087 kubelet[1482]: E1025 09:47:56.172633 1482 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-controller-manager-scheduled-stop-120087\" already exists" pod="kube-system/kube-controller-manager-scheduled-stop-120087"
Oct 25 09:47:56 scheduled-stop-120087 kubelet[1482]: E1025 09:47:56.172925 1482 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"etcd-scheduled-stop-120087\" already exists" pod="kube-system/etcd-scheduled-stop-120087"
Oct 25 09:47:56 scheduled-stop-120087 kubelet[1482]: I1025 09:47:56.186369 1482 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-scheduled-stop-120087" podStartSLOduration=1.1863546839999999 podStartE2EDuration="1.186354684s" podCreationTimestamp="2025-10-25 09:47:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-25 09:47:56.186262494 +0000 UTC m=+1.117908394" watchObservedRunningTime="2025-10-25 09:47:56.186354684 +0000 UTC m=+1.118000582"
Oct 25 09:47:56 scheduled-stop-120087 kubelet[1482]: I1025 09:47:56.207966 1482 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-scheduled-stop-120087" podStartSLOduration=1.207943363 podStartE2EDuration="1.207943363s" podCreationTimestamp="2025-10-25 09:47:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-25 09:47:56.207934442 +0000 UTC m=+1.139580343" watchObservedRunningTime="2025-10-25 09:47:56.207943363 +0000 UTC m=+1.139589263"
Oct 25 09:47:56 scheduled-stop-120087 kubelet[1482]: I1025 09:47:56.208131 1482 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/etcd-scheduled-stop-120087" podStartSLOduration=2.208118876 podStartE2EDuration="2.208118876s" podCreationTimestamp="2025-10-25 09:47:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-25 09:47:56.197745529 +0000 UTC m=+1.129391477" watchObservedRunningTime="2025-10-25 09:47:56.208118876 +0000 UTC m=+1.139764775"
Oct 25 09:47:56 scheduled-stop-120087 kubelet[1482]: I1025 09:47:56.217579 1482 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-scheduled-stop-120087" podStartSLOduration=1.217556012 podStartE2EDuration="1.217556012s" podCreationTimestamp="2025-10-25 09:47:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-25 09:47:56.217166832 +0000 UTC m=+1.148812732" watchObservedRunningTime="2025-10-25 09:47:56.217556012 +0000 UTC m=+1.149201892"
-- /stdout --
helpers_test.go:262: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p scheduled-stop-120087 -n scheduled-stop-120087
helpers_test.go:269: (dbg) Run: kubectl --context scheduled-stop-120087 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:280: non-running pods: storage-provisioner
helpers_test.go:282: ======> post-mortem[TestScheduledStopUnix]: describe non-running pods <======
helpers_test.go:285: (dbg) Run: kubectl --context scheduled-stop-120087 describe pod storage-provisioner
helpers_test.go:285: (dbg) Non-zero exit: kubectl --context scheduled-stop-120087 describe pod storage-provisioner: exit status 1 (60.207943ms)
** stderr **
Error from server (NotFound): pods "storage-provisioner" not found
** /stderr **
helpers_test.go:287: kubectl --context scheduled-stop-120087 describe pod storage-provisioner: exit status 1
helpers_test.go:175: Cleaning up "scheduled-stop-120087" profile ...
helpers_test.go:178: (dbg) Run: out/minikube-linux-amd64 delete -p scheduled-stop-120087
helpers_test.go:178: (dbg) Done: out/minikube-linux-amd64 delete -p scheduled-stop-120087: (2.349579474s)
--- FAIL: TestScheduledStopUnix (25.75s)