=== RUN TestScheduledStopUnix
scheduled_stop_test.go:128: (dbg) Run: out/minikube-linux-arm64 start -p scheduled-stop-579771 --memory=2048 --driver=docker --container-runtime=docker
scheduled_stop_test.go:128: (dbg) Done: out/minikube-linux-arm64 start -p scheduled-stop-579771 --memory=2048 --driver=docker --container-runtime=docker: (32.842620162s)
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-579771 --schedule 5m
scheduled_stop_test.go:191: (dbg) Run: out/minikube-linux-arm64 status --format={{.TimeToStop}} -p scheduled-stop-579771 -n scheduled-stop-579771
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-579771 --schedule 15s
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:98: process 673761 running but should have been killed on reschedule of stop
panic.go:631: *** TestScheduledStopUnix FAILED at 2025-05-10 17:32:15.208072324 +0000 UTC m=+2297.891615931
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestScheduledStopUnix]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect scheduled-stop-579771
helpers_test.go:235: (dbg) docker inspect scheduled-stop-579771:
-- stdout --
[
{
"Id": "8acdd6504c218a03911716c278c7e39bdd433945c619bbaf56e19033659d052c",
"Created": "2025-05-10T17:31:46.865182455Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 670689,
"ExitCode": 0,
"Error": "",
"StartedAt": "2025-05-10T17:31:46.927205453Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:8d0051f61c70b78ce0d27f52d63d179360406349b05070abb34de548473ac66d",
"ResolvConfPath": "/var/lib/docker/containers/8acdd6504c218a03911716c278c7e39bdd433945c619bbaf56e19033659d052c/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/8acdd6504c218a03911716c278c7e39bdd433945c619bbaf56e19033659d052c/hostname",
"HostsPath": "/var/lib/docker/containers/8acdd6504c218a03911716c278c7e39bdd433945c619bbaf56e19033659d052c/hosts",
"LogPath": "/var/lib/docker/containers/8acdd6504c218a03911716c278c7e39bdd433945c619bbaf56e19033659d052c/8acdd6504c218a03911716c278c7e39bdd433945c619bbaf56e19033659d052c-json.log",
"Name": "/scheduled-stop-579771",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"scheduled-stop-579771:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {}
},
"NetworkMode": "scheduled-stop-579771",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 2147483648,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 4294967296,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"ID": "8acdd6504c218a03911716c278c7e39bdd433945c619bbaf56e19033659d052c",
"LowerDir": "/var/lib/docker/overlay2/795569356da99589a90bde22b9106234a8ea03b755770f195d4d8550f6146d2d-init/diff:/var/lib/docker/overlay2/7602a725c5ff3328a98ae2d468b9aab3fb98b26cc00795aed37c457fc8241ca1/diff",
"MergedDir": "/var/lib/docker/overlay2/795569356da99589a90bde22b9106234a8ea03b755770f195d4d8550f6146d2d/merged",
"UpperDir": "/var/lib/docker/overlay2/795569356da99589a90bde22b9106234a8ea03b755770f195d4d8550f6146d2d/diff",
"WorkDir": "/var/lib/docker/overlay2/795569356da99589a90bde22b9106234a8ea03b755770f195d4d8550f6146d2d/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "scheduled-stop-579771",
"Source": "/var/lib/docker/volumes/scheduled-stop-579771/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "scheduled-stop-579771",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1746731792-20718@sha256:074d9afa1e8827ea0e101248fc55098d304814b5d8bf485882a81afc90084155",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "scheduled-stop-579771",
"name.minikube.sigs.k8s.io": "scheduled-stop-579771",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "f16967f1a1f080aba0a95a6f86b2fae944a7bdf505b91c709996ab724ed097ab",
"SandboxKey": "/var/run/docker/netns/f16967f1a1f0",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33367"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33368"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33371"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33369"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33370"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"scheduled-stop-579771": {
"IPAMConfig": {
"IPv4Address": "192.168.76.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "26:82:51:08:ad:5f",
"DriverOpts": null,
"GwPriority": 0,
"NetworkID": "031ff4ed3bfc55b5ffaf111abb5e6650d3cfe4dd6fee3589b4f4eecdca391b6f",
"EndpointID": "92ddbca4a9b44b242fa85ecb72965fb6e16fed389f9f314608ca2ff0a63dc228",
"Gateway": "192.168.76.1",
"IPAddress": "192.168.76.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"scheduled-stop-579771",
"8acdd6504c21"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-arm64 status --format={{.Host}} -p scheduled-stop-579771 -n scheduled-stop-579771
helpers_test.go:244: <<< TestScheduledStopUnix FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestScheduledStopUnix]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-arm64 -p scheduled-stop-579771 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-arm64 -p scheduled-stop-579771 logs -n 25: (1.040327591s)
helpers_test.go:252: TestScheduledStopUnix logs:
-- stdout --
==> Audit <==
|---------|--------------------------------|-----------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|--------------------------------|-----------------------|---------|---------|---------------------|---------------------|
| stop | -p multinode-298968 | multinode-298968 | jenkins | v1.35.0 | 10 May 25 17:25 UTC | 10 May 25 17:25 UTC |
| start | -p multinode-298968 | multinode-298968 | jenkins | v1.35.0 | 10 May 25 17:25 UTC | 10 May 25 17:26 UTC |
| | --wait=true -v=5 | | | | | |
| | --alsologtostderr | | | | | |
| node | list -p multinode-298968 | multinode-298968 | jenkins | v1.35.0 | 10 May 25 17:26 UTC | |
| node | multinode-298968 node delete | multinode-298968 | jenkins | v1.35.0 | 10 May 25 17:26 UTC | 10 May 25 17:26 UTC |
| | m03 | | | | | |
| stop | multinode-298968 stop | multinode-298968 | jenkins | v1.35.0 | 10 May 25 17:26 UTC | 10 May 25 17:27 UTC |
| start | -p multinode-298968 | multinode-298968 | jenkins | v1.35.0 | 10 May 25 17:27 UTC | 10 May 25 17:28 UTC |
| | --wait=true -v=5 | | | | | |
| | --alsologtostderr | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| node | list -p multinode-298968 | multinode-298968 | jenkins | v1.35.0 | 10 May 25 17:28 UTC | |
| start | -p multinode-298968-m02 | multinode-298968-m02 | jenkins | v1.35.0 | 10 May 25 17:28 UTC | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| start | -p multinode-298968-m03 | multinode-298968-m03 | jenkins | v1.35.0 | 10 May 25 17:28 UTC | 10 May 25 17:28 UTC |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| node | add -p multinode-298968 | multinode-298968 | jenkins | v1.35.0 | 10 May 25 17:28 UTC | |
| delete | -p multinode-298968-m03 | multinode-298968-m03 | jenkins | v1.35.0 | 10 May 25 17:28 UTC | 10 May 25 17:28 UTC |
| delete | -p multinode-298968 | multinode-298968 | jenkins | v1.35.0 | 10 May 25 17:28 UTC | 10 May 25 17:28 UTC |
| start | -p test-preload-048602 | test-preload-048602 | jenkins | v1.35.0 | 10 May 25 17:28 UTC | 10 May 25 17:30 UTC |
| | --memory=2200 | | | | | |
| | --alsologtostderr | | | | | |
| | --wait=true --preload=false | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --kubernetes-version=v1.24.4 | | | | | |
| image | test-preload-048602 image pull | test-preload-048602 | jenkins | v1.35.0 | 10 May 25 17:30 UTC | 10 May 25 17:30 UTC |
| | gcr.io/k8s-minikube/busybox | | | | | |
| stop | -p test-preload-048602 | test-preload-048602 | jenkins | v1.35.0 | 10 May 25 17:30 UTC | 10 May 25 17:30 UTC |
| start | -p test-preload-048602 | test-preload-048602 | jenkins | v1.35.0 | 10 May 25 17:30 UTC | 10 May 25 17:31 UTC |
| | --memory=2200 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| | --wait=true --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| image | test-preload-048602 image list | test-preload-048602 | jenkins | v1.35.0 | 10 May 25 17:31 UTC | 10 May 25 17:31 UTC |
| delete | -p test-preload-048602 | test-preload-048602 | jenkins | v1.35.0 | 10 May 25 17:31 UTC | 10 May 25 17:31 UTC |
| start | -p scheduled-stop-579771 | scheduled-stop-579771 | jenkins | v1.35.0 | 10 May 25 17:31 UTC | 10 May 25 17:32 UTC |
| | --memory=2048 --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| stop | -p scheduled-stop-579771 | scheduled-stop-579771 | jenkins | v1.35.0 | 10 May 25 17:32 UTC | |
| | --schedule 5m | | | | | |
| stop | -p scheduled-stop-579771 | scheduled-stop-579771 | jenkins | v1.35.0 | 10 May 25 17:32 UTC | |
| | --schedule 5m | | | | | |
| stop | -p scheduled-stop-579771 | scheduled-stop-579771 | jenkins | v1.35.0 | 10 May 25 17:32 UTC | |
| | --schedule 5m | | | | | |
| stop | -p scheduled-stop-579771 | scheduled-stop-579771 | jenkins | v1.35.0 | 10 May 25 17:32 UTC | |
| | --schedule 15s | | | | | |
| stop | -p scheduled-stop-579771 | scheduled-stop-579771 | jenkins | v1.35.0 | 10 May 25 17:32 UTC | |
| | --schedule 15s | | | | | |
| stop | -p scheduled-stop-579771 | scheduled-stop-579771 | jenkins | v1.35.0 | 10 May 25 17:32 UTC | |
| | --schedule 15s | | | | | |
|---------|--------------------------------|-----------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2025/05/10 17:31:41
Running on machine: ip-172-31-30-239
Binary: Built with gc go1.24.0 for linux/arm64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0510 17:31:41.880380 670298 out.go:345] Setting OutFile to fd 1 ...
I0510 17:31:41.880504 670298 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0510 17:31:41.880507 670298 out.go:358] Setting ErrFile to fd 2...
I0510 17:31:41.880511 670298 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0510 17:31:41.880768 670298 root.go:338] Updating PATH: /home/jenkins/minikube-integration/20720-453270/.minikube/bin
I0510 17:31:41.881162 670298 out.go:352] Setting JSON to false
I0510 17:31:41.882023 670298 start.go:130] hostinfo: {"hostname":"ip-172-31-30-239","uptime":8053,"bootTime":1746890249,"procs":154,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1083-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"92f46a7d-c249-4c12-924a-77f64874c910"}
I0510 17:31:41.882082 670298 start.go:140] virtualization:
I0510 17:31:41.885821 670298 out.go:177] * [scheduled-stop-579771] minikube v1.35.0 on Ubuntu 20.04 (arm64)
I0510 17:31:41.890360 670298 out.go:177] - MINIKUBE_LOCATION=20720
I0510 17:31:41.890538 670298 notify.go:220] Checking for updates...
I0510 17:31:41.896826 670298 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0510 17:31:41.900031 670298 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/20720-453270/kubeconfig
I0510 17:31:41.903241 670298 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/20720-453270/.minikube
I0510 17:31:41.906354 670298 out.go:177] - MINIKUBE_BIN=out/minikube-linux-arm64
I0510 17:31:41.909448 670298 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0510 17:31:41.912584 670298 driver.go:404] Setting default libvirt URI to qemu:///system
I0510 17:31:41.932847 670298 docker.go:123] docker version: linux-28.1.1:Docker Engine - Community
I0510 17:31:41.932958 670298 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0510 17:31:41.993528 670298 info.go:266] docker info: {ID:6ZPO:QZND:VNGE:LUKL:4Y3K:XELL:AAX4:2GTK:E6LM:MPRN:3ZXR:TTMR Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:24 OomKillDisable:true NGoroutines:42 SystemTime:2025-05-10 17:31:41.981831327 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1083-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214831104 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-30-239 Labels:[] ExperimentalBuild:false ServerVersion:28.1.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:05044ec0a9a75232cad458027ca83437aae3f4da Expected:} RuncCommit:{ID:v1.2.5-0-g59923ef Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx P
ath:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.23.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.35.1]] Warnings:<nil>}}
I0510 17:31:41.993636 670298 docker.go:318] overlay module found
I0510 17:31:41.997187 670298 out.go:177] * Using the docker driver based on user configuration
I0510 17:31:42.000453 670298 start.go:304] selected driver: docker
I0510 17:31:42.000480 670298 start.go:908] validating driver "docker" against <nil>
I0510 17:31:42.000493 670298 start.go:919] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0510 17:31:42.001226 670298 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0510 17:31:42.060578 670298 info.go:266] docker info: {ID:6ZPO:QZND:VNGE:LUKL:4Y3K:XELL:AAX4:2GTK:E6LM:MPRN:3ZXR:TTMR Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:24 OomKillDisable:true NGoroutines:42 SystemTime:2025-05-10 17:31:42.050607289 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1083-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214831104 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-30-239 Labels:[] ExperimentalBuild:false ServerVersion:28.1.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:05044ec0a9a75232cad458027ca83437aae3f4da Expected:} RuncCommit:{ID:v1.2.5-0-g59923ef Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx P
ath:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.23.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.35.1]] Warnings:<nil>}}
I0510 17:31:42.060744 670298 start_flags.go:311] no existing cluster config was found, will generate one from the flags
I0510 17:31:42.060983 670298 start_flags.go:957] Wait components to verify : map[apiserver:true system_pods:true]
I0510 17:31:42.064083 670298 out.go:177] * Using Docker driver with root privileges
I0510 17:31:42.067284 670298 cni.go:84] Creating CNI manager for ""
I0510 17:31:42.067361 670298 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0510 17:31:42.067375 670298 start_flags.go:320] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0510 17:31:42.067472 670298 start.go:347] cluster config:
{Name:scheduled-stop-579771 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1746731792-20718@sha256:074d9afa1e8827ea0e101248fc55098d304814b5d8bf485882a81afc90084155 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.33.0 ClusterName:scheduled-stop-579771 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local
ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.33.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0510 17:31:42.072828 670298 out.go:177] * Starting "scheduled-stop-579771" primary control-plane node in "scheduled-stop-579771" cluster
I0510 17:31:42.075949 670298 cache.go:121] Beginning downloading kic base image for docker with docker
I0510 17:31:42.079022 670298 out.go:177] * Pulling base image v0.0.46-1746731792-20718 ...
I0510 17:31:42.082159 670298 preload.go:131] Checking if preload exists for k8s version v1.33.0 and runtime docker
I0510 17:31:42.082251 670298 image.go:81] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1746731792-20718@sha256:074d9afa1e8827ea0e101248fc55098d304814b5d8bf485882a81afc90084155 in local docker daemon
I0510 17:31:42.082280 670298 preload.go:146] Found local preload: /home/jenkins/minikube-integration/20720-453270/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.33.0-docker-overlay2-arm64.tar.lz4
I0510 17:31:42.082289 670298 cache.go:56] Caching tarball of preloaded images
I0510 17:31:42.082424 670298 preload.go:172] Found /home/jenkins/minikube-integration/20720-453270/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.33.0-docker-overlay2-arm64.tar.lz4 in cache, skipping download
I0510 17:31:42.082437 670298 cache.go:59] Finished verifying existence of preloaded tar for v1.33.0 on docker
I0510 17:31:42.082887 670298 profile.go:143] Saving config to /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/config.json ...
I0510 17:31:42.082917 670298 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/config.json: {Name:mkc0cb38cbc4793821826ced7a76146f5fd9f01f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0510 17:31:42.106231 670298 image.go:100] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1746731792-20718@sha256:074d9afa1e8827ea0e101248fc55098d304814b5d8bf485882a81afc90084155 in local docker daemon, skipping pull
I0510 17:31:42.106246 670298 cache.go:145] gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1746731792-20718@sha256:074d9afa1e8827ea0e101248fc55098d304814b5d8bf485882a81afc90084155 exists in daemon, skipping load
I0510 17:31:42.106269 670298 cache.go:230] Successfully downloaded all kic artifacts
I0510 17:31:42.106299 670298 start.go:360] acquireMachinesLock for scheduled-stop-579771: {Name:mk9288978120dd577b861e362b90e5bc729f5084 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0510 17:31:42.106426 670298 start.go:364] duration metric: took 111.311µs to acquireMachinesLock for "scheduled-stop-579771"
I0510 17:31:42.106457 670298 start.go:93] Provisioning new machine with config: &{Name:scheduled-stop-579771 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1746731792-20718@sha256:074d9afa1e8827ea0e101248fc55098d304814b5d8bf485882a81afc90084155 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.33.0 ClusterName:scheduled-stop-579771 Namespace:default APIServerHAVIP: A
PIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.33.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath
: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.33.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0510 17:31:42.106537 670298 start.go:125] createHost starting for "" (driver="docker")
I0510 17:31:42.112163 670298 out.go:235] * Creating docker container (CPUs=2, Memory=2048MB) ...
I0510 17:31:42.112444 670298 start.go:159] libmachine.API.Create for "scheduled-stop-579771" (driver="docker")
I0510 17:31:42.112481 670298 client.go:168] LocalClient.Create starting
I0510 17:31:42.112577 670298 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/20720-453270/.minikube/certs/ca.pem
I0510 17:31:42.112629 670298 main.go:141] libmachine: Decoding PEM data...
I0510 17:31:42.112643 670298 main.go:141] libmachine: Parsing certificate...
I0510 17:31:42.112722 670298 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/20720-453270/.minikube/certs/cert.pem
I0510 17:31:42.112742 670298 main.go:141] libmachine: Decoding PEM data...
I0510 17:31:42.112750 670298 main.go:141] libmachine: Parsing certificate...
I0510 17:31:42.113211 670298 cli_runner.go:164] Run: docker network inspect scheduled-stop-579771 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0510 17:31:42.131515 670298 cli_runner.go:211] docker network inspect scheduled-stop-579771 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0510 17:31:42.131600 670298 network_create.go:284] running [docker network inspect scheduled-stop-579771] to gather additional debugging logs...
I0510 17:31:42.131618 670298 cli_runner.go:164] Run: docker network inspect scheduled-stop-579771
W0510 17:31:42.151036 670298 cli_runner.go:211] docker network inspect scheduled-stop-579771 returned with exit code 1
I0510 17:31:42.151057 670298 network_create.go:287] error running [docker network inspect scheduled-stop-579771]: docker network inspect scheduled-stop-579771: exit status 1
stdout:
[]
stderr:
Error response from daemon: network scheduled-stop-579771 not found
I0510 17:31:42.151071 670298 network_create.go:289] output of [docker network inspect scheduled-stop-579771]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network scheduled-stop-579771 not found
** /stderr **
I0510 17:31:42.151227 670298 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0510 17:31:42.178443 670298 network.go:211] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-d5754c5efbf3 IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:8a:b6:74:41:aa:b3} reservation:<nil>}
I0510 17:31:42.178814 670298 network.go:211] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-1b5976f2ca9c IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:62:7b:aa:c2:cc:7a} reservation:<nil>}
I0510 17:31:42.179028 670298 network.go:211] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-657f5414fda5 IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:4a:ad:0e:cf:e6:60} reservation:<nil>}
I0510 17:31:42.179433 670298 network.go:206] using free private subnet 192.168.76.0/24: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x400198f120}
I0510 17:31:42.179452 670298 network_create.go:124] attempt to create docker network scheduled-stop-579771 192.168.76.0/24 with gateway 192.168.76.1 and MTU of 1500 ...
I0510 17:31:42.179513 670298 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.76.0/24 --gateway=192.168.76.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=scheduled-stop-579771 scheduled-stop-579771
I0510 17:31:42.267157 670298 network_create.go:108] docker network scheduled-stop-579771 192.168.76.0/24 created
I0510 17:31:42.267179 670298 kic.go:121] calculated static IP "192.168.76.2" for the "scheduled-stop-579771" container
I0510 17:31:42.267265 670298 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0510 17:31:42.284773 670298 cli_runner.go:164] Run: docker volume create scheduled-stop-579771 --label name.minikube.sigs.k8s.io=scheduled-stop-579771 --label created_by.minikube.sigs.k8s.io=true
I0510 17:31:42.303584 670298 oci.go:103] Successfully created a docker volume scheduled-stop-579771
I0510 17:31:42.303666 670298 cli_runner.go:164] Run: docker run --rm --name scheduled-stop-579771-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-579771 --entrypoint /usr/bin/test -v scheduled-stop-579771:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1746731792-20718@sha256:074d9afa1e8827ea0e101248fc55098d304814b5d8bf485882a81afc90084155 -d /var/lib
I0510 17:31:42.816387 670298 oci.go:107] Successfully prepared a docker volume scheduled-stop-579771
I0510 17:31:42.816423 670298 preload.go:131] Checking if preload exists for k8s version v1.33.0 and runtime docker
I0510 17:31:42.816442 670298 kic.go:194] Starting extracting preloaded images to volume ...
I0510 17:31:42.816515 670298 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/20720-453270/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.33.0-docker-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-579771:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1746731792-20718@sha256:074d9afa1e8827ea0e101248fc55098d304814b5d8bf485882a81afc90084155 -I lz4 -xf /preloaded.tar -C /extractDir
I0510 17:31:46.794005 670298 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/20720-453270/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.33.0-docker-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-579771:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1746731792-20718@sha256:074d9afa1e8827ea0e101248fc55098d304814b5d8bf485882a81afc90084155 -I lz4 -xf /preloaded.tar -C /extractDir: (3.977451621s)
I0510 17:31:46.794025 670298 kic.go:203] duration metric: took 3.977579399s to extract preloaded images to volume ...
W0510 17:31:46.794184 670298 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0510 17:31:46.794288 670298 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0510 17:31:46.850455 670298 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname scheduled-stop-579771 --name scheduled-stop-579771 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-579771 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=scheduled-stop-579771 --network scheduled-stop-579771 --ip 192.168.76.2 --volume scheduled-stop-579771:/var --security-opt apparmor=unconfined --memory=2048mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1746731792-20718@sha256:074d9afa1e8827ea0e101248fc55098d304814b5d8bf485882a81afc90084155
I0510 17:31:47.154525 670298 cli_runner.go:164] Run: docker container inspect scheduled-stop-579771 --format={{.State.Running}}
I0510 17:31:47.174607 670298 cli_runner.go:164] Run: docker container inspect scheduled-stop-579771 --format={{.State.Status}}
I0510 17:31:47.195412 670298 cli_runner.go:164] Run: docker exec scheduled-stop-579771 stat /var/lib/dpkg/alternatives/iptables
I0510 17:31:47.248648 670298 oci.go:144] the created container "scheduled-stop-579771" has a running status.
I0510 17:31:47.248667 670298 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/20720-453270/.minikube/machines/scheduled-stop-579771/id_rsa...
I0510 17:31:48.015145 670298 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/20720-453270/.minikube/machines/scheduled-stop-579771/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0510 17:31:48.065426 670298 cli_runner.go:164] Run: docker container inspect scheduled-stop-579771 --format={{.State.Status}}
I0510 17:31:48.085266 670298 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0510 17:31:48.085281 670298 kic_runner.go:114] Args: [docker exec --privileged scheduled-stop-579771 chown docker:docker /home/docker/.ssh/authorized_keys]
I0510 17:31:48.132935 670298 cli_runner.go:164] Run: docker container inspect scheduled-stop-579771 --format={{.State.Status}}
I0510 17:31:48.157898 670298 machine.go:93] provisionDockerMachine start ...
I0510 17:31:48.157980 670298 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-579771
I0510 17:31:48.179652 670298 main.go:141] libmachine: Using SSH client type: native
I0510 17:31:48.180153 670298 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3e66e0] 0x3e8ea0 <nil> [] 0s} 127.0.0.1 33367 <nil> <nil>}
I0510 17:31:48.180161 670298 main.go:141] libmachine: About to run SSH command:
hostname
I0510 17:31:48.322872 670298 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-579771
I0510 17:31:48.322885 670298 ubuntu.go:169] provisioning hostname "scheduled-stop-579771"
I0510 17:31:48.322953 670298 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-579771
I0510 17:31:48.343522 670298 main.go:141] libmachine: Using SSH client type: native
I0510 17:31:48.343832 670298 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3e66e0] 0x3e8ea0 <nil> [] 0s} 127.0.0.1 33367 <nil> <nil>}
I0510 17:31:48.343841 670298 main.go:141] libmachine: About to run SSH command:
sudo hostname scheduled-stop-579771 && echo "scheduled-stop-579771" | sudo tee /etc/hostname
I0510 17:31:48.490895 670298 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-579771
I0510 17:31:48.490977 670298 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-579771
I0510 17:31:48.508791 670298 main.go:141] libmachine: Using SSH client type: native
I0510 17:31:48.509089 670298 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3e66e0] 0x3e8ea0 <nil> [] 0s} 127.0.0.1 33367 <nil> <nil>}
I0510 17:31:48.509105 670298 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sscheduled-stop-579771' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 scheduled-stop-579771/g' /etc/hosts;
else
echo '127.0.1.1 scheduled-stop-579771' | sudo tee -a /etc/hosts;
fi
fi
I0510 17:31:48.646858 670298 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0510 17:31:48.646896 670298 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/20720-453270/.minikube CaCertPath:/home/jenkins/minikube-integration/20720-453270/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/20720-453270/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/20720-453270/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/20720-453270/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/20720-453270/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/20720-453270/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/20720-453270/.minikube}
I0510 17:31:48.646921 670298 ubuntu.go:177] setting up certificates
I0510 17:31:48.646930 670298 provision.go:84] configureAuth start
I0510 17:31:48.646995 670298 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-579771
I0510 17:31:48.666466 670298 provision.go:143] copyHostCerts
I0510 17:31:48.666533 670298 exec_runner.go:144] found /home/jenkins/minikube-integration/20720-453270/.minikube/ca.pem, removing ...
I0510 17:31:48.666541 670298 exec_runner.go:203] rm: /home/jenkins/minikube-integration/20720-453270/.minikube/ca.pem
I0510 17:31:48.666616 670298 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20720-453270/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/20720-453270/.minikube/ca.pem (1082 bytes)
I0510 17:31:48.666742 670298 exec_runner.go:144] found /home/jenkins/minikube-integration/20720-453270/.minikube/cert.pem, removing ...
I0510 17:31:48.666747 670298 exec_runner.go:203] rm: /home/jenkins/minikube-integration/20720-453270/.minikube/cert.pem
I0510 17:31:48.666776 670298 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20720-453270/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/20720-453270/.minikube/cert.pem (1123 bytes)
I0510 17:31:48.666826 670298 exec_runner.go:144] found /home/jenkins/minikube-integration/20720-453270/.minikube/key.pem, removing ...
I0510 17:31:48.666829 670298 exec_runner.go:203] rm: /home/jenkins/minikube-integration/20720-453270/.minikube/key.pem
I0510 17:31:48.666852 670298 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20720-453270/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/20720-453270/.minikube/key.pem (1675 bytes)
I0510 17:31:48.666897 670298 provision.go:117] generating server cert: /home/jenkins/minikube-integration/20720-453270/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/20720-453270/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/20720-453270/.minikube/certs/ca-key.pem org=jenkins.scheduled-stop-579771 san=[127.0.0.1 192.168.76.2 localhost minikube scheduled-stop-579771]
I0510 17:31:49.316445 670298 provision.go:177] copyRemoteCerts
I0510 17:31:49.316504 670298 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0510 17:31:49.316543 670298 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-579771
I0510 17:31:49.334125 670298 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33367 SSHKeyPath:/home/jenkins/minikube-integration/20720-453270/.minikube/machines/scheduled-stop-579771/id_rsa Username:docker}
I0510 17:31:49.435619 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
I0510 17:31:49.459434 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/machines/server.pem --> /etc/docker/server.pem (1233 bytes)
I0510 17:31:49.483974 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0510 17:31:49.507987 670298 provision.go:87] duration metric: took 861.044827ms to configureAuth
I0510 17:31:49.508003 670298 ubuntu.go:193] setting minikube options for container-runtime
I0510 17:31:49.508193 670298 config.go:182] Loaded profile config "scheduled-stop-579771": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.33.0
I0510 17:31:49.508243 670298 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-579771
I0510 17:31:49.526076 670298 main.go:141] libmachine: Using SSH client type: native
I0510 17:31:49.526392 670298 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3e66e0] 0x3e8ea0 <nil> [] 0s} 127.0.0.1 33367 <nil> <nil>}
I0510 17:31:49.526399 670298 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0510 17:31:49.659067 670298 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0510 17:31:49.659079 670298 ubuntu.go:71] root file system type: overlay
I0510 17:31:49.659207 670298 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0510 17:31:49.659275 670298 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-579771
I0510 17:31:49.676774 670298 main.go:141] libmachine: Using SSH client type: native
I0510 17:31:49.677100 670298 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3e66e0] 0x3e8ea0 <nil> [] 0s} 127.0.0.1 33367 <nil> <nil>}
I0510 17:31:49.677172 670298 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0510 17:31:49.822965 670298 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0510 17:31:49.823051 670298 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-579771
I0510 17:31:49.840174 670298 main.go:141] libmachine: Using SSH client type: native
I0510 17:31:49.840477 670298 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3e66e0] 0x3e8ea0 <nil> [] 0s} 127.0.0.1 33367 <nil> <nil>}
I0510 17:31:49.840492 670298 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0510 17:31:50.682461 670298 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2025-04-18 09:50:43.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2025-05-10 17:31:49.817261178 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target nss-lookup.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0510 17:31:50.682483 670298 machine.go:96] duration metric: took 2.524572267s to provisionDockerMachine
I0510 17:31:50.682493 670298 client.go:171] duration metric: took 8.570006269s to LocalClient.Create
I0510 17:31:50.682512 670298 start.go:167] duration metric: took 8.570070326s to libmachine.API.Create "scheduled-stop-579771"
I0510 17:31:50.682518 670298 start.go:293] postStartSetup for "scheduled-stop-579771" (driver="docker")
I0510 17:31:50.682527 670298 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0510 17:31:50.682603 670298 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0510 17:31:50.682718 670298 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-579771
I0510 17:31:50.700949 670298 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33367 SSHKeyPath:/home/jenkins/minikube-integration/20720-453270/.minikube/machines/scheduled-stop-579771/id_rsa Username:docker}
I0510 17:31:50.796248 670298 ssh_runner.go:195] Run: cat /etc/os-release
I0510 17:31:50.799784 670298 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0510 17:31:50.799806 670298 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0510 17:31:50.799816 670298 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0510 17:31:50.799822 670298 info.go:137] Remote host: Ubuntu 22.04.5 LTS
I0510 17:31:50.799836 670298 filesync.go:126] Scanning /home/jenkins/minikube-integration/20720-453270/.minikube/addons for local assets ...
I0510 17:31:50.799890 670298 filesync.go:126] Scanning /home/jenkins/minikube-integration/20720-453270/.minikube/files for local assets ...
I0510 17:31:50.799995 670298 filesync.go:149] local asset: /home/jenkins/minikube-integration/20720-453270/.minikube/files/etc/ssl/certs/4586462.pem -> 4586462.pem in /etc/ssl/certs
I0510 17:31:50.800113 670298 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I0510 17:31:50.809009 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/files/etc/ssl/certs/4586462.pem --> /etc/ssl/certs/4586462.pem (1708 bytes)
I0510 17:31:50.833661 670298 start.go:296] duration metric: took 151.128846ms for postStartSetup
I0510 17:31:50.834015 670298 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-579771
I0510 17:31:50.855211 670298 profile.go:143] Saving config to /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/config.json ...
I0510 17:31:50.855487 670298 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0510 17:31:50.855528 670298 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-579771
I0510 17:31:50.872557 670298 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33367 SSHKeyPath:/home/jenkins/minikube-integration/20720-453270/.minikube/machines/scheduled-stop-579771/id_rsa Username:docker}
I0510 17:31:50.968424 670298 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0510 17:31:50.973102 670298 start.go:128] duration metric: took 8.866550388s to createHost
I0510 17:31:50.973118 670298 start.go:83] releasing machines lock for "scheduled-stop-579771", held for 8.866682007s
I0510 17:31:50.973192 670298 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-579771
I0510 17:31:50.999554 670298 ssh_runner.go:195] Run: cat /version.json
I0510 17:31:50.999603 670298 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-579771
I0510 17:31:50.999858 670298 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0510 17:31:50.999910 670298 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-579771
I0510 17:31:51.027171 670298 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33367 SSHKeyPath:/home/jenkins/minikube-integration/20720-453270/.minikube/machines/scheduled-stop-579771/id_rsa Username:docker}
I0510 17:31:51.030769 670298 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33367 SSHKeyPath:/home/jenkins/minikube-integration/20720-453270/.minikube/machines/scheduled-stop-579771/id_rsa Username:docker}
I0510 17:31:51.122597 670298 ssh_runner.go:195] Run: systemctl --version
I0510 17:31:51.253775 670298 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0510 17:31:51.258053 670298 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0510 17:31:51.283928 670298 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0510 17:31:51.283996 670298 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0510 17:31:51.315520 670298 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0510 17:31:51.315535 670298 start.go:495] detecting cgroup driver to use...
I0510 17:31:51.315567 670298 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0510 17:31:51.315659 670298 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0510 17:31:51.332093 670298 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0510 17:31:51.342328 670298 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0510 17:31:51.352601 670298 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0510 17:31:51.352662 670298 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0510 17:31:51.362312 670298 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0510 17:31:51.372096 670298 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0510 17:31:51.382059 670298 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0510 17:31:51.392748 670298 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0510 17:31:51.402744 670298 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0510 17:31:51.412718 670298 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0510 17:31:51.423128 670298 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0510 17:31:51.433284 670298 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0510 17:31:51.442292 670298 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0510 17:31:51.451307 670298 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0510 17:31:51.533771 670298 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0510 17:31:51.639849 670298 start.go:495] detecting cgroup driver to use...
I0510 17:31:51.639886 670298 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0510 17:31:51.639941 670298 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0510 17:31:51.654030 670298 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0510 17:31:51.654099 670298 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0510 17:31:51.667720 670298 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0510 17:31:51.689995 670298 ssh_runner.go:195] Run: which cri-dockerd
I0510 17:31:51.694224 670298 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0510 17:31:51.703397 670298 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0510 17:31:51.722951 670298 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0510 17:31:51.837971 670298 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0510 17:31:51.945910 670298 docker.go:582] configuring docker to use "cgroupfs" as cgroup driver...
I0510 17:31:51.946002 670298 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0510 17:31:51.969476 670298 ssh_runner.go:195] Run: sudo systemctl reset-failed docker
I0510 17:31:51.986274 670298 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0510 17:31:52.094309 670298 ssh_runner.go:195] Run: sudo systemctl restart docker
I0510 17:31:52.404728 670298 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0510 17:31:52.416884 670298 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0510 17:31:52.428955 670298 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0510 17:31:52.523378 670298 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0510 17:31:52.614754 670298 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0510 17:31:52.706717 670298 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0510 17:31:52.721383 670298 ssh_runner.go:195] Run: sudo systemctl reset-failed cri-docker.service
I0510 17:31:52.734376 670298 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0510 17:31:52.827047 670298 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0510 17:31:52.905229 670298 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0510 17:31:52.918729 670298 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0510 17:31:52.918790 670298 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0510 17:31:52.922558 670298 start.go:563] Will wait 60s for crictl version
I0510 17:31:52.922612 670298 ssh_runner.go:195] Run: which crictl
I0510 17:31:52.926169 670298 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0510 17:31:52.963118 670298 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 28.1.1
RuntimeApiVersion: v1
I0510 17:31:52.963178 670298 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0510 17:31:52.989675 670298 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0510 17:31:53.017908 670298 out.go:235] * Preparing Kubernetes v1.33.0 on Docker 28.1.1 ...
I0510 17:31:53.018019 670298 cli_runner.go:164] Run: docker network inspect scheduled-stop-579771 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0510 17:31:53.034826 670298 ssh_runner.go:195] Run: grep 192.168.76.1 host.minikube.internal$ /etc/hosts
I0510 17:31:53.038464 670298 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.76.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0510 17:31:53.050301 670298 kubeadm.go:875] updating cluster {Name:scheduled-stop-579771 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1746731792-20718@sha256:074d9afa1e8827ea0e101248fc55098d304814b5d8bf485882a81afc90084155 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.33.0 ClusterName:scheduled-stop-579771 Namespace:default APIServerHAVIP: APIServerName:miniku
beCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.33.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: Stati
cIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0510 17:31:53.050400 670298 preload.go:131] Checking if preload exists for k8s version v1.33.0 and runtime docker
I0510 17:31:53.050454 670298 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0510 17:31:53.069972 670298 docker.go:697] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.33.0
registry.k8s.io/kube-controller-manager:v1.33.0
registry.k8s.io/kube-scheduler:v1.33.0
registry.k8s.io/kube-proxy:v1.33.0
registry.k8s.io/etcd:3.5.21-0
registry.k8s.io/coredns/coredns:v1.12.0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0510 17:31:53.069985 670298 docker.go:627] Images already preloaded, skipping extraction
I0510 17:31:53.070046 670298 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0510 17:31:53.093348 670298 docker.go:697] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.33.0
registry.k8s.io/kube-scheduler:v1.33.0
registry.k8s.io/kube-controller-manager:v1.33.0
registry.k8s.io/kube-proxy:v1.33.0
registry.k8s.io/etcd:3.5.21-0
registry.k8s.io/coredns/coredns:v1.12.0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0510 17:31:53.093362 670298 cache_images.go:84] Images are preloaded, skipping loading
I0510 17:31:53.093371 670298 kubeadm.go:926] updating node { 192.168.76.2 8443 v1.33.0 docker true true} ...
I0510 17:31:53.093460 670298 kubeadm.go:938] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.33.0/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=scheduled-stop-579771 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.76.2
[Install]
config:
{KubernetesVersion:v1.33.0 ClusterName:scheduled-stop-579771 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0510 17:31:53.093526 670298 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0510 17:31:53.145579 670298 cni.go:84] Creating CNI manager for ""
I0510 17:31:53.145595 670298 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0510 17:31:53.145605 670298 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0510 17:31:53.145623 670298 kubeadm.go:189] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.76.2 APIServerPort:8443 KubernetesVersion:v1.33.0 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:scheduled-stop-579771 NodeName:scheduled-stop-579771 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.76.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.76.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPo
dPath:/etc/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0510 17:31:53.145754 670298 kubeadm.go:195] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta4
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.76.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "scheduled-stop-579771"
kubeletExtraArgs:
- name: "node-ip"
value: "192.168.76.2"
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta4
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.76.2"]
extraArgs:
- name: "enable-admission-plugins"
value: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
- name: "allocate-node-cidrs"
value: "true"
- name: "leader-elect"
value: "false"
scheduler:
extraArgs:
- name: "leader-elect"
value: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
- name: "proxy-refresh-interval"
value: "70000"
kubernetesVersion: v1.33.0
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0510 17:31:53.145827 670298 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.33.0
I0510 17:31:53.155192 670298 binaries.go:44] Found k8s binaries, skipping transfer
I0510 17:31:53.155255 670298 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0510 17:31:53.164129 670298 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (320 bytes)
I0510 17:31:53.183458 670298 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0510 17:31:53.201566 670298 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2299 bytes)
I0510 17:31:53.220454 670298 ssh_runner.go:195] Run: grep 192.168.76.2 control-plane.minikube.internal$ /etc/hosts
I0510 17:31:53.223936 670298 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.76.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0510 17:31:53.235063 670298 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0510 17:31:53.315794 670298 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0510 17:31:53.331425 670298 certs.go:68] Setting up /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771 for IP: 192.168.76.2
I0510 17:31:53.331435 670298 certs.go:194] generating shared ca certs ...
I0510 17:31:53.331450 670298 certs.go:226] acquiring lock for ca certs: {Name:mk6fc393a4c26de0104f16184748c2e2225128de Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0510 17:31:53.331577 670298 certs.go:235] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/20720-453270/.minikube/ca.key
I0510 17:31:53.331618 670298 certs.go:235] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/20720-453270/.minikube/proxy-client-ca.key
I0510 17:31:53.331623 670298 certs.go:256] generating profile certs ...
I0510 17:31:53.331674 670298 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/client.key
I0510 17:31:53.331683 670298 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/client.crt with IP's: []
I0510 17:31:54.090897 670298 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/client.crt ...
I0510 17:31:54.090913 670298 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/client.crt: {Name:mk7b6cd84324705df809f4ccb1f51c06f6b47973 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0510 17:31:54.091121 670298 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/client.key ...
I0510 17:31:54.091129 670298 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/client.key: {Name:mkfad0cd7031241c9da1a242e8a46220ae4f1bce Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0510 17:31:54.091220 670298 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/apiserver.key.2f87821e
I0510 17:31:54.091233 670298 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/apiserver.crt.2f87821e with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.76.2]
I0510 17:31:54.637098 670298 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/apiserver.crt.2f87821e ...
I0510 17:31:54.637115 670298 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/apiserver.crt.2f87821e: {Name:mk6a77b43ef4f5d70fbbc3e8c9662cc10f359255 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0510 17:31:54.637311 670298 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/apiserver.key.2f87821e ...
I0510 17:31:54.637322 670298 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/apiserver.key.2f87821e: {Name:mk475cd5b620d556d22e334592f875567ca5945c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0510 17:31:54.637419 670298 certs.go:381] copying /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/apiserver.crt.2f87821e -> /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/apiserver.crt
I0510 17:31:54.637492 670298 certs.go:385] copying /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/apiserver.key.2f87821e -> /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/apiserver.key
I0510 17:31:54.637548 670298 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/proxy-client.key
I0510 17:31:54.637560 670298 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/proxy-client.crt with IP's: []
I0510 17:31:55.822792 670298 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/proxy-client.crt ...
I0510 17:31:55.822807 670298 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/proxy-client.crt: {Name:mk83278b74bd6e2081482cebcb4f4c995eec8f0e Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0510 17:31:55.822995 670298 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/proxy-client.key ...
I0510 17:31:55.823003 670298 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/proxy-client.key: {Name:mk1f6cce48d23bc6a107c60d69cf80dc65ef0ef3 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0510 17:31:55.823232 670298 certs.go:484] found cert: /home/jenkins/minikube-integration/20720-453270/.minikube/certs/458646.pem (1338 bytes)
W0510 17:31:55.823267 670298 certs.go:480] ignoring /home/jenkins/minikube-integration/20720-453270/.minikube/certs/458646_empty.pem, impossibly tiny 0 bytes
I0510 17:31:55.823276 670298 certs.go:484] found cert: /home/jenkins/minikube-integration/20720-453270/.minikube/certs/ca-key.pem (1679 bytes)
I0510 17:31:55.823301 670298 certs.go:484] found cert: /home/jenkins/minikube-integration/20720-453270/.minikube/certs/ca.pem (1082 bytes)
I0510 17:31:55.823323 670298 certs.go:484] found cert: /home/jenkins/minikube-integration/20720-453270/.minikube/certs/cert.pem (1123 bytes)
I0510 17:31:55.823348 670298 certs.go:484] found cert: /home/jenkins/minikube-integration/20720-453270/.minikube/certs/key.pem (1675 bytes)
I0510 17:31:55.823389 670298 certs.go:484] found cert: /home/jenkins/minikube-integration/20720-453270/.minikube/files/etc/ssl/certs/4586462.pem (1708 bytes)
I0510 17:31:55.824025 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0510 17:31:55.849516 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I0510 17:31:55.874184 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0510 17:31:55.898795 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0510 17:31:55.923695 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1432 bytes)
I0510 17:31:55.948277 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0510 17:31:55.972621 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0510 17:31:56.000725 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/profiles/scheduled-stop-579771/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0510 17:31:56.025524 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/certs/458646.pem --> /usr/share/ca-certificates/458646.pem (1338 bytes)
I0510 17:31:56.050820 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/files/etc/ssl/certs/4586462.pem --> /usr/share/ca-certificates/4586462.pem (1708 bytes)
I0510 17:31:56.075795 670298 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20720-453270/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0510 17:31:56.100570 670298 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0510 17:31:56.118987 670298 ssh_runner.go:195] Run: openssl version
I0510 17:31:56.124758 670298 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/458646.pem && ln -fs /usr/share/ca-certificates/458646.pem /etc/ssl/certs/458646.pem"
I0510 17:31:56.134053 670298 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/458646.pem
I0510 17:31:56.137480 670298 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 May 10 17:01 /usr/share/ca-certificates/458646.pem
I0510 17:31:56.137533 670298 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/458646.pem
I0510 17:31:56.144426 670298 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/458646.pem /etc/ssl/certs/51391683.0"
I0510 17:31:56.153837 670298 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/4586462.pem && ln -fs /usr/share/ca-certificates/4586462.pem /etc/ssl/certs/4586462.pem"
I0510 17:31:56.163534 670298 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/4586462.pem
I0510 17:31:56.166986 670298 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 May 10 17:01 /usr/share/ca-certificates/4586462.pem
I0510 17:31:56.167042 670298 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/4586462.pem
I0510 17:31:56.174888 670298 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/4586462.pem /etc/ssl/certs/3ec20f2e.0"
I0510 17:31:56.184619 670298 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0510 17:31:56.193646 670298 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0510 17:31:56.197174 670298 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 May 10 16:54 /usr/share/ca-certificates/minikubeCA.pem
I0510 17:31:56.197229 670298 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0510 17:31:56.204249 670298 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0510 17:31:56.213314 670298 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0510 17:31:56.216714 670298 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0510 17:31:56.216755 670298 kubeadm.go:392] StartCluster: {Name:scheduled-stop-579771 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1746731792-20718@sha256:074d9afa1e8827ea0e101248fc55098d304814b5d8bf485882a81afc90084155 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.33.0 ClusterName:scheduled-stop-579771 Namespace:default APIServerHAVIP: APIServerName:minikubeC
A APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.33.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP
: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0510 17:31:56.216869 670298 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0510 17:31:56.235082 670298 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0510 17:31:56.244069 670298 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0510 17:31:56.252662 670298 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0510 17:31:56.252721 670298 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0510 17:31:56.261347 670298 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0510 17:31:56.261355 670298 kubeadm.go:157] found existing configuration files:
I0510 17:31:56.261403 670298 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0510 17:31:56.270177 670298 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0510 17:31:56.270233 670298 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0510 17:31:56.278732 670298 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0510 17:31:56.287686 670298 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0510 17:31:56.287750 670298 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0510 17:31:56.296380 670298 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0510 17:31:56.305218 670298 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0510 17:31:56.305282 670298 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0510 17:31:56.314318 670298 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0510 17:31:56.323493 670298 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0510 17:31:56.323549 670298 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0510 17:31:56.332190 670298 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.33.0:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0510 17:31:56.398088 670298 kubeadm.go:310] [WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2
I0510 17:31:56.398322 670298 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1083-aws\n", err: exit status 1
I0510 17:31:56.463979 670298 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0510 17:32:13.141846 670298 kubeadm.go:310] [init] Using Kubernetes version: v1.33.0
I0510 17:32:13.141898 670298 kubeadm.go:310] [preflight] Running pre-flight checks
I0510 17:32:13.141992 670298 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0510 17:32:13.142053 670298 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1083-aws[0m
I0510 17:32:13.142086 670298 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0510 17:32:13.142145 670298 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0510 17:32:13.142212 670298 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0510 17:32:13.142289 670298 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0510 17:32:13.142339 670298 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0510 17:32:13.142398 670298 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0510 17:32:13.142447 670298 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0510 17:32:13.142491 670298 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0510 17:32:13.142538 670298 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0510 17:32:13.142583 670298 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0510 17:32:13.142667 670298 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0510 17:32:13.142759 670298 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0510 17:32:13.142846 670298 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0510 17:32:13.142906 670298 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0510 17:32:13.145927 670298 out.go:235] - Generating certificates and keys ...
I0510 17:32:13.146018 670298 kubeadm.go:310] [certs] Using existing ca certificate authority
I0510 17:32:13.146088 670298 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0510 17:32:13.146155 670298 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0510 17:32:13.146217 670298 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0510 17:32:13.146276 670298 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0510 17:32:13.146324 670298 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0510 17:32:13.146376 670298 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0510 17:32:13.146497 670298 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [localhost scheduled-stop-579771] and IPs [192.168.76.2 127.0.0.1 ::1]
I0510 17:32:13.146548 670298 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0510 17:32:13.146689 670298 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [localhost scheduled-stop-579771] and IPs [192.168.76.2 127.0.0.1 ::1]
I0510 17:32:13.146752 670298 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0510 17:32:13.146813 670298 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0510 17:32:13.146856 670298 kubeadm.go:310] [certs] Generating "sa" key and public key
I0510 17:32:13.146911 670298 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0510 17:32:13.146960 670298 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0510 17:32:13.147014 670298 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0510 17:32:13.147068 670298 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0510 17:32:13.147129 670298 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0510 17:32:13.147182 670298 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0510 17:32:13.147261 670298 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0510 17:32:13.147327 670298 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0510 17:32:13.152205 670298 out.go:235] - Booting up control plane ...
I0510 17:32:13.152356 670298 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0510 17:32:13.152433 670298 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0510 17:32:13.152497 670298 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0510 17:32:13.152597 670298 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0510 17:32:13.152680 670298 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0510 17:32:13.152717 670298 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0510 17:32:13.152844 670298 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0510 17:32:13.152945 670298 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0510 17:32:13.153001 670298 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 1.001739286s
I0510 17:32:13.153119 670298 kubeadm.go:310] [control-plane-check] Waiting for healthy control plane components. This can take up to 4m0s
I0510 17:32:13.153205 670298 kubeadm.go:310] [control-plane-check] Checking kube-apiserver at https://192.168.76.2:8443/livez
I0510 17:32:13.153300 670298 kubeadm.go:310] [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz
I0510 17:32:13.153377 670298 kubeadm.go:310] [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez
I0510 17:32:13.153449 670298 kubeadm.go:310] [control-plane-check] kube-controller-manager is healthy after 5.3331108s
I0510 17:32:13.153514 670298 kubeadm.go:310] [control-plane-check] kube-scheduler is healthy after 6.379358732s
I0510 17:32:13.153579 670298 kubeadm.go:310] [control-plane-check] kube-apiserver is healthy after 7.501184941s
I0510 17:32:13.153681 670298 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0510 17:32:13.153803 670298 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0510 17:32:13.153859 670298 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0510 17:32:13.154058 670298 kubeadm.go:310] [mark-control-plane] Marking the node scheduled-stop-579771 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0510 17:32:13.154112 670298 kubeadm.go:310] [bootstrap-token] Using token: epqwd2.fkgxfbivnzbmy3fr
I0510 17:32:13.157093 670298 out.go:235] - Configuring RBAC rules ...
I0510 17:32:13.157224 670298 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0510 17:32:13.157319 670298 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0510 17:32:13.157457 670298 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0510 17:32:13.157607 670298 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0510 17:32:13.157748 670298 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0510 17:32:13.157837 670298 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0510 17:32:13.157950 670298 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0510 17:32:13.157991 670298 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0510 17:32:13.158036 670298 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0510 17:32:13.158055 670298 kubeadm.go:310]
I0510 17:32:13.158114 670298 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0510 17:32:13.158117 670298 kubeadm.go:310]
I0510 17:32:13.158192 670298 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0510 17:32:13.158196 670298 kubeadm.go:310]
I0510 17:32:13.158219 670298 kubeadm.go:310] mkdir -p $HOME/.kube
I0510 17:32:13.158279 670298 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0510 17:32:13.158328 670298 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0510 17:32:13.158332 670298 kubeadm.go:310]
I0510 17:32:13.158384 670298 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0510 17:32:13.158388 670298 kubeadm.go:310]
I0510 17:32:13.158434 670298 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0510 17:32:13.158437 670298 kubeadm.go:310]
I0510 17:32:13.158496 670298 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0510 17:32:13.158569 670298 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0510 17:32:13.158636 670298 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0510 17:32:13.158639 670298 kubeadm.go:310]
I0510 17:32:13.158834 670298 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0510 17:32:13.158909 670298 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0510 17:32:13.158913 670298 kubeadm.go:310]
I0510 17:32:13.159009 670298 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token epqwd2.fkgxfbivnzbmy3fr \
I0510 17:32:13.159111 670298 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:249dd50f996888b2961e66f106be33678f92233172413c745bf009bcf93efb1b \
I0510 17:32:13.159131 670298 kubeadm.go:310] --control-plane
I0510 17:32:13.159134 670298 kubeadm.go:310]
I0510 17:32:13.159217 670298 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0510 17:32:13.159220 670298 kubeadm.go:310]
I0510 17:32:13.159300 670298 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token epqwd2.fkgxfbivnzbmy3fr \
I0510 17:32:13.159400 670298 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:249dd50f996888b2961e66f106be33678f92233172413c745bf009bcf93efb1b
I0510 17:32:13.159421 670298 cni.go:84] Creating CNI manager for ""
I0510 17:32:13.159435 670298 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0510 17:32:13.162541 670298 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0510 17:32:13.166031 670298 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0510 17:32:13.174927 670298 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0510 17:32:13.193431 670298 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0510 17:32:13.193568 670298 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.33.0/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0510 17:32:13.193646 670298 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.33.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes scheduled-stop-579771 minikube.k8s.io/updated_at=2025_05_10T17_32_13_0700 minikube.k8s.io/version=v1.35.0 minikube.k8s.io/commit=e96c83983357cd8557f3cdfe077a25cc73d485a4 minikube.k8s.io/name=scheduled-stop-579771 minikube.k8s.io/primary=true
I0510 17:32:13.208365 670298 ops.go:34] apiserver oom_adj: -16
I0510 17:32:13.333812 670298 kubeadm.go:1105] duration metric: took 140.295478ms to wait for elevateKubeSystemPrivileges
I0510 17:32:13.338481 670298 kubeadm.go:394] duration metric: took 17.121723187s to StartCluster
I0510 17:32:13.338507 670298 settings.go:142] acquiring lock: {Name:mk73409f27eabc06be6e26db0de886eff4b4bb84 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0510 17:32:13.338582 670298 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/20720-453270/kubeconfig
I0510 17:32:13.339283 670298 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20720-453270/kubeconfig: {Name:mk61b0fc2c46863342156febfe0b8261bb6e49c3 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0510 17:32:13.339494 670298 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.33.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0510 17:32:13.339594 670298 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.33.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0510 17:32:13.339829 670298 config.go:182] Loaded profile config "scheduled-stop-579771": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.33.0
I0510 17:32:13.339859 670298 addons.go:511] enable addons start: toEnable=map[ambassador:false amd-gpu-device-plugin:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:false volcano:false volumesnapshots:false yakd:false]
I0510 17:32:13.339914 670298 addons.go:69] Setting storage-provisioner=true in profile "scheduled-stop-579771"
I0510 17:32:13.339926 670298 addons.go:238] Setting addon storage-provisioner=true in "scheduled-stop-579771"
I0510 17:32:13.339952 670298 addons.go:69] Setting default-storageclass=true in profile "scheduled-stop-579771"
I0510 17:32:13.339964 670298 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "scheduled-stop-579771"
I0510 17:32:13.340304 670298 cli_runner.go:164] Run: docker container inspect scheduled-stop-579771 --format={{.State.Status}}
I0510 17:32:13.340482 670298 host.go:66] Checking if "scheduled-stop-579771" exists ...
I0510 17:32:13.340886 670298 cli_runner.go:164] Run: docker container inspect scheduled-stop-579771 --format={{.State.Status}}
I0510 17:32:13.343182 670298 out.go:177] * Verifying Kubernetes components...
I0510 17:32:13.350886 670298 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0510 17:32:13.378511 670298 addons.go:238] Setting addon default-storageclass=true in "scheduled-stop-579771"
I0510 17:32:13.378596 670298 host.go:66] Checking if "scheduled-stop-579771" exists ...
I0510 17:32:13.379119 670298 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0510 17:32:13.379195 670298 cli_runner.go:164] Run: docker container inspect scheduled-stop-579771 --format={{.State.Status}}
I0510 17:32:13.382136 670298 addons.go:435] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0510 17:32:13.382148 670298 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0510 17:32:13.382215 670298 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-579771
I0510 17:32:13.417920 670298 addons.go:435] installing /etc/kubernetes/addons/storageclass.yaml
I0510 17:32:13.417933 670298 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0510 17:32:13.417998 670298 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-579771
I0510 17:32:13.425431 670298 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33367 SSHKeyPath:/home/jenkins/minikube-integration/20720-453270/.minikube/machines/scheduled-stop-579771/id_rsa Username:docker}
I0510 17:32:13.452433 670298 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33367 SSHKeyPath:/home/jenkins/minikube-integration/20720-453270/.minikube/machines/scheduled-stop-579771/id_rsa Username:docker}
I0510 17:32:13.555308 670298 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.33.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.76.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.33.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0510 17:32:13.568843 670298 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0510 17:32:13.621441 670298 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.33.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0510 17:32:13.700090 670298 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.33.0/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0510 17:32:14.088196 670298 start.go:971] {"host.minikube.internal": 192.168.76.1} host record injected into CoreDNS's ConfigMap
I0510 17:32:14.088960 670298 api_server.go:52] waiting for apiserver process to appear ...
I0510 17:32:14.089015 670298 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0510 17:32:14.269025 670298 api_server.go:72] duration metric: took 929.505745ms to wait for apiserver process to appear ...
I0510 17:32:14.269036 670298 api_server.go:88] waiting for apiserver healthz status ...
I0510 17:32:14.269052 670298 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
I0510 17:32:14.278011 670298 api_server.go:279] https://192.168.76.2:8443/healthz returned 200:
ok
I0510 17:32:14.279097 670298 api_server.go:141] control plane version: v1.33.0
I0510 17:32:14.279122 670298 api_server.go:131] duration metric: took 10.079757ms to wait for apiserver health ...
I0510 17:32:14.279130 670298 system_pods.go:43] waiting for kube-system pods to appear ...
I0510 17:32:14.283277 670298 system_pods.go:59] 5 kube-system pods found
I0510 17:32:14.283295 670298 system_pods.go:61] "etcd-scheduled-stop-579771" [0f8bb30e-e1d0-449e-9548-ac532383560b] Running / Ready:ContainersNotReady (containers with unready status: [etcd]) / ContainersReady:ContainersNotReady (containers with unready status: [etcd])
I0510 17:32:14.283304 670298 system_pods.go:61] "kube-apiserver-scheduled-stop-579771" [3ae8bcbd-4132-490e-8561-b2036b2867cf] Running / Ready:ContainersNotReady (containers with unready status: [kube-apiserver]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-apiserver])
I0510 17:32:14.283311 670298 system_pods.go:61] "kube-controller-manager-scheduled-stop-579771" [c173c24e-9104-4a53-b1eb-a2da3c90da1e] Running / Ready:ContainersNotReady (containers with unready status: [kube-controller-manager]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-controller-manager])
I0510 17:32:14.283317 670298 system_pods.go:61] "kube-scheduler-scheduled-stop-579771" [d851561f-901b-4e10-af5d-9e34caca5bcc] Running / Ready:ContainersNotReady (containers with unready status: [kube-scheduler]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-scheduler])
I0510 17:32:14.283321 670298 system_pods.go:61] "storage-provisioner" [f0ef7b13-41ec-4158-be21-09aebcf77eeb] Pending: PodScheduled:Unschedulable (0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.)
I0510 17:32:14.283326 670298 system_pods.go:74] duration metric: took 4.191136ms to wait for pod list to return data ...
I0510 17:32:14.283359 670298 kubeadm.go:578] duration metric: took 943.821241ms to wait for: map[apiserver:true system_pods:true]
I0510 17:32:14.283372 670298 node_conditions.go:102] verifying NodePressure condition ...
I0510 17:32:14.283388 670298 out.go:177] * Enabled addons: storage-provisioner, default-storageclass
I0510 17:32:14.285686 670298 node_conditions.go:122] node storage ephemeral capacity is 203034800Ki
I0510 17:32:14.285703 670298 node_conditions.go:123] node cpu capacity is 2
I0510 17:32:14.285715 670298 node_conditions.go:105] duration metric: took 2.337287ms to run NodePressure ...
I0510 17:32:14.285725 670298 start.go:241] waiting for startup goroutines ...
I0510 17:32:14.286429 670298 addons.go:514] duration metric: took 946.56373ms for enable addons: enabled=[storage-provisioner default-storageclass]
I0510 17:32:14.591567 670298 kapi.go:214] "coredns" deployment in "kube-system" namespace and "scheduled-stop-579771" context rescaled to 1 replicas
I0510 17:32:14.591597 670298 start.go:246] waiting for cluster config update ...
I0510 17:32:14.591607 670298 start.go:255] writing updated cluster config ...
I0510 17:32:14.591898 670298 ssh_runner.go:195] Run: rm -f paused
I0510 17:32:14.654699 670298 start.go:607] kubectl: 1.33.0, cluster: 1.33.0 (minor skew: 0)
I0510 17:32:14.657784 670298 out.go:177] * Done! kubectl is now configured to use "scheduled-stop-579771" cluster and "default" namespace by default
==> Docker <==
May 10 17:31:52 scheduled-stop-579771 dockerd[1397]: time="2025-05-10T17:31:52.332774378Z" level=warning msg="Error (Unable to complete atomic operation, key modified) deleting object [endpoint_count 4fff1700fc6537979c0f55ff8a2c1bb544332cab0ddd1e264238a93b85cd7a78], retrying...."
May 10 17:31:52 scheduled-stop-579771 dockerd[1397]: time="2025-05-10T17:31:52.367493424Z" level=info msg="Loading containers: done."
May 10 17:31:52 scheduled-stop-579771 dockerd[1397]: time="2025-05-10T17:31:52.377704857Z" level=info msg="Docker daemon" commit=01f442b containerd-snapshotter=false storage-driver=overlay2 version=28.1.1
May 10 17:31:52 scheduled-stop-579771 dockerd[1397]: time="2025-05-10T17:31:52.377782190Z" level=info msg="Initializing buildkit"
May 10 17:31:52 scheduled-stop-579771 dockerd[1397]: time="2025-05-10T17:31:52.396344215Z" level=info msg="Completed buildkit initialization"
May 10 17:31:52 scheduled-stop-579771 dockerd[1397]: time="2025-05-10T17:31:52.401871818Z" level=info msg="Daemon has completed initialization"
May 10 17:31:52 scheduled-stop-579771 dockerd[1397]: time="2025-05-10T17:31:52.402118571Z" level=info msg="API listen on [::]:2376"
May 10 17:31:52 scheduled-stop-579771 systemd[1]: Started Docker Application Container Engine.
May 10 17:31:52 scheduled-stop-579771 dockerd[1397]: time="2025-05-10T17:31:52.402240105Z" level=info msg="API listen on /var/run/docker.sock"
May 10 17:31:52 scheduled-stop-579771 systemd[1]: Starting CRI Interface for Docker Application Container Engine...
May 10 17:31:52 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:31:52Z" level=info msg="Starting cri-dockerd dev (HEAD)"
May 10 17:31:52 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:31:52Z" level=info msg="Connecting to docker on the Endpoint unix:///var/run/docker.sock"
May 10 17:31:52 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:31:52Z" level=info msg="Start docker client with request timeout 0s"
May 10 17:31:52 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:31:52Z" level=info msg="Hairpin mode is set to hairpin-veth"
May 10 17:31:52 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:31:52Z" level=info msg="Loaded network plugin cni"
May 10 17:31:52 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:31:52Z" level=info msg="Docker cri networking managed by network plugin cni"
May 10 17:31:52 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:31:52Z" level=info msg="Setting cgroupDriver cgroupfs"
May 10 17:31:52 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:31:52Z" level=info msg="Docker cri received runtime config &RuntimeConfig{NetworkConfig:&NetworkConfig{PodCidr:,},}"
May 10 17:31:52 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:31:52Z" level=info msg="Starting the GRPC backend for the Docker CRI interface."
May 10 17:31:52 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:31:52Z" level=info msg="Start cri-dockerd grpc backend"
May 10 17:31:52 scheduled-stop-579771 systemd[1]: Started CRI Interface for Docker Application Container Engine.
May 10 17:32:04 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:32:04Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/477f644987bf103fdb808048e81f505dbc34ad2d4326388cabab7e5d1605b0b2/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options edns0 trust-ad ndots:0]"
May 10 17:32:04 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:32:04Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/051156cdcd3171ea3d86f4f222c86e67c41097843196565da21573028cbcc2b1/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options edns0 trust-ad ndots:0]"
May 10 17:32:04 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:32:04Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/f85a184b75617ba7fcf302687345357cc656bd736dc33ee8886ff07b83231e85/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options edns0 trust-ad ndots:0]"
May 10 17:32:04 scheduled-stop-579771 cri-dockerd[1695]: time="2025-05-10T17:32:04Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/96629f3e1ca7dba002b81d840379bc3a3d147a2f0260be4ce5611fa6e8b82fe4/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options edns0 trust-ad ndots:0]"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
59de35a41ff30 f8984990ac26b 12 seconds ago Running kube-scheduler 0 96629f3e1ca7d kube-scheduler-scheduled-stop-579771
4fd152ef84f56 5372350fd0a1e 12 seconds ago Running kube-controller-manager 0 f85a184b75617 kube-controller-manager-scheduled-stop-579771
7e14b8cb7dab6 61f3acc54bb62 12 seconds ago Running kube-apiserver 0 051156cdcd317 kube-apiserver-scheduled-stop-579771
6be2d0c7d6965 31747a36ce712 12 seconds ago Running etcd 0 477f644987bf1 etcd-scheduled-stop-579771
==> describe nodes <==
Name: scheduled-stop-579771
Roles: control-plane
Labels: beta.kubernetes.io/arch=arm64
beta.kubernetes.io/os=linux
kubernetes.io/arch=arm64
kubernetes.io/hostname=scheduled-stop-579771
kubernetes.io/os=linux
minikube.k8s.io/commit=e96c83983357cd8557f3cdfe077a25cc73d485a4
minikube.k8s.io/name=scheduled-stop-579771
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2025_05_10T17_32_13_0700
minikube.k8s.io/version=v1.35.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Sat, 10 May 2025 17:32:09 +0000
Taints: node.kubernetes.io/not-ready:NoSchedule
Unschedulable: false
Lease:
HolderIdentity: scheduled-stop-579771
AcquireTime: <unset>
RenewTime: Sat, 10 May 2025 17:32:12 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Sat, 10 May 2025 17:32:12 +0000 Sat, 10 May 2025 17:32:05 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Sat, 10 May 2025 17:32:12 +0000 Sat, 10 May 2025 17:32:05 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Sat, 10 May 2025 17:32:12 +0000 Sat, 10 May 2025 17:32:05 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Sat, 10 May 2025 17:32:12 +0000 Sat, 10 May 2025 17:32:09 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.76.2
Hostname: scheduled-stop-579771
Capacity:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022296Ki
pods: 110
Allocatable:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022296Ki
pods: 110
System Info:
Machine ID: b035532049cb42ad81bf64f01d3d2e6f
System UUID: d282a8e9-7121-453b-beb3-3c72f3cb5fac
Boot ID: fb90d26d-c40d-4d7a-8f91-70daff9c23aa
Kernel Version: 5.15.0-1083-aws
OS Image: Ubuntu 22.04.5 LTS
Operating System: linux
Architecture: arm64
Container Runtime Version: docker://28.1.1
Kubelet Version: v1.33.0
Kube-Proxy Version:
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (4 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
kube-system etcd-scheduled-stop-579771 100m (5%) 0 (0%) 100Mi (1%) 0 (0%) 4s
kube-system kube-apiserver-scheduled-stop-579771 250m (12%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-controller-manager-scheduled-stop-579771 200m (10%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-scheduler-scheduled-stop-579771 100m (5%) 0 (0%) 0 (0%) 0 (0%) 4s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 650m (32%) 0 (0%)
memory 100Mi (1%) 0 (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
hugepages-32Mi 0 (0%) 0 (0%)
hugepages-64Ki 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal NodeHasSufficientMemory 13s (x8 over 13s) kubelet Node scheduled-stop-579771 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 13s (x8 over 13s) kubelet Node scheduled-stop-579771 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 13s (x7 over 13s) kubelet Node scheduled-stop-579771 status is now: NodeHasSufficientPID
Normal NodeAllocatableEnforced 13s kubelet Updated Node Allocatable limit across pods
Normal Starting 4s kubelet Starting kubelet.
Warning CgroupV1 4s kubelet cgroup v1 support is in maintenance mode, please migrate to cgroup v2
Normal NodeAllocatableEnforced 4s kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 4s kubelet Node scheduled-stop-579771 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 4s kubelet Node scheduled-stop-579771 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 4s kubelet Node scheduled-stop-579771 status is now: NodeHasSufficientPID
==> dmesg <==
[May10 16:06] kmem.limit_in_bytes is deprecated and will be removed. Please report your usecase to linux-mm@kvack.org if you depend on this functionality.
[May10 16:21] systemd-journald[223]: Failed to send stream file descriptor to service manager: Connection refused
==> etcd [6be2d0c7d696] <==
{"level":"info","ts":"2025-05-10T17:32:04.801699Z","caller":"embed/etcd.go:762","msg":"starting with client TLS","tls-info":"cert = /var/lib/minikube/certs/etcd/server.crt, key = /var/lib/minikube/certs/etcd/server.key, client-cert=, client-key=, trusted-ca = /var/lib/minikube/certs/etcd/ca.crt, client-cert-auth = true, crl-file = ","cipher-suites":[]}
{"level":"info","ts":"2025-05-10T17:32:04.801895Z","caller":"embed/etcd.go:633","msg":"serving peer traffic","address":"192.168.76.2:2380"}
{"level":"info","ts":"2025-05-10T17:32:04.801921Z","caller":"embed/etcd.go:603","msg":"cmux::serve","address":"192.168.76.2:2380"}
{"level":"info","ts":"2025-05-10T17:32:04.803126Z","caller":"embed/etcd.go:908","msg":"serving metrics","address":"http://127.0.0.1:2381"}
{"level":"info","ts":"2025-05-10T17:32:04.803084Z","caller":"embed/etcd.go:292","msg":"now serving peer/client/metrics","local-member-id":"ea7e25599daad906","initial-advertise-peer-urls":["https://192.168.76.2:2380"],"listen-peer-urls":["https://192.168.76.2:2380"],"advertise-client-urls":["https://192.168.76.2:2379"],"listen-client-urls":["https://127.0.0.1:2379","https://192.168.76.2:2379"],"listen-metrics-urls":["http://127.0.0.1:2381"]}
{"level":"info","ts":"2025-05-10T17:32:05.786702Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 is starting a new election at term 1"}
{"level":"info","ts":"2025-05-10T17:32:05.786824Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 became pre-candidate at term 1"}
{"level":"info","ts":"2025-05-10T17:32:05.786949Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 received MsgPreVoteResp from ea7e25599daad906 at term 1"}
{"level":"info","ts":"2025-05-10T17:32:05.787008Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 became candidate at term 2"}
{"level":"info","ts":"2025-05-10T17:32:05.787055Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 received MsgVoteResp from ea7e25599daad906 at term 2"}
{"level":"info","ts":"2025-05-10T17:32:05.787120Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 became leader at term 2"}
{"level":"info","ts":"2025-05-10T17:32:05.787149Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: ea7e25599daad906 elected leader ea7e25599daad906 at term 2"}
{"level":"info","ts":"2025-05-10T17:32:05.790892Z","caller":"etcdserver/server.go:2144","msg":"published local member to cluster through raft","local-member-id":"ea7e25599daad906","local-member-attributes":"{Name:scheduled-stop-579771 ClientURLs:[https://192.168.76.2:2379]}","request-path":"/0/members/ea7e25599daad906/attributes","cluster-id":"6f20f2c4b2fb5f8a","publish-timeout":"7s"}
{"level":"info","ts":"2025-05-10T17:32:05.790997Z","caller":"embed/serve.go:124","msg":"ready to serve client requests"}
{"level":"info","ts":"2025-05-10T17:32:05.791372Z","caller":"etcdserver/server.go:2697","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2025-05-10T17:32:05.792098Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2025-05-10T17:32:05.793031Z","caller":"embed/serve.go:275","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2025-05-10T17:32:05.793480Z","caller":"membership/cluster.go:587","msg":"set initial cluster version","cluster-id":"6f20f2c4b2fb5f8a","local-member-id":"ea7e25599daad906","cluster-version":"3.5"}
{"level":"info","ts":"2025-05-10T17:32:05.793631Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2025-05-10T17:32:05.793716Z","caller":"etcdserver/server.go:2721","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2025-05-10T17:32:05.794675Z","caller":"embed/serve.go:124","msg":"ready to serve client requests"}
{"level":"info","ts":"2025-05-10T17:32:05.803315Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2025-05-10T17:32:05.807439Z","caller":"embed/serve.go:275","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.76.2:2379"}
{"level":"info","ts":"2025-05-10T17:32:05.807803Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2025-05-10T17:32:05.807945Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
==> kernel <==
17:32:16 up 2:14, 0 users, load average: 1.53, 1.78, 2.45
Linux scheduled-stop-579771 5.15.0-1083-aws #90~20.04.1-Ubuntu SMP Tue Apr 22 09:59:55 UTC 2025 aarch64 aarch64 aarch64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.5 LTS"
==> kube-apiserver [7e14b8cb7dab] <==
I0510 17:32:09.546199 1 shared_informer.go:357] "Caches are synced" controller="kubernetes-service-cidr-controller"
I0510 17:32:09.546439 1 default_servicecidr_controller.go:165] Creating default ServiceCIDR with CIDRs: [10.96.0.0/12]
I0510 17:32:09.546956 1 cache.go:39] Caches are synced for LocalAvailability controller
I0510 17:32:09.547131 1 cache.go:39] Caches are synced for RemoteAvailability controller
I0510 17:32:09.547331 1 apf_controller.go:382] Running API Priority and Fairness config worker
I0510 17:32:09.547450 1 apf_controller.go:385] Running API Priority and Fairness periodic rebalancing process
I0510 17:32:09.547691 1 shared_informer.go:357] "Caches are synced" controller="configmaps"
I0510 17:32:09.564530 1 cidrallocator.go:301] created ClusterIP allocator for Service CIDR 10.96.0.0/12
I0510 17:32:09.568024 1 default_servicecidr_controller.go:214] Setting default ServiceCIDR condition Ready to True
I0510 17:32:09.591812 1 cidrallocator.go:277] updated ClusterIP allocator for Service CIDR 10.96.0.0/12
I0510 17:32:09.595249 1 controller.go:667] quota admission added evaluator for: leases.coordination.k8s.io
I0510 17:32:09.595303 1 default_servicecidr_controller.go:136] Shutting down kubernetes-service-cidr-controller
I0510 17:32:10.342572 1 storage_scheduling.go:95] created PriorityClass system-node-critical with value 2000001000
I0510 17:32:10.348040 1 storage_scheduling.go:95] created PriorityClass system-cluster-critical with value 2000000000
I0510 17:32:10.348060 1 storage_scheduling.go:111] all system priority classes are created successfully or already exist.
I0510 17:32:11.049114 1 controller.go:667] quota admission added evaluator for: roles.rbac.authorization.k8s.io
I0510 17:32:11.100878 1 controller.go:667] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
I0510 17:32:11.264147 1 alloc.go:328] "allocated clusterIPs" service="default/kubernetes" clusterIPs={"IPv4":"10.96.0.1"}
W0510 17:32:11.273435 1 lease.go:265] Resetting endpoints for master service "kubernetes" to [192.168.76.2]
I0510 17:32:11.274740 1 controller.go:667] quota admission added evaluator for: endpoints
I0510 17:32:11.280213 1 controller.go:667] quota admission added evaluator for: endpointslices.discovery.k8s.io
I0510 17:32:11.367139 1 controller.go:667] quota admission added evaluator for: serviceaccounts
I0510 17:32:12.544865 1 controller.go:667] quota admission added evaluator for: deployments.apps
I0510 17:32:12.562443 1 alloc.go:328] "allocated clusterIPs" service="kube-system/kube-dns" clusterIPs={"IPv4":"10.96.0.10"}
I0510 17:32:12.729498 1 controller.go:667] quota admission added evaluator for: daemonsets.apps
==> kube-controller-manager [4fd152ef84f5] <==
I0510 17:32:16.264221 1 shared_informer.go:357] "Caches are synced" controller="cronjob"
I0510 17:32:16.324022 1 shared_informer.go:357] "Caches are synced" controller="ephemeral"
I0510 17:32:16.336315 1 shared_informer.go:357] "Caches are synced" controller="endpoint"
I0510 17:32:16.336856 1 shared_informer.go:357] "Caches are synced" controller="resource quota"
I0510 17:32:16.346169 1 shared_informer.go:357] "Caches are synced" controller="ReplicaSet"
I0510 17:32:16.362821 1 shared_informer.go:357] "Caches are synced" controller="taint"
I0510 17:32:16.363264 1 node_lifecycle_controller.go:1221] "Initializing eviction metric for zone" logger="node-lifecycle-controller" zone=""
I0510 17:32:16.363334 1 node_lifecycle_controller.go:873] "Missing timestamp for Node. Assuming now as a timestamp" logger="node-lifecycle-controller" node="scheduled-stop-579771"
I0510 17:32:16.363367 1 node_lifecycle_controller.go:1067] "Controller detected that zone is now in new state" logger="node-lifecycle-controller" zone="" newState="Normal"
I0510 17:32:16.363747 1 shared_informer.go:357] "Caches are synced" controller="PV protection"
I0510 17:32:16.367348 1 shared_informer.go:357] "Caches are synced" controller="ReplicationController"
I0510 17:32:16.368090 1 shared_informer.go:357] "Caches are synced" controller="taint-eviction-controller"
I0510 17:32:16.368750 1 shared_informer.go:357] "Caches are synced" controller="daemon sets"
I0510 17:32:16.368889 1 shared_informer.go:357] "Caches are synced" controller="GC"
I0510 17:32:16.372167 1 shared_informer.go:357] "Caches are synced" controller="attach detach"
I0510 17:32:16.397055 1 shared_informer.go:357] "Caches are synced" controller="resource quota"
I0510 17:32:16.397725 1 shared_informer.go:357] "Caches are synced" controller="persistent volume"
I0510 17:32:16.411854 1 shared_informer.go:357] "Caches are synced" controller="HPA"
I0510 17:32:16.412408 1 shared_informer.go:357] "Caches are synced" controller="disruption"
I0510 17:32:16.426187 1 shared_informer.go:357] "Caches are synced" controller="legacy-service-account-token-cleaner"
I0510 17:32:16.426790 1 shared_informer.go:357] "Caches are synced" controller="deployment"
I0510 17:32:16.427970 1 shared_informer.go:357] "Caches are synced" controller="endpoint_slice"
I0510 17:32:16.429279 1 shared_informer.go:357] "Caches are synced" controller="PVC protection"
I0510 17:32:16.429296 1 shared_informer.go:357] "Caches are synced" controller="job"
I0510 17:32:16.429306 1 shared_informer.go:357] "Caches are synced" controller="stateful set"
==> kube-scheduler [59de35a41ff3] <==
W0510 17:32:10.238444 1 authentication.go:398] Continuing without authentication configuration. This may treat all requests as anonymous.
W0510 17:32:10.238452 1 authentication.go:399] To require authentication configuration lookup to succeed, set --authentication-tolerate-lookup-failure=false
I0510 17:32:10.260696 1 server.go:171] "Starting Kubernetes Scheduler" version="v1.33.0"
I0510 17:32:10.260728 1 server.go:173] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0510 17:32:10.263103 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
I0510 17:32:10.263294 1 shared_informer.go:350] "Waiting for caches to sync" controller="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
I0510 17:32:10.263758 1 secure_serving.go:211] Serving securely on 127.0.0.1:10259
I0510 17:32:10.263979 1 tlsconfig.go:243] "Starting DynamicServingCertificateController"
E0510 17:32:10.271006 1 reflector.go:200] "Failed to watch" err="failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod"
E0510 17:32:10.271107 1 reflector.go:200] "Failed to watch" err="failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.StorageClass"
E0510 17:32:10.271327 1 reflector.go:200] "Failed to watch" err="failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace"
E0510 17:32:10.271437 1 reflector.go:200] "Failed to watch" err="failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicaSet"
E0510 17:32:10.271518 1 reflector.go:200] "Failed to watch" err="failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.StatefulSet"
E0510 17:32:10.271750 1 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node"
E0510 17:32:10.271810 1 reflector.go:200] "Failed to watch" err="failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSINode"
E0510 17:32:10.271854 1 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIStorageCapacity"
E0510 17:32:10.271905 1 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver"
E0510 17:32:10.271945 1 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service"
E0510 17:32:10.271986 1 reflector.go:200] "Failed to watch" err="failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PodDisruptionBudget"
E0510 17:32:10.272089 1 reflector.go:200] "Failed to watch" err="failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolume"
E0510 17:32:10.272140 1 reflector.go:200] "Failed to watch" err="failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.ReplicationController"
E0510 17:32:10.272197 1 reflector.go:200] "Failed to watch" err="failed to list *v1.VolumeAttachment: volumeattachments.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"volumeattachments\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.VolumeAttachment"
E0510 17:32:10.272283 1 reflector.go:200] "Failed to watch" err="failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError" reflector="runtime/asm_arm64.s:1223" type="*v1.ConfigMap"
E0510 17:32:10.272304 1 reflector.go:200] "Failed to watch" err="failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.PersistentVolumeClaim"
I0510 17:32:11.564299 1 shared_informer.go:357] "Caches are synced" controller="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
==> kubelet <==
May 10 17:32:12 scheduled-stop-579771 kubelet[2560]: I0510 17:32:12.893628 2560 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/72d722a3be85f56d06c33af54d1293d1-kubeconfig\") pod \"kube-controller-manager-scheduled-stop-579771\" (UID: \"72d722a3be85f56d06c33af54d1293d1\") " pod="kube-system/kube-controller-manager-scheduled-stop-579771"
May 10 17:32:12 scheduled-stop-579771 kubelet[2560]: I0510 17:32:12.893650 2560 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/72d722a3be85f56d06c33af54d1293d1-usr-share-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-579771\" (UID: \"72d722a3be85f56d06c33af54d1293d1\") " pod="kube-system/kube-controller-manager-scheduled-stop-579771"
May 10 17:32:12 scheduled-stop-579771 kubelet[2560]: I0510 17:32:12.893674 2560 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/495e89af0d412c3d81fab15a289d83dc-ca-certs\") pod \"kube-apiserver-scheduled-stop-579771\" (UID: \"495e89af0d412c3d81fab15a289d83dc\") " pod="kube-system/kube-apiserver-scheduled-stop-579771"
May 10 17:32:12 scheduled-stop-579771 kubelet[2560]: I0510 17:32:12.893693 2560 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/72d722a3be85f56d06c33af54d1293d1-flexvolume-dir\") pod \"kube-controller-manager-scheduled-stop-579771\" (UID: \"72d722a3be85f56d06c33af54d1293d1\") " pod="kube-system/kube-controller-manager-scheduled-stop-579771"
May 10 17:32:12 scheduled-stop-579771 kubelet[2560]: I0510 17:32:12.893711 2560 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/72d722a3be85f56d06c33af54d1293d1-k8s-certs\") pod \"kube-controller-manager-scheduled-stop-579771\" (UID: \"72d722a3be85f56d06c33af54d1293d1\") " pod="kube-system/kube-controller-manager-scheduled-stop-579771"
May 10 17:32:12 scheduled-stop-579771 kubelet[2560]: I0510 17:32:12.893738 2560 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/72d722a3be85f56d06c33af54d1293d1-usr-local-share-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-579771\" (UID: \"72d722a3be85f56d06c33af54d1293d1\") " pod="kube-system/kube-controller-manager-scheduled-stop-579771"
May 10 17:32:12 scheduled-stop-579771 kubelet[2560]: I0510 17:32:12.893760 2560 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-data\" (UniqueName: \"kubernetes.io/host-path/cf5c1af43f7d57094562158fc3b55b1b-etcd-data\") pod \"etcd-scheduled-stop-579771\" (UID: \"cf5c1af43f7d57094562158fc3b55b1b\") " pod="kube-system/etcd-scheduled-stop-579771"
May 10 17:32:12 scheduled-stop-579771 kubelet[2560]: I0510 17:32:12.893778 2560 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/495e89af0d412c3d81fab15a289d83dc-etc-ca-certificates\") pod \"kube-apiserver-scheduled-stop-579771\" (UID: \"495e89af0d412c3d81fab15a289d83dc\") " pod="kube-system/kube-apiserver-scheduled-stop-579771"
May 10 17:32:12 scheduled-stop-579771 kubelet[2560]: I0510 17:32:12.893801 2560 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/495e89af0d412c3d81fab15a289d83dc-usr-share-ca-certificates\") pod \"kube-apiserver-scheduled-stop-579771\" (UID: \"495e89af0d412c3d81fab15a289d83dc\") " pod="kube-system/kube-apiserver-scheduled-stop-579771"
May 10 17:32:12 scheduled-stop-579771 kubelet[2560]: I0510 17:32:12.893821 2560 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/72d722a3be85f56d06c33af54d1293d1-etc-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-579771\" (UID: \"72d722a3be85f56d06c33af54d1293d1\") " pod="kube-system/kube-controller-manager-scheduled-stop-579771"
May 10 17:32:13 scheduled-stop-579771 kubelet[2560]: I0510 17:32:13.437159 2560 apiserver.go:52] "Watching apiserver"
May 10 17:32:13 scheduled-stop-579771 kubelet[2560]: I0510 17:32:13.488561 2560 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world"
May 10 17:32:13 scheduled-stop-579771 kubelet[2560]: I0510 17:32:13.605375 2560 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-scheduled-stop-579771"
May 10 17:32:13 scheduled-stop-579771 kubelet[2560]: I0510 17:32:13.606748 2560 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-scheduled-stop-579771"
May 10 17:32:13 scheduled-stop-579771 kubelet[2560]: I0510 17:32:13.607016 2560 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-scheduled-stop-579771"
May 10 17:32:13 scheduled-stop-579771 kubelet[2560]: E0510 17:32:13.625866 2560 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-scheduled-stop-579771\" already exists" pod="kube-system/kube-scheduler-scheduled-stop-579771"
May 10 17:32:13 scheduled-stop-579771 kubelet[2560]: E0510 17:32:13.631328 2560 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-scheduled-stop-579771\" already exists" pod="kube-system/kube-apiserver-scheduled-stop-579771"
May 10 17:32:13 scheduled-stop-579771 kubelet[2560]: E0510 17:32:13.631696 2560 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-scheduled-stop-579771\" already exists" pod="kube-system/kube-controller-manager-scheduled-stop-579771"
May 10 17:32:13 scheduled-stop-579771 kubelet[2560]: I0510 17:32:13.665597 2560 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-scheduled-stop-579771" podStartSLOduration=1.665578826 podStartE2EDuration="1.665578826s" podCreationTimestamp="2025-05-10 17:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-10 17:32:13.64954996 +0000 UTC m=+1.303291631" watchObservedRunningTime="2025-05-10 17:32:13.665578826 +0000 UTC m=+1.319320513"
May 10 17:32:13 scheduled-stop-579771 kubelet[2560]: I0510 17:32:13.677552 2560 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/etcd-scheduled-stop-579771" podStartSLOduration=1.677532551 podStartE2EDuration="1.677532551s" podCreationTimestamp="2025-05-10 17:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-10 17:32:13.667160913 +0000 UTC m=+1.320902601" watchObservedRunningTime="2025-05-10 17:32:13.677532551 +0000 UTC m=+1.331274230"
May 10 17:32:13 scheduled-stop-579771 kubelet[2560]: I0510 17:32:13.690144 2560 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-scheduled-stop-579771" podStartSLOduration=1.690128533 podStartE2EDuration="1.690128533s" podCreationTimestamp="2025-05-10 17:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-10 17:32:13.689892643 +0000 UTC m=+1.343634339" watchObservedRunningTime="2025-05-10 17:32:13.690128533 +0000 UTC m=+1.343870253"
May 10 17:32:13 scheduled-stop-579771 kubelet[2560]: I0510 17:32:13.690400 2560 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-scheduled-stop-579771" podStartSLOduration=1.690390883 podStartE2EDuration="1.690390883s" podCreationTimestamp="2025-05-10 17:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-10 17:32:13.677987583 +0000 UTC m=+1.331729262" watchObservedRunningTime="2025-05-10 17:32:13.690390883 +0000 UTC m=+1.344132611"
May 10 17:32:16 scheduled-stop-579771 kubelet[2560]: I0510 17:32:16.525845 2560 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/host-path/f0ef7b13-41ec-4158-be21-09aebcf77eeb-tmp\") pod \"storage-provisioner\" (UID: \"f0ef7b13-41ec-4158-be21-09aebcf77eeb\") " pod="kube-system/storage-provisioner"
May 10 17:32:16 scheduled-stop-579771 kubelet[2560]: I0510 17:32:16.525923 2560 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2lf2\" (UniqueName: \"kubernetes.io/projected/f0ef7b13-41ec-4158-be21-09aebcf77eeb-kube-api-access-b2lf2\") pod \"storage-provisioner\" (UID: \"f0ef7b13-41ec-4158-be21-09aebcf77eeb\") " pod="kube-system/storage-provisioner"
May 10 17:32:16 scheduled-stop-579771 kubelet[2560]: I0510 17:32:16.636779 2560 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory"
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-arm64 status --format={{.APIServer}} -p scheduled-stop-579771 -n scheduled-stop-579771
helpers_test.go:261: (dbg) Run: kubectl --context scheduled-stop-579771 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: coredns-674b8bbfcf-4d2bb kube-proxy-mz8vd storage-provisioner
helpers_test.go:274: ======> post-mortem[TestScheduledStopUnix]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context scheduled-stop-579771 describe pod coredns-674b8bbfcf-4d2bb kube-proxy-mz8vd storage-provisioner
helpers_test.go:277: (dbg) Non-zero exit: kubectl --context scheduled-stop-579771 describe pod coredns-674b8bbfcf-4d2bb kube-proxy-mz8vd storage-provisioner: exit status 1 (108.13401ms)
** stderr **
Error from server (NotFound): pods "coredns-674b8bbfcf-4d2bb" not found
Error from server (NotFound): pods "kube-proxy-mz8vd" not found
Error from server (NotFound): pods "storage-provisioner" not found
** /stderr **
helpers_test.go:279: kubectl --context scheduled-stop-579771 describe pod coredns-674b8bbfcf-4d2bb kube-proxy-mz8vd storage-provisioner: exit status 1
helpers_test.go:175: Cleaning up "scheduled-stop-579771" profile ...
helpers_test.go:178: (dbg) Run: out/minikube-linux-arm64 delete -p scheduled-stop-579771
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p scheduled-stop-579771: (2.201474412s)
--- FAIL: TestScheduledStopUnix (37.89s)