=== RUN TestScheduledStopUnix
scheduled_stop_test.go:128: (dbg) Run: out/minikube-linux-arm64 start -p scheduled-stop-375543 --memory=2048 --driver=docker --container-runtime=docker
scheduled_stop_test.go:128: (dbg) Done: out/minikube-linux-arm64 start -p scheduled-stop-375543 --memory=2048 --driver=docker --container-runtime=docker: (31.507267704s)
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-375543 --schedule 5m
scheduled_stop_test.go:191: (dbg) Run: out/minikube-linux-arm64 status --format={{.TimeToStop}} -p scheduled-stop-375543 -n scheduled-stop-375543
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-375543 --schedule 15s
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:98: process 217553 running but should have been killed on reschedule of stop
panic.go:631: *** TestScheduledStopUnix FAILED at 2025-03-17 11:03:59.952189245 +0000 UTC m=+2317.909338440
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestScheduledStopUnix]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect scheduled-stop-375543
helpers_test.go:235: (dbg) docker inspect scheduled-stop-375543:
-- stdout --
[
{
"Id": "51c64c237ddd7d0b5ba7bcff042adb73d5b0a3340789318647e3048cc1cf008c",
"Created": "2025-03-17T11:03:32.948922354Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 214553,
"ExitCode": 0,
"Error": "",
"StartedAt": "2025-03-17T11:03:33.020319469Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:df0c2544fb3106b890f0a9ab81fcf49f97edb092b83e47f42288ad5dfe1f4b40",
"ResolvConfPath": "/var/lib/docker/containers/51c64c237ddd7d0b5ba7bcff042adb73d5b0a3340789318647e3048cc1cf008c/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/51c64c237ddd7d0b5ba7bcff042adb73d5b0a3340789318647e3048cc1cf008c/hostname",
"HostsPath": "/var/lib/docker/containers/51c64c237ddd7d0b5ba7bcff042adb73d5b0a3340789318647e3048cc1cf008c/hosts",
"LogPath": "/var/lib/docker/containers/51c64c237ddd7d0b5ba7bcff042adb73d5b0a3340789318647e3048cc1cf008c/51c64c237ddd7d0b5ba7bcff042adb73d5b0a3340789318647e3048cc1cf008c-json.log",
"Name": "/scheduled-stop-375543",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"scheduled-stop-375543:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {}
},
"NetworkMode": "scheduled-stop-375543",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 2147483648,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 4294967296,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"ID": "51c64c237ddd7d0b5ba7bcff042adb73d5b0a3340789318647e3048cc1cf008c",
"LowerDir": "/var/lib/docker/overlay2/591d6fa7ea63d87a4222366043b9f8be29447bdb975f7a6f5f9e5b75dcf265a7-init/diff:/var/lib/docker/overlay2/42b510827f2f8842517c989dc5e1a994764ef51a8389d9062b6808cb43da9c38/diff",
"MergedDir": "/var/lib/docker/overlay2/591d6fa7ea63d87a4222366043b9f8be29447bdb975f7a6f5f9e5b75dcf265a7/merged",
"UpperDir": "/var/lib/docker/overlay2/591d6fa7ea63d87a4222366043b9f8be29447bdb975f7a6f5f9e5b75dcf265a7/diff",
"WorkDir": "/var/lib/docker/overlay2/591d6fa7ea63d87a4222366043b9f8be29447bdb975f7a6f5f9e5b75dcf265a7/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "scheduled-stop-375543",
"Source": "/var/lib/docker/volumes/scheduled-stop-375543/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "scheduled-stop-375543",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1741860993-20523@sha256:cd976907fa4d517c84fff1e5ef773b9fb3c738c4e1ded824ea5133470a66e185",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "scheduled-stop-375543",
"name.minikube.sigs.k8s.io": "scheduled-stop-375543",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "c6425a1a007f276aedbdc66b2613764bf167156b40fd7d93cdf2a8e4eb381ba6",
"SandboxKey": "/var/run/docker/netns/c6425a1a007f",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32970"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32971"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32974"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32972"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32973"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"scheduled-stop-375543": {
"IPAMConfig": {
"IPv4Address": "192.168.76.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "c6:41:b9:dc:8c:3d",
"DriverOpts": null,
"GwPriority": 0,
"NetworkID": "7b9766396531bfc1aab99165eebdd81d84d616df3af1e6ea2db9344c68e6509e",
"EndpointID": "43e0d10df55fe5aaeb5739d25289d3a83fac8a3669e2d573896a4ca8e09243f8",
"Gateway": "192.168.76.1",
"IPAddress": "192.168.76.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"scheduled-stop-375543",
"51c64c237ddd"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-arm64 status --format={{.Host}} -p scheduled-stop-375543 -n scheduled-stop-375543
helpers_test.go:244: <<< TestScheduledStopUnix FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestScheduledStopUnix]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-arm64 -p scheduled-stop-375543 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-arm64 -p scheduled-stop-375543 logs -n 25: (1.147686556s)
helpers_test.go:252: TestScheduledStopUnix logs:
-- stdout --
==> Audit <==
|---------|--------------------------------|-----------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|--------------------------------|-----------------------|---------|---------|---------------------|---------------------|
| stop | -p multinode-120432 | multinode-120432 | jenkins | v1.35.0 | 17 Mar 25 10:57 UTC | 17 Mar 25 10:57 UTC |
| start | -p multinode-120432 | multinode-120432 | jenkins | v1.35.0 | 17 Mar 25 10:57 UTC | 17 Mar 25 10:58 UTC |
| | --wait=true -v=8 | | | | | |
| | --alsologtostderr | | | | | |
| node | list -p multinode-120432 | multinode-120432 | jenkins | v1.35.0 | 17 Mar 25 10:58 UTC | |
| node | multinode-120432 node delete | multinode-120432 | jenkins | v1.35.0 | 17 Mar 25 10:58 UTC | 17 Mar 25 10:59 UTC |
| | m03 | | | | | |
| stop | multinode-120432 stop | multinode-120432 | jenkins | v1.35.0 | 17 Mar 25 10:59 UTC | 17 Mar 25 10:59 UTC |
| start | -p multinode-120432 | multinode-120432 | jenkins | v1.35.0 | 17 Mar 25 10:59 UTC | 17 Mar 25 11:00 UTC |
| | --wait=true -v=8 | | | | | |
| | --alsologtostderr | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| node | list -p multinode-120432 | multinode-120432 | jenkins | v1.35.0 | 17 Mar 25 11:00 UTC | |
| start | -p multinode-120432-m02 | multinode-120432-m02 | jenkins | v1.35.0 | 17 Mar 25 11:00 UTC | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| start | -p multinode-120432-m03 | multinode-120432-m03 | jenkins | v1.35.0 | 17 Mar 25 11:00 UTC | 17 Mar 25 11:00 UTC |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| node | add -p multinode-120432 | multinode-120432 | jenkins | v1.35.0 | 17 Mar 25 11:00 UTC | |
| delete | -p multinode-120432-m03 | multinode-120432-m03 | jenkins | v1.35.0 | 17 Mar 25 11:00 UTC | 17 Mar 25 11:00 UTC |
| delete | -p multinode-120432 | multinode-120432 | jenkins | v1.35.0 | 17 Mar 25 11:00 UTC | 17 Mar 25 11:00 UTC |
| start | -p test-preload-243342 | test-preload-243342 | jenkins | v1.35.0 | 17 Mar 25 11:00 UTC | 17 Mar 25 11:02 UTC |
| | --memory=2200 | | | | | |
| | --alsologtostderr | | | | | |
| | --wait=true --preload=false | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --kubernetes-version=v1.24.4 | | | | | |
| image | test-preload-243342 image pull | test-preload-243342 | jenkins | v1.35.0 | 17 Mar 25 11:02 UTC | 17 Mar 25 11:02 UTC |
| | gcr.io/k8s-minikube/busybox | | | | | |
| stop | -p test-preload-243342 | test-preload-243342 | jenkins | v1.35.0 | 17 Mar 25 11:02 UTC | 17 Mar 25 11:03 UTC |
| start | -p test-preload-243342 | test-preload-243342 | jenkins | v1.35.0 | 17 Mar 25 11:03 UTC | 17 Mar 25 11:03 UTC |
| | --memory=2200 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| | --wait=true --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| image | test-preload-243342 image list | test-preload-243342 | jenkins | v1.35.0 | 17 Mar 25 11:03 UTC | 17 Mar 25 11:03 UTC |
| delete | -p test-preload-243342 | test-preload-243342 | jenkins | v1.35.0 | 17 Mar 25 11:03 UTC | 17 Mar 25 11:03 UTC |
| start | -p scheduled-stop-375543 | scheduled-stop-375543 | jenkins | v1.35.0 | 17 Mar 25 11:03 UTC | 17 Mar 25 11:03 UTC |
| | --memory=2048 --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| stop | -p scheduled-stop-375543 | scheduled-stop-375543 | jenkins | v1.35.0 | 17 Mar 25 11:03 UTC | |
| | --schedule 5m | | | | | |
| stop | -p scheduled-stop-375543 | scheduled-stop-375543 | jenkins | v1.35.0 | 17 Mar 25 11:03 UTC | |
| | --schedule 5m | | | | | |
| stop | -p scheduled-stop-375543 | scheduled-stop-375543 | jenkins | v1.35.0 | 17 Mar 25 11:03 UTC | |
| | --schedule 5m | | | | | |
| stop | -p scheduled-stop-375543 | scheduled-stop-375543 | jenkins | v1.35.0 | 17 Mar 25 11:03 UTC | |
| | --schedule 15s | | | | | |
| stop | -p scheduled-stop-375543 | scheduled-stop-375543 | jenkins | v1.35.0 | 17 Mar 25 11:04 UTC | |
| | --schedule 15s | | | | | |
| stop | -p scheduled-stop-375543 | scheduled-stop-375543 | jenkins | v1.35.0 | 17 Mar 25 11:04 UTC | |
| | --schedule 15s | | | | | |
|---------|--------------------------------|-----------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2025/03/17 11:03:27
Running on machine: ip-172-31-29-130
Binary: Built with gc go1.24.0 for linux/arm64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0317 11:03:27.981401 214170 out.go:345] Setting OutFile to fd 1 ...
I0317 11:03:27.981544 214170 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0317 11:03:27.981547 214170 out.go:358] Setting ErrFile to fd 2...
I0317 11:03:27.981551 214170 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0317 11:03:27.981796 214170 root.go:338] Updating PATH: /home/jenkins/minikube-integration/20535-2275/.minikube/bin
I0317 11:03:27.982193 214170 out.go:352] Setting JSON to false
I0317 11:03:27.983019 214170 start.go:129] hostinfo: {"hostname":"ip-172-31-29-130","uptime":2751,"bootTime":1742206657,"procs":175,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1077-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"36adf542-ef4f-4e2d-a0c8-6868d1383ff9"}
I0317 11:03:27.983074 214170 start.go:139] virtualization:
I0317 11:03:27.988957 214170 out.go:177] * [scheduled-stop-375543] minikube v1.35.0 on Ubuntu 20.04 (arm64)
I0317 11:03:27.992531 214170 out.go:177] - MINIKUBE_LOCATION=20535
I0317 11:03:27.992572 214170 notify.go:220] Checking for updates...
I0317 11:03:27.996249 214170 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0317 11:03:27.999509 214170 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/20535-2275/kubeconfig
I0317 11:03:28.002552 214170 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/20535-2275/.minikube
I0317 11:03:28.005650 214170 out.go:177] - MINIKUBE_BIN=out/minikube-linux-arm64
I0317 11:03:28.009601 214170 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0317 11:03:28.013080 214170 driver.go:394] Setting default libvirt URI to qemu:///system
I0317 11:03:28.040677 214170 docker.go:123] docker version: linux-28.0.1:Docker Engine - Community
I0317 11:03:28.040807 214170 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0317 11:03:28.100283 214170 info.go:266] docker info: {ID:U5VK:ZNT5:35M3:FHLW:Q7TL:ELFX:BNAG:AV4T:UD2H:SK5L:SEJV:SJJL Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:23 OomKillDisable:true NGoroutines:42 SystemTime:2025-03-17 11:03:28.09110935 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1077-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:aa
rch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214831104 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-29-130 Labels:[] ExperimentalBuild:false ServerVersion:28.0.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:bcc810d6b9066471b0b6fa75f557a15a1cbf31bb Expected:bcc810d6b9066471b0b6fa75f557a15a1cbf31bb} RuncCommit:{ID:v1.2.4-0-g6c52b3f Expected:v1.2.4-0-g6c52b3f} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> Ser
verErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.21.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.33.1]] Warnings:<nil>}}
I0317 11:03:28.100378 214170 docker.go:318] overlay module found
I0317 11:03:28.103943 214170 out.go:177] * Using the docker driver based on user configuration
I0317 11:03:28.106780 214170 start.go:297] selected driver: docker
I0317 11:03:28.106793 214170 start.go:901] validating driver "docker" against <nil>
I0317 11:03:28.106805 214170 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0317 11:03:28.107584 214170 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0317 11:03:28.161929 214170 info.go:266] docker info: {ID:U5VK:ZNT5:35M3:FHLW:Q7TL:ELFX:BNAG:AV4T:UD2H:SK5L:SEJV:SJJL Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:23 OomKillDisable:true NGoroutines:42 SystemTime:2025-03-17 11:03:28.153312358 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1077-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214831104 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-29-130 Labels:[] ExperimentalBuild:false ServerVersion:28.0.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:bcc810d6b9066471b0b6fa75f557a15a1cbf31bb Expected:bcc810d6b9066471b0b6fa75f557a15a1cbf31bb} RuncCommit:{ID:v1.2.4-0-g6c52b3f Expected:v1.2.4-0-g6c52b3f} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> Se
rverErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.21.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.33.1]] Warnings:<nil>}}
I0317 11:03:28.162066 214170 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0317 11:03:28.162286 214170 start_flags.go:929] Wait components to verify : map[apiserver:true system_pods:true]
I0317 11:03:28.165284 214170 out.go:177] * Using Docker driver with root privileges
I0317 11:03:28.168151 214170 cni.go:84] Creating CNI manager for ""
I0317 11:03:28.168210 214170 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0317 11:03:28.168217 214170 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0317 11:03:28.168292 214170 start.go:340] cluster config:
{Name:scheduled-stop-375543 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1741860993-20523@sha256:cd976907fa4d517c84fff1e5ef773b9fb3c738c4e1ded824ea5133470a66e185 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.32.2 ClusterName:scheduled-stop-375543 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local
ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.32.2 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0317 11:03:28.171395 214170 out.go:177] * Starting "scheduled-stop-375543" primary control-plane node in "scheduled-stop-375543" cluster
I0317 11:03:28.174294 214170 cache.go:121] Beginning downloading kic base image for docker with docker
I0317 11:03:28.177175 214170 out.go:177] * Pulling base image v0.0.46-1741860993-20523 ...
I0317 11:03:28.180109 214170 preload.go:131] Checking if preload exists for k8s version v1.32.2 and runtime docker
I0317 11:03:28.180152 214170 preload.go:146] Found local preload: /home/jenkins/minikube-integration/20535-2275/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.32.2-docker-overlay2-arm64.tar.lz4
I0317 11:03:28.180159 214170 cache.go:56] Caching tarball of preloaded images
I0317 11:03:28.180187 214170 image.go:81] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1741860993-20523@sha256:cd976907fa4d517c84fff1e5ef773b9fb3c738c4e1ded824ea5133470a66e185 in local docker daemon
I0317 11:03:28.180238 214170 preload.go:172] Found /home/jenkins/minikube-integration/20535-2275/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.32.2-docker-overlay2-arm64.tar.lz4 in cache, skipping download
I0317 11:03:28.180246 214170 cache.go:59] Finished verifying existence of preloaded tar for v1.32.2 on docker
I0317 11:03:28.180574 214170 profile.go:143] Saving config to /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/config.json ...
I0317 11:03:28.180591 214170 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/config.json: {Name:mk346bd4a0b279619504b2ae9b01b5efb5fdbfbe Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0317 11:03:28.198732 214170 image.go:100] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1741860993-20523@sha256:cd976907fa4d517c84fff1e5ef773b9fb3c738c4e1ded824ea5133470a66e185 in local docker daemon, skipping pull
I0317 11:03:28.198744 214170 cache.go:145] gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1741860993-20523@sha256:cd976907fa4d517c84fff1e5ef773b9fb3c738c4e1ded824ea5133470a66e185 exists in daemon, skipping load
I0317 11:03:28.198761 214170 cache.go:230] Successfully downloaded all kic artifacts
I0317 11:03:28.198786 214170 start.go:360] acquireMachinesLock for scheduled-stop-375543: {Name:mk1bdedfe9422b0af9282df267e4fffc1d06ec0c Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0317 11:03:28.198893 214170 start.go:364] duration metric: took 93.892µs to acquireMachinesLock for "scheduled-stop-375543"
I0317 11:03:28.198923 214170 start.go:93] Provisioning new machine with config: &{Name:scheduled-stop-375543 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1741860993-20523@sha256:cd976907fa4d517c84fff1e5ef773b9fb3c738c4e1ded824ea5133470a66e185 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.32.2 ClusterName:scheduled-stop-375543 Namespace:default APIServerHAVIP: A
PIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.32.2 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath
: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.32.2 ContainerRuntime:docker ControlPlane:true Worker:true}
I0317 11:03:28.198996 214170 start.go:125] createHost starting for "" (driver="docker")
I0317 11:03:28.202370 214170 out.go:235] * Creating docker container (CPUs=2, Memory=2048MB) ...
I0317 11:03:28.202579 214170 start.go:159] libmachine.API.Create for "scheduled-stop-375543" (driver="docker")
I0317 11:03:28.202608 214170 client.go:168] LocalClient.Create starting
I0317 11:03:28.202669 214170 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/20535-2275/.minikube/certs/ca.pem
I0317 11:03:28.202703 214170 main.go:141] libmachine: Decoding PEM data...
I0317 11:03:28.202717 214170 main.go:141] libmachine: Parsing certificate...
I0317 11:03:28.202768 214170 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/20535-2275/.minikube/certs/cert.pem
I0317 11:03:28.202783 214170 main.go:141] libmachine: Decoding PEM data...
I0317 11:03:28.202791 214170 main.go:141] libmachine: Parsing certificate...
I0317 11:03:28.203163 214170 cli_runner.go:164] Run: docker network inspect scheduled-stop-375543 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0317 11:03:28.218250 214170 cli_runner.go:211] docker network inspect scheduled-stop-375543 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0317 11:03:28.218342 214170 network_create.go:284] running [docker network inspect scheduled-stop-375543] to gather additional debugging logs...
I0317 11:03:28.218355 214170 cli_runner.go:164] Run: docker network inspect scheduled-stop-375543
W0317 11:03:28.233145 214170 cli_runner.go:211] docker network inspect scheduled-stop-375543 returned with exit code 1
I0317 11:03:28.233164 214170 network_create.go:287] error running [docker network inspect scheduled-stop-375543]: docker network inspect scheduled-stop-375543: exit status 1
stdout:
[]
stderr:
Error response from daemon: network scheduled-stop-375543 not found
I0317 11:03:28.233174 214170 network_create.go:289] output of [docker network inspect scheduled-stop-375543]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network scheduled-stop-375543 not found
** /stderr **
I0317 11:03:28.233281 214170 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0317 11:03:28.249001 214170 network.go:211] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-4537de9dd09b IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:f6:3f:dc:bd:c3:10} reservation:<nil>}
I0317 11:03:28.249261 214170 network.go:211] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-2ef24a570a1a IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:02:a2:b0:df:5e:b6} reservation:<nil>}
I0317 11:03:28.249479 214170 network.go:211] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-5b990df68c1b IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:76:7c:32:72:0b:3f} reservation:<nil>}
I0317 11:03:28.249779 214170 network.go:206] using free private subnet 192.168.76.0/24: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x4001984b40}
I0317 11:03:28.249794 214170 network_create.go:124] attempt to create docker network scheduled-stop-375543 192.168.76.0/24 with gateway 192.168.76.1 and MTU of 1500 ...
I0317 11:03:28.249852 214170 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.76.0/24 --gateway=192.168.76.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=scheduled-stop-375543 scheduled-stop-375543
I0317 11:03:28.307506 214170 network_create.go:108] docker network scheduled-stop-375543 192.168.76.0/24 created
I0317 11:03:28.307525 214170 kic.go:121] calculated static IP "192.168.76.2" for the "scheduled-stop-375543" container
I0317 11:03:28.307611 214170 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0317 11:03:28.322225 214170 cli_runner.go:164] Run: docker volume create scheduled-stop-375543 --label name.minikube.sigs.k8s.io=scheduled-stop-375543 --label created_by.minikube.sigs.k8s.io=true
I0317 11:03:28.339571 214170 oci.go:103] Successfully created a docker volume scheduled-stop-375543
I0317 11:03:28.339644 214170 cli_runner.go:164] Run: docker run --rm --name scheduled-stop-375543-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-375543 --entrypoint /usr/bin/test -v scheduled-stop-375543:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1741860993-20523@sha256:cd976907fa4d517c84fff1e5ef773b9fb3c738c4e1ded824ea5133470a66e185 -d /var/lib
I0317 11:03:28.848259 214170 oci.go:107] Successfully prepared a docker volume scheduled-stop-375543
I0317 11:03:28.848305 214170 preload.go:131] Checking if preload exists for k8s version v1.32.2 and runtime docker
I0317 11:03:28.848322 214170 kic.go:194] Starting extracting preloaded images to volume ...
I0317 11:03:28.848385 214170 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/20535-2275/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.32.2-docker-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-375543:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1741860993-20523@sha256:cd976907fa4d517c84fff1e5ef773b9fb3c738c4e1ded824ea5133470a66e185 -I lz4 -xf /preloaded.tar -C /extractDir
I0317 11:03:32.874962 214170 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/20535-2275/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.32.2-docker-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-375543:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1741860993-20523@sha256:cd976907fa4d517c84fff1e5ef773b9fb3c738c4e1ded824ea5133470a66e185 -I lz4 -xf /preloaded.tar -C /extractDir: (4.026533863s)
I0317 11:03:32.875012 214170 kic.go:203] duration metric: took 4.026685733s to extract preloaded images to volume ...
W0317 11:03:32.875148 214170 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0317 11:03:32.875253 214170 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0317 11:03:32.933376 214170 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname scheduled-stop-375543 --name scheduled-stop-375543 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-375543 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=scheduled-stop-375543 --network scheduled-stop-375543 --ip 192.168.76.2 --volume scheduled-stop-375543:/var --security-opt apparmor=unconfined --memory=2048mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1741860993-20523@sha256:cd976907fa4d517c84fff1e5ef773b9fb3c738c4e1ded824ea5133470a66e185
I0317 11:03:33.257824 214170 cli_runner.go:164] Run: docker container inspect scheduled-stop-375543 --format={{.State.Running}}
I0317 11:03:33.278395 214170 cli_runner.go:164] Run: docker container inspect scheduled-stop-375543 --format={{.State.Status}}
I0317 11:03:33.303097 214170 cli_runner.go:164] Run: docker exec scheduled-stop-375543 stat /var/lib/dpkg/alternatives/iptables
I0317 11:03:33.357733 214170 oci.go:144] the created container "scheduled-stop-375543" has a running status.
I0317 11:03:33.357764 214170 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/20535-2275/.minikube/machines/scheduled-stop-375543/id_rsa...
I0317 11:03:33.940466 214170 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/20535-2275/.minikube/machines/scheduled-stop-375543/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0317 11:03:33.970287 214170 cli_runner.go:164] Run: docker container inspect scheduled-stop-375543 --format={{.State.Status}}
I0317 11:03:33.997431 214170 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0317 11:03:33.997442 214170 kic_runner.go:114] Args: [docker exec --privileged scheduled-stop-375543 chown docker:docker /home/docker/.ssh/authorized_keys]
I0317 11:03:34.064608 214170 cli_runner.go:164] Run: docker container inspect scheduled-stop-375543 --format={{.State.Status}}
I0317 11:03:34.085706 214170 machine.go:93] provisionDockerMachine start ...
I0317 11:03:34.085787 214170 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-375543
I0317 11:03:34.104841 214170 main.go:141] libmachine: Using SSH client type: native
I0317 11:03:34.106245 214170 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3e66c0] 0x3e8e80 <nil> [] 0s} 127.0.0.1 32970 <nil> <nil>}
I0317 11:03:34.106255 214170 main.go:141] libmachine: About to run SSH command:
hostname
I0317 11:03:34.242436 214170 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-375543
I0317 11:03:34.242451 214170 ubuntu.go:169] provisioning hostname "scheduled-stop-375543"
I0317 11:03:34.242509 214170 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-375543
I0317 11:03:34.266432 214170 main.go:141] libmachine: Using SSH client type: native
I0317 11:03:34.266727 214170 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3e66c0] 0x3e8e80 <nil> [] 0s} 127.0.0.1 32970 <nil> <nil>}
I0317 11:03:34.266736 214170 main.go:141] libmachine: About to run SSH command:
sudo hostname scheduled-stop-375543 && echo "scheduled-stop-375543" | sudo tee /etc/hostname
I0317 11:03:34.406880 214170 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-375543
I0317 11:03:34.406954 214170 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-375543
I0317 11:03:34.425327 214170 main.go:141] libmachine: Using SSH client type: native
I0317 11:03:34.425638 214170 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3e66c0] 0x3e8e80 <nil> [] 0s} 127.0.0.1 32970 <nil> <nil>}
I0317 11:03:34.425653 214170 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sscheduled-stop-375543' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 scheduled-stop-375543/g' /etc/hosts;
else
echo '127.0.1.1 scheduled-stop-375543' | sudo tee -a /etc/hosts;
fi
fi
I0317 11:03:34.555091 214170 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0317 11:03:34.555107 214170 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/20535-2275/.minikube CaCertPath:/home/jenkins/minikube-integration/20535-2275/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/20535-2275/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/20535-2275/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/20535-2275/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/20535-2275/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/20535-2275/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/20535-2275/.minikube}
I0317 11:03:34.555133 214170 ubuntu.go:177] setting up certificates
I0317 11:03:34.555140 214170 provision.go:84] configureAuth start
I0317 11:03:34.555199 214170 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-375543
I0317 11:03:34.573386 214170 provision.go:143] copyHostCerts
I0317 11:03:34.573448 214170 exec_runner.go:144] found /home/jenkins/minikube-integration/20535-2275/.minikube/ca.pem, removing ...
I0317 11:03:34.573455 214170 exec_runner.go:203] rm: /home/jenkins/minikube-integration/20535-2275/.minikube/ca.pem
I0317 11:03:34.573533 214170 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20535-2275/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/20535-2275/.minikube/ca.pem (1078 bytes)
I0317 11:03:34.573631 214170 exec_runner.go:144] found /home/jenkins/minikube-integration/20535-2275/.minikube/cert.pem, removing ...
I0317 11:03:34.573635 214170 exec_runner.go:203] rm: /home/jenkins/minikube-integration/20535-2275/.minikube/cert.pem
I0317 11:03:34.573661 214170 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20535-2275/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/20535-2275/.minikube/cert.pem (1123 bytes)
I0317 11:03:34.573724 214170 exec_runner.go:144] found /home/jenkins/minikube-integration/20535-2275/.minikube/key.pem, removing ...
I0317 11:03:34.573727 214170 exec_runner.go:203] rm: /home/jenkins/minikube-integration/20535-2275/.minikube/key.pem
I0317 11:03:34.573750 214170 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20535-2275/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/20535-2275/.minikube/key.pem (1679 bytes)
I0317 11:03:34.573804 214170 provision.go:117] generating server cert: /home/jenkins/minikube-integration/20535-2275/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/20535-2275/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/20535-2275/.minikube/certs/ca-key.pem org=jenkins.scheduled-stop-375543 san=[127.0.0.1 192.168.76.2 localhost minikube scheduled-stop-375543]
I0317 11:03:34.816755 214170 provision.go:177] copyRemoteCerts
I0317 11:03:34.816816 214170 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0317 11:03:34.816855 214170 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-375543
I0317 11:03:34.834213 214170 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/20535-2275/.minikube/machines/scheduled-stop-375543/id_rsa Username:docker}
I0317 11:03:34.923692 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0317 11:03:34.948010 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/machines/server.pem --> /etc/docker/server.pem (1229 bytes)
I0317 11:03:34.972159 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0317 11:03:34.996542 214170 provision.go:87] duration metric: took 441.38842ms to configureAuth
I0317 11:03:34.996558 214170 ubuntu.go:193] setting minikube options for container-runtime
I0317 11:03:34.996746 214170 config.go:182] Loaded profile config "scheduled-stop-375543": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.32.2
I0317 11:03:34.996795 214170 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-375543
I0317 11:03:35.014538 214170 main.go:141] libmachine: Using SSH client type: native
I0317 11:03:35.014842 214170 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3e66c0] 0x3e8e80 <nil> [] 0s} 127.0.0.1 32970 <nil> <nil>}
I0317 11:03:35.014850 214170 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0317 11:03:35.139929 214170 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0317 11:03:35.139940 214170 ubuntu.go:71] root file system type: overlay
I0317 11:03:35.140054 214170 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0317 11:03:35.140116 214170 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-375543
I0317 11:03:35.157792 214170 main.go:141] libmachine: Using SSH client type: native
I0317 11:03:35.158087 214170 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3e66c0] 0x3e8e80 <nil> [] 0s} 127.0.0.1 32970 <nil> <nil>}
I0317 11:03:35.158162 214170 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0317 11:03:35.290162 214170 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0317 11:03:35.290248 214170 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-375543
I0317 11:03:35.308169 214170 main.go:141] libmachine: Using SSH client type: native
I0317 11:03:35.308456 214170 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x3e66c0] 0x3e8e80 <nil> [] 0s} 127.0.0.1 32970 <nil> <nil>}
I0317 11:03:35.308471 214170 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0317 11:03:36.147280 214170 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2025-02-26 10:39:24.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2025-03-17 11:03:35.285651190 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target nss-lookup.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0317 11:03:36.147299 214170 machine.go:96] duration metric: took 2.061581859s to provisionDockerMachine
I0317 11:03:36.147309 214170 client.go:171] duration metric: took 7.94469677s to LocalClient.Create
I0317 11:03:36.147321 214170 start.go:167] duration metric: took 7.944742563s to libmachine.API.Create "scheduled-stop-375543"
I0317 11:03:36.147328 214170 start.go:293] postStartSetup for "scheduled-stop-375543" (driver="docker")
I0317 11:03:36.147337 214170 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0317 11:03:36.147394 214170 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0317 11:03:36.147430 214170 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-375543
I0317 11:03:36.165097 214170 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/20535-2275/.minikube/machines/scheduled-stop-375543/id_rsa Username:docker}
I0317 11:03:36.256050 214170 ssh_runner.go:195] Run: cat /etc/os-release
I0317 11:03:36.259418 214170 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0317 11:03:36.259440 214170 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0317 11:03:36.259449 214170 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0317 11:03:36.259455 214170 info.go:137] Remote host: Ubuntu 22.04.5 LTS
I0317 11:03:36.259464 214170 filesync.go:126] Scanning /home/jenkins/minikube-integration/20535-2275/.minikube/addons for local assets ...
I0317 11:03:36.259517 214170 filesync.go:126] Scanning /home/jenkins/minikube-integration/20535-2275/.minikube/files for local assets ...
I0317 11:03:36.259598 214170 filesync.go:149] local asset: /home/jenkins/minikube-integration/20535-2275/.minikube/files/etc/ssl/certs/75852.pem -> 75852.pem in /etc/ssl/certs
I0317 11:03:36.259696 214170 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I0317 11:03:36.268379 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/files/etc/ssl/certs/75852.pem --> /etc/ssl/certs/75852.pem (1708 bytes)
I0317 11:03:36.293322 214170 start.go:296] duration metric: took 145.981168ms for postStartSetup
I0317 11:03:36.293689 214170 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-375543
I0317 11:03:36.311308 214170 profile.go:143] Saving config to /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/config.json ...
I0317 11:03:36.311594 214170 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0317 11:03:36.311631 214170 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-375543
I0317 11:03:36.328238 214170 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/20535-2275/.minikube/machines/scheduled-stop-375543/id_rsa Username:docker}
I0317 11:03:36.415755 214170 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0317 11:03:36.420192 214170 start.go:128] duration metric: took 8.221182442s to createHost
I0317 11:03:36.420206 214170 start.go:83] releasing machines lock for "scheduled-stop-375543", held for 8.221306544s
I0317 11:03:36.420283 214170 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-375543
I0317 11:03:36.437169 214170 ssh_runner.go:195] Run: cat /version.json
I0317 11:03:36.437227 214170 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0317 11:03:36.437269 214170 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-375543
I0317 11:03:36.437290 214170 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-375543
I0317 11:03:36.462237 214170 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/20535-2275/.minikube/machines/scheduled-stop-375543/id_rsa Username:docker}
I0317 11:03:36.472541 214170 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/20535-2275/.minikube/machines/scheduled-stop-375543/id_rsa Username:docker}
I0317 11:03:36.554510 214170 ssh_runner.go:195] Run: systemctl --version
I0317 11:03:36.686549 214170 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0317 11:03:36.690934 214170 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0317 11:03:36.716605 214170 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0317 11:03:36.716679 214170 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0317 11:03:36.748966 214170 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0317 11:03:36.748998 214170 start.go:495] detecting cgroup driver to use...
I0317 11:03:36.749031 214170 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0317 11:03:36.749184 214170 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0317 11:03:36.767247 214170 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0317 11:03:36.779407 214170 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0317 11:03:36.789668 214170 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0317 11:03:36.789726 214170 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0317 11:03:36.800226 214170 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0317 11:03:36.810455 214170 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0317 11:03:36.820805 214170 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0317 11:03:36.830772 214170 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0317 11:03:36.840544 214170 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0317 11:03:36.850912 214170 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0317 11:03:36.860528 214170 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0317 11:03:36.871112 214170 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0317 11:03:36.880149 214170 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0317 11:03:36.888926 214170 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0317 11:03:36.987504 214170 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0317 11:03:37.108540 214170 start.go:495] detecting cgroup driver to use...
I0317 11:03:37.108579 214170 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0317 11:03:37.108629 214170 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0317 11:03:37.124946 214170 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0317 11:03:37.125010 214170 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0317 11:03:37.146015 214170 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0317 11:03:37.163040 214170 ssh_runner.go:195] Run: which cri-dockerd
I0317 11:03:37.166646 214170 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0317 11:03:37.178173 214170 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0317 11:03:37.197536 214170 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0317 11:03:37.305506 214170 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0317 11:03:37.407002 214170 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0317 11:03:37.407088 214170 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0317 11:03:37.428306 214170 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0317 11:03:37.535249 214170 ssh_runner.go:195] Run: sudo systemctl restart docker
I0317 11:03:37.848231 214170 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0317 11:03:37.860509 214170 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0317 11:03:37.872646 214170 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0317 11:03:37.967829 214170 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0317 11:03:38.066768 214170 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0317 11:03:38.162106 214170 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0317 11:03:38.176050 214170 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0317 11:03:38.187450 214170 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0317 11:03:38.277603 214170 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0317 11:03:38.351137 214170 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0317 11:03:38.351213 214170 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0317 11:03:38.356130 214170 start.go:563] Will wait 60s for crictl version
I0317 11:03:38.356194 214170 ssh_runner.go:195] Run: which crictl
I0317 11:03:38.359767 214170 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0317 11:03:38.399335 214170 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 28.0.1
RuntimeApiVersion: v1
I0317 11:03:38.399398 214170 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0317 11:03:38.422312 214170 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0317 11:03:38.453099 214170 out.go:235] * Preparing Kubernetes v1.32.2 on Docker 28.0.1 ...
I0317 11:03:38.453257 214170 cli_runner.go:164] Run: docker network inspect scheduled-stop-375543 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0317 11:03:38.469980 214170 ssh_runner.go:195] Run: grep 192.168.76.1 host.minikube.internal$ /etc/hosts
I0317 11:03:38.473880 214170 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.76.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0317 11:03:38.485086 214170 kubeadm.go:883] updating cluster {Name:scheduled-stop-375543 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1741860993-20523@sha256:cd976907fa4d517c84fff1e5ef773b9fb3c738c4e1ded824ea5133470a66e185 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.32.2 ClusterName:scheduled-stop-375543 Namespace:default APIServerHAVIP: APIServerName:miniku
beCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.32.2 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: Stati
cIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0317 11:03:38.485188 214170 preload.go:131] Checking if preload exists for k8s version v1.32.2 and runtime docker
I0317 11:03:38.485240 214170 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0317 11:03:38.505209 214170 docker.go:689] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.32.2
registry.k8s.io/kube-controller-manager:v1.32.2
registry.k8s.io/kube-scheduler:v1.32.2
registry.k8s.io/kube-proxy:v1.32.2
registry.k8s.io/etcd:3.5.16-0
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0317 11:03:38.505222 214170 docker.go:619] Images already preloaded, skipping extraction
I0317 11:03:38.505284 214170 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0317 11:03:38.524612 214170 docker.go:689] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.32.2
registry.k8s.io/kube-scheduler:v1.32.2
registry.k8s.io/kube-controller-manager:v1.32.2
registry.k8s.io/kube-proxy:v1.32.2
registry.k8s.io/etcd:3.5.16-0
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0317 11:03:38.524625 214170 cache_images.go:84] Images are preloaded, skipping loading
I0317 11:03:38.524636 214170 kubeadm.go:934] updating node { 192.168.76.2 8443 v1.32.2 docker true true} ...
I0317 11:03:38.524723 214170 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.32.2/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=scheduled-stop-375543 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.76.2
[Install]
config:
{KubernetesVersion:v1.32.2 ClusterName:scheduled-stop-375543 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0317 11:03:38.524792 214170 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0317 11:03:38.571745 214170 cni.go:84] Creating CNI manager for ""
I0317 11:03:38.571760 214170 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0317 11:03:38.571769 214170 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0317 11:03:38.571790 214170 kubeadm.go:189] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.76.2 APIServerPort:8443 KubernetesVersion:v1.32.2 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:scheduled-stop-375543 NodeName:scheduled-stop-375543 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.76.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.76.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPo
dPath:/etc/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0317 11:03:38.571912 214170 kubeadm.go:195] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta4
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.76.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "scheduled-stop-375543"
kubeletExtraArgs:
- name: "node-ip"
value: "192.168.76.2"
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta4
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.76.2"]
extraArgs:
- name: "enable-admission-plugins"
value: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
- name: "allocate-node-cidrs"
value: "true"
- name: "leader-elect"
value: "false"
scheduler:
extraArgs:
- name: "leader-elect"
value: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
- name: "proxy-refresh-interval"
value: "70000"
kubernetesVersion: v1.32.2
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0317 11:03:38.572016 214170 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.32.2
I0317 11:03:38.581213 214170 binaries.go:44] Found k8s binaries, skipping transfer
I0317 11:03:38.581280 214170 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0317 11:03:38.590233 214170 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (320 bytes)
I0317 11:03:38.609154 214170 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0317 11:03:38.627234 214170 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2299 bytes)
I0317 11:03:38.645650 214170 ssh_runner.go:195] Run: grep 192.168.76.2 control-plane.minikube.internal$ /etc/hosts
I0317 11:03:38.649214 214170 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.76.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0317 11:03:38.659980 214170 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0317 11:03:38.749590 214170 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0317 11:03:38.765474 214170 certs.go:68] Setting up /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543 for IP: 192.168.76.2
I0317 11:03:38.765485 214170 certs.go:194] generating shared ca certs ...
I0317 11:03:38.765499 214170 certs.go:226] acquiring lock for ca certs: {Name:mk3596b40ac59f57e92ee44d1d2c92250574921d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0317 11:03:38.765633 214170 certs.go:235] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/20535-2275/.minikube/ca.key
I0317 11:03:38.765672 214170 certs.go:235] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/20535-2275/.minikube/proxy-client-ca.key
I0317 11:03:38.765678 214170 certs.go:256] generating profile certs ...
I0317 11:03:38.765737 214170 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/client.key
I0317 11:03:38.765749 214170 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/client.crt with IP's: []
I0317 11:03:38.937132 214170 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/client.crt ...
I0317 11:03:38.937148 214170 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/client.crt: {Name:mk69b77bd51ea4c9754f91558d862a6b7ac1e1f8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0317 11:03:38.937361 214170 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/client.key ...
I0317 11:03:38.937369 214170 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/client.key: {Name:mked8ca864b9af6b31447e582047334262956216 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0317 11:03:38.937474 214170 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/apiserver.key.5eb4fc4e
I0317 11:03:38.937487 214170 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/apiserver.crt.5eb4fc4e with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.76.2]
I0317 11:03:39.185533 214170 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/apiserver.crt.5eb4fc4e ...
I0317 11:03:39.185548 214170 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/apiserver.crt.5eb4fc4e: {Name:mk7769d29d7859ae09ceddeec66b9de857ea26f9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0317 11:03:39.185739 214170 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/apiserver.key.5eb4fc4e ...
I0317 11:03:39.185747 214170 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/apiserver.key.5eb4fc4e: {Name:mk51057cbcf4987baa42b66ddeb623c3c47498cd Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0317 11:03:39.185870 214170 certs.go:381] copying /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/apiserver.crt.5eb4fc4e -> /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/apiserver.crt
I0317 11:03:39.185942 214170 certs.go:385] copying /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/apiserver.key.5eb4fc4e -> /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/apiserver.key
I0317 11:03:39.185993 214170 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/proxy-client.key
I0317 11:03:39.186007 214170 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/proxy-client.crt with IP's: []
I0317 11:03:39.506630 214170 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/proxy-client.crt ...
I0317 11:03:39.506645 214170 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/proxy-client.crt: {Name:mk22ab4ac6da465d4767c0853bef0fc718dad648 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0317 11:03:39.506847 214170 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/proxy-client.key ...
I0317 11:03:39.506858 214170 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/proxy-client.key: {Name:mkab9c73fb0996f297c8d00fd9a321c779c566ed Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0317 11:03:39.507072 214170 certs.go:484] found cert: /home/jenkins/minikube-integration/20535-2275/.minikube/certs/7585.pem (1338 bytes)
W0317 11:03:39.507106 214170 certs.go:480] ignoring /home/jenkins/minikube-integration/20535-2275/.minikube/certs/7585_empty.pem, impossibly tiny 0 bytes
I0317 11:03:39.507113 214170 certs.go:484] found cert: /home/jenkins/minikube-integration/20535-2275/.minikube/certs/ca-key.pem (1679 bytes)
I0317 11:03:39.507135 214170 certs.go:484] found cert: /home/jenkins/minikube-integration/20535-2275/.minikube/certs/ca.pem (1078 bytes)
I0317 11:03:39.507154 214170 certs.go:484] found cert: /home/jenkins/minikube-integration/20535-2275/.minikube/certs/cert.pem (1123 bytes)
I0317 11:03:39.507177 214170 certs.go:484] found cert: /home/jenkins/minikube-integration/20535-2275/.minikube/certs/key.pem (1679 bytes)
I0317 11:03:39.507218 214170 certs.go:484] found cert: /home/jenkins/minikube-integration/20535-2275/.minikube/files/etc/ssl/certs/75852.pem (1708 bytes)
I0317 11:03:39.507779 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0317 11:03:39.533866 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I0317 11:03:39.558866 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0317 11:03:39.584359 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I0317 11:03:39.609751 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1432 bytes)
I0317 11:03:39.633966 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0317 11:03:39.660197 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0317 11:03:39.686029 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/profiles/scheduled-stop-375543/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
I0317 11:03:39.713703 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0317 11:03:39.740475 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/certs/7585.pem --> /usr/share/ca-certificates/7585.pem (1338 bytes)
I0317 11:03:39.765953 214170 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20535-2275/.minikube/files/etc/ssl/certs/75852.pem --> /usr/share/ca-certificates/75852.pem (1708 bytes)
I0317 11:03:39.791844 214170 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0317 11:03:39.811097 214170 ssh_runner.go:195] Run: openssl version
I0317 11:03:39.817049 214170 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/7585.pem && ln -fs /usr/share/ca-certificates/7585.pem /etc/ssl/certs/7585.pem"
I0317 11:03:39.827818 214170 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/7585.pem
I0317 11:03:39.831840 214170 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 Mar 17 10:33 /usr/share/ca-certificates/7585.pem
I0317 11:03:39.831895 214170 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/7585.pem
I0317 11:03:39.839134 214170 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/7585.pem /etc/ssl/certs/51391683.0"
I0317 11:03:39.849233 214170 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/75852.pem && ln -fs /usr/share/ca-certificates/75852.pem /etc/ssl/certs/75852.pem"
I0317 11:03:39.859054 214170 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/75852.pem
I0317 11:03:39.862844 214170 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 Mar 17 10:33 /usr/share/ca-certificates/75852.pem
I0317 11:03:39.862900 214170 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/75852.pem
I0317 11:03:39.870040 214170 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/75852.pem /etc/ssl/certs/3ec20f2e.0"
I0317 11:03:39.880263 214170 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0317 11:03:39.890261 214170 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0317 11:03:39.893942 214170 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Mar 17 10:26 /usr/share/ca-certificates/minikubeCA.pem
I0317 11:03:39.893999 214170 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0317 11:03:39.901619 214170 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0317 11:03:39.911728 214170 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0317 11:03:39.915102 214170 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0317 11:03:39.915154 214170 kubeadm.go:392] StartCluster: {Name:scheduled-stop-375543 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.46-1741860993-20523@sha256:cd976907fa4d517c84fff1e5ef773b9fb3c738c4e1ded824ea5133470a66e185 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.32.2 ClusterName:scheduled-stop-375543 Namespace:default APIServerHAVIP: APIServerName:minikubeC
A APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.32.2 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP
: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0317 11:03:39.915264 214170 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0317 11:03:39.933125 214170 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0317 11:03:39.942785 214170 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0317 11:03:39.952563 214170 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0317 11:03:39.952623 214170 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0317 11:03:39.961864 214170 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0317 11:03:39.961872 214170 kubeadm.go:157] found existing configuration files:
I0317 11:03:39.961922 214170 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0317 11:03:39.970938 214170 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0317 11:03:39.971021 214170 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0317 11:03:39.980019 214170 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0317 11:03:39.989386 214170 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0317 11:03:39.989439 214170 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0317 11:03:39.998881 214170 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0317 11:03:40.020074 214170 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0317 11:03:40.020148 214170 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0317 11:03:40.032669 214170 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0317 11:03:40.044077 214170 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0317 11:03:40.044144 214170 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0317 11:03:40.054963 214170 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.32.2:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0317 11:03:40.128347 214170 kubeadm.go:310] [WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2
I0317 11:03:40.128596 214170 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1077-aws\n", err: exit status 1
I0317 11:03:40.194841 214170 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0317 11:03:57.821336 214170 kubeadm.go:310] [init] Using Kubernetes version: v1.32.2
I0317 11:03:57.821383 214170 kubeadm.go:310] [preflight] Running pre-flight checks
I0317 11:03:57.821478 214170 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0317 11:03:57.821532 214170 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1077-aws[0m
I0317 11:03:57.821583 214170 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0317 11:03:57.821627 214170 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0317 11:03:57.821674 214170 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0317 11:03:57.821720 214170 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0317 11:03:57.821767 214170 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0317 11:03:57.821814 214170 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0317 11:03:57.821861 214170 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0317 11:03:57.821905 214170 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0317 11:03:57.821952 214170 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0317 11:03:57.821997 214170 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0317 11:03:57.822067 214170 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0317 11:03:57.822159 214170 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0317 11:03:57.822252 214170 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0317 11:03:57.822313 214170 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0317 11:03:57.825444 214170 out.go:235] - Generating certificates and keys ...
I0317 11:03:57.825544 214170 kubeadm.go:310] [certs] Using existing ca certificate authority
I0317 11:03:57.825607 214170 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0317 11:03:57.825678 214170 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0317 11:03:57.825734 214170 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0317 11:03:57.825794 214170 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0317 11:03:57.825843 214170 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0317 11:03:57.825895 214170 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0317 11:03:57.826017 214170 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [localhost scheduled-stop-375543] and IPs [192.168.76.2 127.0.0.1 ::1]
I0317 11:03:57.826069 214170 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0317 11:03:57.826188 214170 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [localhost scheduled-stop-375543] and IPs [192.168.76.2 127.0.0.1 ::1]
I0317 11:03:57.826264 214170 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0317 11:03:57.826326 214170 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0317 11:03:57.826369 214170 kubeadm.go:310] [certs] Generating "sa" key and public key
I0317 11:03:57.826423 214170 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0317 11:03:57.826472 214170 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0317 11:03:57.826526 214170 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0317 11:03:57.826599 214170 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0317 11:03:57.826660 214170 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0317 11:03:57.826719 214170 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0317 11:03:57.826799 214170 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0317 11:03:57.826863 214170 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0317 11:03:57.829933 214170 out.go:235] - Booting up control plane ...
I0317 11:03:57.830096 214170 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0317 11:03:57.830208 214170 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0317 11:03:57.830311 214170 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0317 11:03:57.830453 214170 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0317 11:03:57.830576 214170 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0317 11:03:57.830633 214170 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0317 11:03:57.830794 214170 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0317 11:03:57.831121 214170 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0317 11:03:57.831197 214170 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 1.000910324s
I0317 11:03:57.831272 214170 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0317 11:03:57.831339 214170 kubeadm.go:310] [api-check] The API server is healthy after 6.501707079s
I0317 11:03:57.831452 214170 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0317 11:03:57.831585 214170 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0317 11:03:57.831647 214170 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0317 11:03:57.831849 214170 kubeadm.go:310] [mark-control-plane] Marking the node scheduled-stop-375543 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0317 11:03:57.831924 214170 kubeadm.go:310] [bootstrap-token] Using token: wdhg74.494qbhl39gzlb040
I0317 11:03:57.834867 214170 out.go:235] - Configuring RBAC rules ...
I0317 11:03:57.835027 214170 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0317 11:03:57.835110 214170 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0317 11:03:57.835298 214170 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0317 11:03:57.835454 214170 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0317 11:03:57.835578 214170 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0317 11:03:57.835664 214170 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0317 11:03:57.835777 214170 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0317 11:03:57.835819 214170 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0317 11:03:57.835863 214170 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0317 11:03:57.835866 214170 kubeadm.go:310]
I0317 11:03:57.835925 214170 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0317 11:03:57.835928 214170 kubeadm.go:310]
I0317 11:03:57.836004 214170 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0317 11:03:57.836007 214170 kubeadm.go:310]
I0317 11:03:57.836034 214170 kubeadm.go:310] mkdir -p $HOME/.kube
I0317 11:03:57.836091 214170 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0317 11:03:57.836140 214170 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0317 11:03:57.836143 214170 kubeadm.go:310]
I0317 11:03:57.836216 214170 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0317 11:03:57.836221 214170 kubeadm.go:310]
I0317 11:03:57.836273 214170 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0317 11:03:57.836276 214170 kubeadm.go:310]
I0317 11:03:57.836331 214170 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0317 11:03:57.836412 214170 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0317 11:03:57.836483 214170 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0317 11:03:57.836486 214170 kubeadm.go:310]
I0317 11:03:57.836584 214170 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0317 11:03:57.836661 214170 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0317 11:03:57.836664 214170 kubeadm.go:310]
I0317 11:03:57.836750 214170 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token wdhg74.494qbhl39gzlb040 \
I0317 11:03:57.836858 214170 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:3cf1b898defecfe95004a6aeb68435866ffe3e9555f25601e9f36ac23e531178 \
I0317 11:03:57.836878 214170 kubeadm.go:310] --control-plane
I0317 11:03:57.836881 214170 kubeadm.go:310]
I0317 11:03:57.836974 214170 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0317 11:03:57.836978 214170 kubeadm.go:310]
I0317 11:03:57.837058 214170 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token wdhg74.494qbhl39gzlb040 \
I0317 11:03:57.837179 214170 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:3cf1b898defecfe95004a6aeb68435866ffe3e9555f25601e9f36ac23e531178
I0317 11:03:57.837186 214170 cni.go:84] Creating CNI manager for ""
I0317 11:03:57.837198 214170 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0317 11:03:57.842263 214170 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0317 11:03:57.845090 214170 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0317 11:03:57.854357 214170 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0317 11:03:57.875947 214170 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0317 11:03:57.876001 214170 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.32.2/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0317 11:03:57.876082 214170 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.32.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes scheduled-stop-375543 minikube.k8s.io/updated_at=2025_03_17T11_03_57_0700 minikube.k8s.io/version=v1.35.0 minikube.k8s.io/commit=28b3ce799b018a38b7c40f89b465976263272e76 minikube.k8s.io/name=scheduled-stop-375543 minikube.k8s.io/primary=true
I0317 11:03:58.048040 214170 ops.go:34] apiserver oom_adj: -16
I0317 11:03:58.048049 214170 kubeadm.go:1113] duration metric: took 172.100151ms to wait for elevateKubeSystemPrivileges
I0317 11:03:58.048071 214170 kubeadm.go:394] duration metric: took 18.13292051s to StartCluster
I0317 11:03:58.048087 214170 settings.go:142] acquiring lock: {Name:mkf9a97e0ede49aff4670aa5bdac2bb5aae4b749 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0317 11:03:58.048168 214170 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/20535-2275/kubeconfig
I0317 11:03:58.048866 214170 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20535-2275/kubeconfig: {Name:mk24b6be897a3c260c2cdc2291f04289b584d621 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0317 11:03:58.049098 214170 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.32.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0317 11:03:58.049106 214170 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.32.2 ContainerRuntime:docker ControlPlane:true Worker:true}
I0317 11:03:58.049368 214170 config.go:182] Loaded profile config "scheduled-stop-375543": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.32.2
I0317 11:03:58.049412 214170 addons.go:511] enable addons start: toEnable=map[ambassador:false amd-gpu-device-plugin:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:false volcano:false volumesnapshots:false yakd:false]
I0317 11:03:58.049473 214170 addons.go:69] Setting storage-provisioner=true in profile "scheduled-stop-375543"
I0317 11:03:58.049486 214170 addons.go:238] Setting addon storage-provisioner=true in "scheduled-stop-375543"
I0317 11:03:58.049509 214170 host.go:66] Checking if "scheduled-stop-375543" exists ...
I0317 11:03:58.049875 214170 addons.go:69] Setting default-storageclass=true in profile "scheduled-stop-375543"
I0317 11:03:58.049891 214170 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "scheduled-stop-375543"
I0317 11:03:58.050017 214170 cli_runner.go:164] Run: docker container inspect scheduled-stop-375543 --format={{.State.Status}}
I0317 11:03:58.050249 214170 cli_runner.go:164] Run: docker container inspect scheduled-stop-375543 --format={{.State.Status}}
I0317 11:03:58.054933 214170 out.go:177] * Verifying Kubernetes components...
I0317 11:03:58.057808 214170 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0317 11:03:58.086920 214170 addons.go:238] Setting addon default-storageclass=true in "scheduled-stop-375543"
I0317 11:03:58.086950 214170 host.go:66] Checking if "scheduled-stop-375543" exists ...
I0317 11:03:58.087391 214170 cli_runner.go:164] Run: docker container inspect scheduled-stop-375543 --format={{.State.Status}}
I0317 11:03:58.097874 214170 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0317 11:03:58.102111 214170 addons.go:435] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0317 11:03:58.102123 214170 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0317 11:03:58.102192 214170 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-375543
I0317 11:03:58.125537 214170 addons.go:435] installing /etc/kubernetes/addons/storageclass.yaml
I0317 11:03:58.125550 214170 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0317 11:03:58.125612 214170 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-375543
I0317 11:03:58.175976 214170 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/20535-2275/.minikube/machines/scheduled-stop-375543/id_rsa Username:docker}
I0317 11:03:58.201494 214170 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32970 SSHKeyPath:/home/jenkins/minikube-integration/20535-2275/.minikube/machines/scheduled-stop-375543/id_rsa Username:docker}
I0317 11:03:58.408147 214170 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.32.2/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0317 11:03:58.439297 214170 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.32.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.76.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.32.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0317 11:03:58.439427 214170 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0317 11:03:58.465656 214170 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.32.2/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0317 11:03:58.846074 214170 start.go:971] {"host.minikube.internal": 192.168.76.1} host record injected into CoreDNS's ConfigMap
I0317 11:03:58.847873 214170 api_server.go:52] waiting for apiserver process to appear ...
I0317 11:03:58.848034 214170 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0317 11:03:59.042735 214170 api_server.go:72] duration metric: took 993.603257ms to wait for apiserver process to appear ...
I0317 11:03:59.042759 214170 api_server.go:88] waiting for apiserver healthz status ...
I0317 11:03:59.042785 214170 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
I0317 11:03:59.046596 214170 out.go:177] * Enabled addons: default-storageclass, storage-provisioner
I0317 11:03:59.048859 214170 addons.go:514] duration metric: took 999.445789ms for enable addons: enabled=[default-storageclass storage-provisioner]
I0317 11:03:59.055575 214170 api_server.go:279] https://192.168.76.2:8443/healthz returned 200:
ok
I0317 11:03:59.057098 214170 api_server.go:141] control plane version: v1.32.2
I0317 11:03:59.057115 214170 api_server.go:131] duration metric: took 14.350383ms to wait for apiserver health ...
I0317 11:03:59.057122 214170 system_pods.go:43] waiting for kube-system pods to appear ...
I0317 11:03:59.061232 214170 system_pods.go:59] 5 kube-system pods found
I0317 11:03:59.061254 214170 system_pods.go:61] "etcd-scheduled-stop-375543" [6d61f537-b807-45d5-9a76-27b056a9c956] Running / Ready:ContainersNotReady (containers with unready status: [etcd]) / ContainersReady:ContainersNotReady (containers with unready status: [etcd])
I0317 11:03:59.061261 214170 system_pods.go:61] "kube-apiserver-scheduled-stop-375543" [232330a4-3f1e-4523-a812-cf6503ccc27e] Running / Ready:ContainersNotReady (containers with unready status: [kube-apiserver]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-apiserver])
I0317 11:03:59.061268 214170 system_pods.go:61] "kube-controller-manager-scheduled-stop-375543" [7da44fae-f083-4cb7-8129-e22aba9b063b] Running / Ready:ContainersNotReady (containers with unready status: [kube-controller-manager]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-controller-manager])
I0317 11:03:59.061273 214170 system_pods.go:61] "kube-scheduler-scheduled-stop-375543" [690ac331-9b13-4fe6-a2ea-98629bdd5374] Running / Ready:ContainersNotReady (containers with unready status: [kube-scheduler]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-scheduler])
I0317 11:03:59.061279 214170 system_pods.go:61] "storage-provisioner" [b0b03cba-7aea-4a43-9ef1-5658eedc8516] Pending: PodScheduled:Unschedulable (0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.)
I0317 11:03:59.061284 214170 system_pods.go:74] duration metric: took 4.157503ms to wait for pod list to return data ...
I0317 11:03:59.061295 214170 kubeadm.go:582] duration metric: took 1.012167225s to wait for: map[apiserver:true system_pods:true]
I0317 11:03:59.061306 214170 node_conditions.go:102] verifying NodePressure condition ...
I0317 11:03:59.068862 214170 node_conditions.go:122] node storage ephemeral capacity is 203034800Ki
I0317 11:03:59.068881 214170 node_conditions.go:123] node cpu capacity is 2
I0317 11:03:59.068892 214170 node_conditions.go:105] duration metric: took 7.58202ms to run NodePressure ...
I0317 11:03:59.068903 214170 start.go:241] waiting for startup goroutines ...
I0317 11:03:59.349716 214170 kapi.go:214] "coredns" deployment in "kube-system" namespace and "scheduled-stop-375543" context rescaled to 1 replicas
I0317 11:03:59.349751 214170 start.go:246] waiting for cluster config update ...
I0317 11:03:59.349763 214170 start.go:255] writing updated cluster config ...
I0317 11:03:59.350067 214170 ssh_runner.go:195] Run: rm -f paused
I0317 11:03:59.416576 214170 start.go:600] kubectl: 1.32.3, cluster: 1.32.2 (minor skew: 0)
I0317 11:03:59.419760 214170 out.go:177] * Done! kubectl is now configured to use "scheduled-stop-375543" cluster and "default" namespace by default
==> Docker <==
Mar 17 11:03:37 scheduled-stop-375543 dockerd[1354]: time="2025-03-17T11:03:37.635177720Z" level=info msg="Loading containers: start."
Mar 17 11:03:37 scheduled-stop-375543 dockerd[1354]: time="2025-03-17T11:03:37.810642346Z" level=info msg="Loading containers: done."
Mar 17 11:03:37 scheduled-stop-375543 dockerd[1354]: time="2025-03-17T11:03:37.820804015Z" level=info msg="Docker daemon" commit=bbd0a17 containerd-snapshotter=false storage-driver=overlay2 version=28.0.1
Mar 17 11:03:37 scheduled-stop-375543 dockerd[1354]: time="2025-03-17T11:03:37.820892983Z" level=info msg="Initializing buildkit"
Mar 17 11:03:37 scheduled-stop-375543 dockerd[1354]: time="2025-03-17T11:03:37.839310620Z" level=info msg="Completed buildkit initialization"
Mar 17 11:03:37 scheduled-stop-375543 dockerd[1354]: time="2025-03-17T11:03:37.845139302Z" level=info msg="Daemon has completed initialization"
Mar 17 11:03:37 scheduled-stop-375543 dockerd[1354]: time="2025-03-17T11:03:37.845342314Z" level=info msg="API listen on /var/run/docker.sock"
Mar 17 11:03:37 scheduled-stop-375543 systemd[1]: Started Docker Application Container Engine.
Mar 17 11:03:37 scheduled-stop-375543 dockerd[1354]: time="2025-03-17T11:03:37.845477060Z" level=info msg="API listen on [::]:2376"
Mar 17 11:03:38 scheduled-stop-375543 systemd[1]: Starting CRI Interface for Docker Application Container Engine...
Mar 17 11:03:38 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:38Z" level=info msg="Starting cri-dockerd dev (HEAD)"
Mar 17 11:03:38 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:38Z" level=info msg="Connecting to docker on the Endpoint unix:///var/run/docker.sock"
Mar 17 11:03:38 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:38Z" level=info msg="Start docker client with request timeout 0s"
Mar 17 11:03:38 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:38Z" level=info msg="Hairpin mode is set to hairpin-veth"
Mar 17 11:03:38 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:38Z" level=info msg="Loaded network plugin cni"
Mar 17 11:03:38 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:38Z" level=info msg="Docker cri networking managed by network plugin cni"
Mar 17 11:03:38 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:38Z" level=info msg="Setting cgroupDriver cgroupfs"
Mar 17 11:03:38 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:38Z" level=info msg="Docker cri received runtime config &RuntimeConfig{NetworkConfig:&NetworkConfig{PodCidr:,},}"
Mar 17 11:03:38 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:38Z" level=info msg="Starting the GRPC backend for the Docker CRI interface."
Mar 17 11:03:38 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:38Z" level=info msg="Start cri-dockerd grpc backend"
Mar 17 11:03:38 scheduled-stop-375543 systemd[1]: Started CRI Interface for Docker Application Container Engine.
Mar 17 11:03:50 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:50Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/8524bf97d308ea4af876a27047b4649b28f1333deed7b45dcd8e40a90def1c34/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options ndots:0 edns0 trust-ad]"
Mar 17 11:03:50 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:50Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/1d08b05eb98ad868e81df2493a3a764f4113b240f67820142c230c343bb8cc00/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options edns0 trust-ad ndots:0]"
Mar 17 11:03:50 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:50Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/8663ce8457bfbda81917c16bfd009fdd723509cb3dbe5c3ea19b10b19e0abbfd/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options edns0 trust-ad ndots:0]"
Mar 17 11:03:50 scheduled-stop-375543 cri-dockerd[1635]: time="2025-03-17T11:03:50Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/0ef6340d93ad9987020519d4f2c7cdf501cc0169a7cb5503090049b4ade93a8d/resolv.conf as [nameserver 192.168.76.1 search us-east-2.compute.internal options edns0 trust-ad ndots:0]"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
0b9694acb13ea 3c9285acfd2ff 11 seconds ago Running kube-controller-manager 0 0ef6340d93ad9 kube-controller-manager-scheduled-stop-375543
52eb16128508c 7fc9d4aa817aa 11 seconds ago Running etcd 0 8663ce8457bfb etcd-scheduled-stop-375543
3a4e14a94e591 6417e1437b6d9 11 seconds ago Running kube-apiserver 0 1d08b05eb98ad kube-apiserver-scheduled-stop-375543
44548ca91013a 82dfa03f692fb 11 seconds ago Running kube-scheduler 0 8524bf97d308e kube-scheduler-scheduled-stop-375543
==> describe nodes <==
Name: scheduled-stop-375543
Roles: control-plane
Labels: beta.kubernetes.io/arch=arm64
beta.kubernetes.io/os=linux
kubernetes.io/arch=arm64
kubernetes.io/hostname=scheduled-stop-375543
kubernetes.io/os=linux
minikube.k8s.io/commit=28b3ce799b018a38b7c40f89b465976263272e76
minikube.k8s.io/name=scheduled-stop-375543
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2025_03_17T11_03_57_0700
minikube.k8s.io/version=v1.35.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Mon, 17 Mar 2025 11:03:54 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: scheduled-stop-375543
AcquireTime: <unset>
RenewTime: Mon, 17 Mar 2025 11:03:57 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Mon, 17 Mar 2025 11:03:54 +0000 Mon, 17 Mar 2025 11:03:51 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Mon, 17 Mar 2025 11:03:54 +0000 Mon, 17 Mar 2025 11:03:51 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Mon, 17 Mar 2025 11:03:54 +0000 Mon, 17 Mar 2025 11:03:51 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Mon, 17 Mar 2025 11:03:54 +0000 Mon, 17 Mar 2025 11:03:54 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.76.2
Hostname: scheduled-stop-375543
Capacity:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022296Ki
pods: 110
Allocatable:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022296Ki
pods: 110
System Info:
Machine ID: e43703b92a1744279a500c7d662c915b
System UUID: 932fd146-a983-41ed-bf33-885ee464428a
Boot ID: 9f3f55f4-20c8-491e-a739-6b58d0efa6f5
Kernel Version: 5.15.0-1077-aws
OS Image: Ubuntu 22.04.5 LTS
Operating System: linux
Architecture: arm64
Container Runtime Version: docker://28.0.1
Kubelet Version: v1.32.2
Kube-Proxy Version: v1.32.2
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (5 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
kube-system etcd-scheduled-stop-375543 100m (5%) 0 (0%) 100Mi (1%) 0 (0%) 4s
kube-system kube-apiserver-scheduled-stop-375543 250m (12%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-controller-manager-scheduled-stop-375543 200m (10%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-scheduler-scheduled-stop-375543 100m (5%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 2s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 650m (32%) 0 (0%)
memory 100Mi (1%) 0 (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
hugepages-32Mi 0 (0%) 0 (0%)
hugepages-64Ki 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal NodeAllocatableEnforced 12s kubelet Updated Node Allocatable limit across pods
Normal Starting 12s kubelet Starting kubelet.
Warning CgroupV1 12s kubelet cgroup v1 support is in maintenance mode, please migrate to cgroup v2
Normal NodeHasSufficientMemory 12s (x8 over 12s) kubelet Node scheduled-stop-375543 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12s (x8 over 12s) kubelet Node scheduled-stop-375543 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12s (x7 over 12s) kubelet Node scheduled-stop-375543 status is now: NodeHasSufficientPID
Normal Starting 4s kubelet Starting kubelet.
Warning CgroupV1 4s kubelet cgroup v1 support is in maintenance mode, please migrate to cgroup v2
Normal NodeAllocatableEnforced 4s kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 4s kubelet Node scheduled-stop-375543 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 4s kubelet Node scheduled-stop-375543 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 4s kubelet Node scheduled-stop-375543 status is now: NodeHasSufficientPID
Normal RegisteredNode 0s node-controller Node scheduled-stop-375543 event: Registered Node scheduled-stop-375543 in Controller
==> dmesg <==
[Mar17 10:17] ACPI: SRAT not present
[ +0.000000] ACPI: SRAT not present
[ +0.000000] SPI driver altr_a10sr has no spi_device_id for altr,a10sr
[ +0.014732] device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log.
[ +0.507118] systemd[1]: Configuration file /run/systemd/system/netplan-ovs-cleanup.service is marked world-inaccessible. This has no effect as configuration data is accessible via APIs without restrictions. Proceeding anyway.
[ +0.033958] systemd[1]: /lib/systemd/system/snapd.service:23: Unknown key name 'RestartMode' in section 'Service', ignoring.
[ +0.843051] ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy.
[ +6.678231] kauditd_printk_skb: 36 callbacks suppressed
[Mar17 10:38] hrtimer: interrupt took 5033620 ns
==> etcd [52eb16128508] <==
{"level":"info","ts":"2025-03-17T11:03:50.896740Z","caller":"embed/etcd.go:729","msg":"starting with client TLS","tls-info":"cert = /var/lib/minikube/certs/etcd/server.crt, key = /var/lib/minikube/certs/etcd/server.key, client-cert=, client-key=, trusted-ca = /var/lib/minikube/certs/etcd/ca.crt, client-cert-auth = true, crl-file = ","cipher-suites":[]}
{"level":"info","ts":"2025-03-17T11:03:50.897020Z","caller":"embed/etcd.go:280","msg":"now serving peer/client/metrics","local-member-id":"ea7e25599daad906","initial-advertise-peer-urls":["https://192.168.76.2:2380"],"listen-peer-urls":["https://192.168.76.2:2380"],"advertise-client-urls":["https://192.168.76.2:2379"],"listen-client-urls":["https://127.0.0.1:2379","https://192.168.76.2:2379"],"listen-metrics-urls":["http://127.0.0.1:2381"]}
{"level":"info","ts":"2025-03-17T11:03:50.897046Z","caller":"embed/etcd.go:871","msg":"serving metrics","address":"http://127.0.0.1:2381"}
{"level":"info","ts":"2025-03-17T11:03:50.897102Z","caller":"embed/etcd.go:600","msg":"serving peer traffic","address":"192.168.76.2:2380"}
{"level":"info","ts":"2025-03-17T11:03:50.897114Z","caller":"embed/etcd.go:572","msg":"cmux::serve","address":"192.168.76.2:2380"}
{"level":"info","ts":"2025-03-17T11:03:51.271068Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 is starting a new election at term 1"}
{"level":"info","ts":"2025-03-17T11:03:51.271191Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 became pre-candidate at term 1"}
{"level":"info","ts":"2025-03-17T11:03:51.271275Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 received MsgPreVoteResp from ea7e25599daad906 at term 1"}
{"level":"info","ts":"2025-03-17T11:03:51.271314Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 became candidate at term 2"}
{"level":"info","ts":"2025-03-17T11:03:51.271370Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 received MsgVoteResp from ea7e25599daad906 at term 2"}
{"level":"info","ts":"2025-03-17T11:03:51.271408Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 became leader at term 2"}
{"level":"info","ts":"2025-03-17T11:03:51.271458Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: ea7e25599daad906 elected leader ea7e25599daad906 at term 2"}
{"level":"info","ts":"2025-03-17T11:03:51.275169Z","caller":"etcdserver/server.go:2651","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2025-03-17T11:03:51.279248Z","caller":"etcdserver/server.go:2140","msg":"published local member to cluster through raft","local-member-id":"ea7e25599daad906","local-member-attributes":"{Name:scheduled-stop-375543 ClientURLs:[https://192.168.76.2:2379]}","request-path":"/0/members/ea7e25599daad906/attributes","cluster-id":"6f20f2c4b2fb5f8a","publish-timeout":"7s"}
{"level":"info","ts":"2025-03-17T11:03:51.279476Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2025-03-17T11:03:51.280364Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2025-03-17T11:03:51.281279Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.76.2:2379"}
{"level":"info","ts":"2025-03-17T11:03:51.281653Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"6f20f2c4b2fb5f8a","local-member-id":"ea7e25599daad906","cluster-version":"3.5"}
{"level":"info","ts":"2025-03-17T11:03:51.281796Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2025-03-17T11:03:51.281862Z","caller":"etcdserver/server.go:2675","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2025-03-17T11:03:51.281903Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2025-03-17T11:03:51.287727Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2025-03-17T11:03:51.288731Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2025-03-17T11:03:51.288898Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2025-03-17T11:03:51.289016Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
==> kernel <==
11:04:01 up 46 min, 0 users, load average: 1.37, 1.58, 1.80
Linux scheduled-stop-375543 5.15.0-1077-aws #84~20.04.1-Ubuntu SMP Mon Jan 20 22:14:27 UTC 2025 aarch64 aarch64 aarch64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.5 LTS"
==> kube-apiserver [3a4e14a94e59] <==
I0317 11:03:54.695118 1 shared_informer.go:320] Caches are synced for node_authorizer
E0317 11:03:54.707991 1 controller.go:145] "Failed to ensure lease exists, will retry" err="namespaces \"kube-system\" not found" interval="200ms"
I0317 11:03:54.720424 1 cache.go:39] Caches are synced for APIServiceRegistrationController controller
I0317 11:03:54.720692 1 cache.go:39] Caches are synced for RemoteAvailability controller
I0317 11:03:54.721218 1 cache.go:39] Caches are synced for LocalAvailability controller
I0317 11:03:54.722604 1 controller.go:615] quota admission added evaluator for: namespaces
I0317 11:03:54.723625 1 apf_controller.go:382] Running API Priority and Fairness config worker
I0317 11:03:54.724352 1 apf_controller.go:385] Running API Priority and Fairness periodic rebalancing process
I0317 11:03:54.726261 1 shared_informer.go:320] Caches are synced for configmaps
I0317 11:03:54.726683 1 handler_discovery.go:451] Starting ResourceDiscoveryManager
I0317 11:03:54.726922 1 shared_informer.go:320] Caches are synced for cluster_authentication_trust_controller
I0317 11:03:54.911471 1 controller.go:615] quota admission added evaluator for: leases.coordination.k8s.io
I0317 11:03:55.426823 1 storage_scheduling.go:95] created PriorityClass system-node-critical with value 2000001000
I0317 11:03:55.432236 1 storage_scheduling.go:95] created PriorityClass system-cluster-critical with value 2000000000
I0317 11:03:55.432467 1 storage_scheduling.go:111] all system priority classes are created successfully or already exist.
I0317 11:03:56.165641 1 controller.go:615] quota admission added evaluator for: roles.rbac.authorization.k8s.io
I0317 11:03:56.233554 1 controller.go:615] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
I0317 11:03:56.339567 1 alloc.go:330] "allocated clusterIPs" service="default/kubernetes" clusterIPs={"IPv4":"10.96.0.1"}
W0317 11:03:56.351317 1 lease.go:265] Resetting endpoints for master service "kubernetes" to [192.168.76.2]
I0317 11:03:56.352791 1 controller.go:615] quota admission added evaluator for: endpoints
I0317 11:03:56.358066 1 controller.go:615] quota admission added evaluator for: endpointslices.discovery.k8s.io
I0317 11:03:56.520469 1 controller.go:615] quota admission added evaluator for: serviceaccounts
I0317 11:03:57.227402 1 controller.go:615] quota admission added evaluator for: deployments.apps
I0317 11:03:57.240267 1 alloc.go:330] "allocated clusterIPs" service="kube-system/kube-dns" clusterIPs={"IPv4":"10.96.0.10"}
I0317 11:03:57.268005 1 controller.go:615] quota admission added evaluator for: daemonsets.apps
==> kube-controller-manager [0b9694acb13e] <==
I0317 11:04:01.076372 1 shared_informer.go:320] Caches are synced for TTL
I0317 11:04:01.083388 1 shared_informer.go:320] Caches are synced for certificate-csrapproving
I0317 11:04:01.085635 1 shared_informer.go:320] Caches are synced for cronjob
I0317 11:04:01.092294 1 shared_informer.go:320] Caches are synced for expand
I0317 11:04:01.092642 1 shared_informer.go:320] Caches are synced for namespace
I0317 11:04:01.095754 1 shared_informer.go:320] Caches are synced for garbage collector
I0317 11:04:01.103652 1 shared_informer.go:320] Caches are synced for disruption
I0317 11:04:01.114699 1 shared_informer.go:320] Caches are synced for HPA
I0317 11:04:01.115022 1 shared_informer.go:320] Caches are synced for garbage collector
I0317 11:04:01.115155 1 garbagecollector.go:154] "Garbage collector: all resource monitors have synced" logger="garbage-collector-controller"
I0317 11:04:01.115171 1 garbagecollector.go:157] "Proceeding to collect garbage" logger="garbage-collector-controller"
I0317 11:04:01.116062 1 shared_informer.go:320] Caches are synced for taint
I0317 11:04:01.116324 1 node_lifecycle_controller.go:1234] "Initializing eviction metric for zone" logger="node-lifecycle-controller" zone=""
I0317 11:04:01.116506 1 shared_informer.go:320] Caches are synced for deployment
I0317 11:04:01.116717 1 node_lifecycle_controller.go:886] "Missing timestamp for Node. Assuming now as a timestamp" logger="node-lifecycle-controller" node="scheduled-stop-375543"
I0317 11:04:01.116980 1 shared_informer.go:320] Caches are synced for crt configmap
I0317 11:04:01.117598 1 shared_informer.go:320] Caches are synced for TTL after finished
I0317 11:04:01.118304 1 shared_informer.go:320] Caches are synced for ReplicaSet
I0317 11:04:01.118760 1 shared_informer.go:320] Caches are synced for endpoint_slice_mirroring
I0317 11:04:01.118902 1 shared_informer.go:320] Caches are synced for ClusterRoleAggregator
I0317 11:04:01.119297 1 node_lifecycle_controller.go:1080] "Controller detected that zone is now in new state" logger="node-lifecycle-controller" zone="" newState="Normal"
I0317 11:04:01.119355 1 shared_informer.go:320] Caches are synced for PV protection
I0317 11:04:01.120233 1 shared_informer.go:320] Caches are synced for attach detach
I0317 11:04:01.139518 1 shared_informer.go:320] Caches are synced for PVC protection
I0317 11:04:01.139746 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="scheduled-stop-375543"
==> kube-scheduler [44548ca91013] <==
W0317 11:03:55.477881 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E0317 11:03:55.478152 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0317 11:03:55.495847 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0317 11:03:55.496172 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0317 11:03:55.529915 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0317 11:03:55.530186 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0317 11:03:55.594431 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.VolumeAttachment: volumeattachments.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "volumeattachments" in API group "storage.k8s.io" at the cluster scope
E0317 11:03:55.594479 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.VolumeAttachment: failed to list *v1.VolumeAttachment: volumeattachments.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"volumeattachments\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0317 11:03:55.647104 1 reflector.go:569] runtime/asm_arm64.s:1223: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0317 11:03:55.647374 1 reflector.go:166] "Unhandled Error" err="runtime/asm_arm64.s:1223: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
W0317 11:03:55.697804 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0317 11:03:55.697845 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0317 11:03:55.706884 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0317 11:03:55.706931 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0317 11:03:55.752397 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0317 11:03:55.752564 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0317 11:03:55.792223 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0317 11:03:55.792278 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0317 11:03:55.794681 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0317 11:03:55.794727 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0317 11:03:55.839374 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0317 11:03:55.839435 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0317 11:03:55.855166 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0317 11:03:55.855219 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
I0317 11:03:57.658137 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Mar 17 11:03:57 scheduled-stop-375543 kubelet[2479]: I0317 11:03:57.566200 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-data\" (UniqueName: \"kubernetes.io/host-path/abde8736338038cf2e2191b58eb567b3-etcd-data\") pod \"etcd-scheduled-stop-375543\" (UID: \"abde8736338038cf2e2191b58eb567b3\") " pod="kube-system/etcd-scheduled-stop-375543"
Mar 17 11:03:57 scheduled-stop-375543 kubelet[2479]: I0317 11:03:57.566231 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b42bb468076de42e247b02c0d1d376fd-ca-certs\") pod \"kube-apiserver-scheduled-stop-375543\" (UID: \"b42bb468076de42e247b02c0d1d376fd\") " pod="kube-system/kube-apiserver-scheduled-stop-375543"
Mar 17 11:03:57 scheduled-stop-375543 kubelet[2479]: I0317 11:03:57.566261 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b42bb468076de42e247b02c0d1d376fd-usr-share-ca-certificates\") pod \"kube-apiserver-scheduled-stop-375543\" (UID: \"b42bb468076de42e247b02c0d1d376fd\") " pod="kube-system/kube-apiserver-scheduled-stop-375543"
Mar 17 11:03:57 scheduled-stop-375543 kubelet[2479]: I0317 11:03:57.566294 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/da113bf4006ce6881d24cfcd88d12277-etc-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-375543\" (UID: \"da113bf4006ce6881d24cfcd88d12277\") " pod="kube-system/kube-controller-manager-scheduled-stop-375543"
Mar 17 11:03:57 scheduled-stop-375543 kubelet[2479]: I0317 11:03:57.566319 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/da113bf4006ce6881d24cfcd88d12277-ca-certs\") pod \"kube-controller-manager-scheduled-stop-375543\" (UID: \"da113bf4006ce6881d24cfcd88d12277\") " pod="kube-system/kube-controller-manager-scheduled-stop-375543"
Mar 17 11:03:57 scheduled-stop-375543 kubelet[2479]: I0317 11:03:57.566338 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/da113bf4006ce6881d24cfcd88d12277-usr-local-share-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-375543\" (UID: \"da113bf4006ce6881d24cfcd88d12277\") " pod="kube-system/kube-controller-manager-scheduled-stop-375543"
Mar 17 11:03:57 scheduled-stop-375543 kubelet[2479]: I0317 11:03:57.566367 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-certs\" (UniqueName: \"kubernetes.io/host-path/abde8736338038cf2e2191b58eb567b3-etcd-certs\") pod \"etcd-scheduled-stop-375543\" (UID: \"abde8736338038cf2e2191b58eb567b3\") " pod="kube-system/etcd-scheduled-stop-375543"
Mar 17 11:03:57 scheduled-stop-375543 kubelet[2479]: I0317 11:03:57.566386 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b42bb468076de42e247b02c0d1d376fd-k8s-certs\") pod \"kube-apiserver-scheduled-stop-375543\" (UID: \"b42bb468076de42e247b02c0d1d376fd\") " pod="kube-system/kube-apiserver-scheduled-stop-375543"
Mar 17 11:03:57 scheduled-stop-375543 kubelet[2479]: I0317 11:03:57.566416 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/1abc8147fb0e9de3a463fdc45b5eee53-kubeconfig\") pod \"kube-scheduler-scheduled-stop-375543\" (UID: \"1abc8147fb0e9de3a463fdc45b5eee53\") " pod="kube-system/kube-scheduler-scheduled-stop-375543"
Mar 17 11:03:57 scheduled-stop-375543 kubelet[2479]: I0317 11:03:57.566436 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b42bb468076de42e247b02c0d1d376fd-etc-ca-certificates\") pod \"kube-apiserver-scheduled-stop-375543\" (UID: \"b42bb468076de42e247b02c0d1d376fd\") " pod="kube-system/kube-apiserver-scheduled-stop-375543"
Mar 17 11:03:57 scheduled-stop-375543 kubelet[2479]: I0317 11:03:57.566470 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b42bb468076de42e247b02c0d1d376fd-usr-local-share-ca-certificates\") pod \"kube-apiserver-scheduled-stop-375543\" (UID: \"b42bb468076de42e247b02c0d1d376fd\") " pod="kube-system/kube-apiserver-scheduled-stop-375543"
Mar 17 11:03:57 scheduled-stop-375543 kubelet[2479]: I0317 11:03:57.566491 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/da113bf4006ce6881d24cfcd88d12277-flexvolume-dir\") pod \"kube-controller-manager-scheduled-stop-375543\" (UID: \"da113bf4006ce6881d24cfcd88d12277\") " pod="kube-system/kube-controller-manager-scheduled-stop-375543"
Mar 17 11:03:58 scheduled-stop-375543 kubelet[2479]: I0317 11:03:58.144699 2479 apiserver.go:52] "Watching apiserver"
Mar 17 11:03:58 scheduled-stop-375543 kubelet[2479]: I0317 11:03:58.176188 2479 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world"
Mar 17 11:03:58 scheduled-stop-375543 kubelet[2479]: I0317 11:03:58.228712 2479 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-scheduled-stop-375543" podStartSLOduration=1.228686242 podStartE2EDuration="1.228686242s" podCreationTimestamp="2025-03-17 11:03:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 11:03:58.224950263 +0000 UTC m=+1.178656984" watchObservedRunningTime="2025-03-17 11:03:58.228686242 +0000 UTC m=+1.182392954"
Mar 17 11:03:58 scheduled-stop-375543 kubelet[2479]: I0317 11:03:58.242818 2479 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-scheduled-stop-375543" podStartSLOduration=1.242800365 podStartE2EDuration="1.242800365s" podCreationTimestamp="2025-03-17 11:03:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 11:03:58.242283883 +0000 UTC m=+1.195990595" watchObservedRunningTime="2025-03-17 11:03:58.242800365 +0000 UTC m=+1.196507086"
Mar 17 11:03:58 scheduled-stop-375543 kubelet[2479]: I0317 11:03:58.272710 2479 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/etcd-scheduled-stop-375543" podStartSLOduration=1.2726906279999999 podStartE2EDuration="1.272690628s" podCreationTimestamp="2025-03-17 11:03:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 11:03:58.272260743 +0000 UTC m=+1.225967479" watchObservedRunningTime="2025-03-17 11:03:58.272690628 +0000 UTC m=+1.226397332"
Mar 17 11:03:58 scheduled-stop-375543 kubelet[2479]: I0317 11:03:58.272913 2479 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-scheduled-stop-375543" podStartSLOduration=1.272906563 podStartE2EDuration="1.272906563s" podCreationTimestamp="2025-03-17 11:03:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 11:03:58.25494437 +0000 UTC m=+1.208651091" watchObservedRunningTime="2025-03-17 11:03:58.272906563 +0000 UTC m=+1.226613267"
Mar 17 11:03:58 scheduled-stop-375543 kubelet[2479]: I0317 11:03:58.300651 2479 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/etcd-scheduled-stop-375543"
Mar 17 11:03:58 scheduled-stop-375543 kubelet[2479]: E0317 11:03:58.317718 2479 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"etcd-scheduled-stop-375543\" already exists" pod="kube-system/etcd-scheduled-stop-375543"
Mar 17 11:04:01 scheduled-stop-375543 kubelet[2479]: I0317 11:04:01.196364 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/host-path/b0b03cba-7aea-4a43-9ef1-5658eedc8516-tmp\") pod \"storage-provisioner\" (UID: \"b0b03cba-7aea-4a43-9ef1-5658eedc8516\") " pod="kube-system/storage-provisioner"
Mar 17 11:04:01 scheduled-stop-375543 kubelet[2479]: I0317 11:04:01.196420 2479 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjh9b\" (UniqueName: \"kubernetes.io/projected/b0b03cba-7aea-4a43-9ef1-5658eedc8516-kube-api-access-hjh9b\") pod \"storage-provisioner\" (UID: \"b0b03cba-7aea-4a43-9ef1-5658eedc8516\") " pod="kube-system/storage-provisioner"
Mar 17 11:04:01 scheduled-stop-375543 kubelet[2479]: E0317 11:04:01.307614 2479 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found
Mar 17 11:04:01 scheduled-stop-375543 kubelet[2479]: E0317 11:04:01.307652 2479 projected.go:194] Error preparing data for projected volume kube-api-access-hjh9b for pod kube-system/storage-provisioner: configmap "kube-root-ca.crt" not found
Mar 17 11:04:01 scheduled-stop-375543 kubelet[2479]: E0317 11:04:01.307721 2479 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b0b03cba-7aea-4a43-9ef1-5658eedc8516-kube-api-access-hjh9b podName:b0b03cba-7aea-4a43-9ef1-5658eedc8516 nodeName:}" failed. No retries permitted until 2025-03-17 11:04:01.807695915 +0000 UTC m=+4.761402618 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hjh9b" (UniqueName: "kubernetes.io/projected/b0b03cba-7aea-4a43-9ef1-5658eedc8516-kube-api-access-hjh9b") pod "storage-provisioner" (UID: "b0b03cba-7aea-4a43-9ef1-5658eedc8516") : configmap "kube-root-ca.crt" not found
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-arm64 status --format={{.APIServer}} -p scheduled-stop-375543 -n scheduled-stop-375543
helpers_test.go:261: (dbg) Run: kubectl --context scheduled-stop-375543 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: coredns-668d6bf9bc-4fdrt kube-proxy-plhzf storage-provisioner
helpers_test.go:274: ======> post-mortem[TestScheduledStopUnix]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context scheduled-stop-375543 describe pod coredns-668d6bf9bc-4fdrt kube-proxy-plhzf storage-provisioner
helpers_test.go:277: (dbg) Non-zero exit: kubectl --context scheduled-stop-375543 describe pod coredns-668d6bf9bc-4fdrt kube-proxy-plhzf storage-provisioner: exit status 1 (152.749825ms)
** stderr **
Error from server (NotFound): pods "coredns-668d6bf9bc-4fdrt" not found
Error from server (NotFound): pods "kube-proxy-plhzf" not found
Error from server (NotFound): pods "storage-provisioner" not found
** /stderr **
helpers_test.go:279: kubectl --context scheduled-stop-375543 describe pod coredns-668d6bf9bc-4fdrt kube-proxy-plhzf storage-provisioner: exit status 1
helpers_test.go:175: Cleaning up "scheduled-stop-375543" profile ...
helpers_test.go:178: (dbg) Run: out/minikube-linux-arm64 delete -p scheduled-stop-375543
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p scheduled-stop-375543: (2.222313382s)
--- FAIL: TestScheduledStopUnix (37.04s)