=== RUN TestScheduledStopUnix
scheduled_stop_test.go:128: (dbg) Run: out/minikube-linux-arm64 start -p scheduled-stop-735749 --memory=2048 --driver=docker --container-runtime=containerd
E0127 14:20:47.442159 996828 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/addons-084879/client.crt: no such file or directory" logger="UnhandledError"
scheduled_stop_test.go:128: (dbg) Done: out/minikube-linux-arm64 start -p scheduled-stop-735749 --memory=2048 --driver=docker --container-runtime=containerd: (32.284452448s)
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-735749 --schedule 5m
scheduled_stop_test.go:191: (dbg) Run: out/minikube-linux-arm64 status --format={{.TimeToStop}} -p scheduled-stop-735749 -n scheduled-stop-735749
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-735749 --schedule 15s
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:98: process 1145850 running but should have been killed on reschedule of stop
panic.go:629: *** TestScheduledStopUnix FAILED at 2025-01-27 14:21:14.597020527 +0000 UTC m=+2073.453175285
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestScheduledStopUnix]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect scheduled-stop-735749
helpers_test.go:235: (dbg) docker inspect scheduled-stop-735749:
-- stdout --
[
{
"Id": "ad2e0be7832dcc5a59b2018cb9742a52f085f164ce07ea5bf175c02d1db73a12",
"Created": "2025-01-27T14:20:47.293738925Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 1143903,
"ExitCode": 0,
"Error": "",
"StartedAt": "2025-01-27T14:20:47.447693525Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:0434cf58b6dbace281e5de753aa4b2e3fe33dc9a3be53021531403743c3f155a",
"ResolvConfPath": "/var/lib/docker/containers/ad2e0be7832dcc5a59b2018cb9742a52f085f164ce07ea5bf175c02d1db73a12/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/ad2e0be7832dcc5a59b2018cb9742a52f085f164ce07ea5bf175c02d1db73a12/hostname",
"HostsPath": "/var/lib/docker/containers/ad2e0be7832dcc5a59b2018cb9742a52f085f164ce07ea5bf175c02d1db73a12/hosts",
"LogPath": "/var/lib/docker/containers/ad2e0be7832dcc5a59b2018cb9742a52f085f164ce07ea5bf175c02d1db73a12/ad2e0be7832dcc5a59b2018cb9742a52f085f164ce07ea5bf175c02d1db73a12-json.log",
"Name": "/scheduled-stop-735749",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"scheduled-stop-735749:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {}
},
"NetworkMode": "scheduled-stop-735749",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 2147483648,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 4294967296,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/2f765c866da6aa539f308f2197ad46aa795316047e1b1ee8e67e0bf1dd805a5d-init/diff:/var/lib/docker/overlay2/b004c1c82f1902f99b3b285e67dcfa27cc70c62db1cf8fb7b3885e9b2248dbf3/diff",
"MergedDir": "/var/lib/docker/overlay2/2f765c866da6aa539f308f2197ad46aa795316047e1b1ee8e67e0bf1dd805a5d/merged",
"UpperDir": "/var/lib/docker/overlay2/2f765c866da6aa539f308f2197ad46aa795316047e1b1ee8e67e0bf1dd805a5d/diff",
"WorkDir": "/var/lib/docker/overlay2/2f765c866da6aa539f308f2197ad46aa795316047e1b1ee8e67e0bf1dd805a5d/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "scheduled-stop-735749",
"Source": "/var/lib/docker/volumes/scheduled-stop-735749/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "scheduled-stop-735749",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "scheduled-stop-735749",
"name.minikube.sigs.k8s.io": "scheduled-stop-735749",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "05935358a6973877344ffc3be4dac00e111df21e92bb5930e51065afda5ac5d1",
"SandboxKey": "/var/run/docker/netns/05935358a697",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "34096"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "34097"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "34100"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "34098"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "34099"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"scheduled-stop-735749": {
"IPAMConfig": {
"IPv4Address": "192.168.76.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:4c:02",
"DriverOpts": null,
"NetworkID": "54759d86d37eedb571d4cfb6b91cf22f4ebf820e0d41e9238fcf3648c62acc1b",
"EndpointID": "e42c1425898bbdc9ea5a513b2c20b34f449ac4a8956600be5ee2794794d5c329",
"Gateway": "192.168.76.1",
"IPAddress": "192.168.76.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"scheduled-stop-735749",
"ad2e0be7832d"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-arm64 status --format={{.Host}} -p scheduled-stop-735749 -n scheduled-stop-735749
helpers_test.go:244: <<< TestScheduledStopUnix FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestScheduledStopUnix]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-arm64 -p scheduled-stop-735749 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-arm64 -p scheduled-stop-735749 logs -n 25: (1.34816794s)
helpers_test.go:252: TestScheduledStopUnix logs:
-- stdout --
==> Audit <==
|---------|--------------------------------|-----------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|--------------------------------|-----------------------|---------|---------|---------------------|---------------------|
| stop | -p multinode-978001 | multinode-978001 | jenkins | v1.35.0 | 27 Jan 25 14:15 UTC | 27 Jan 25 14:15 UTC |
| start | -p multinode-978001 | multinode-978001 | jenkins | v1.35.0 | 27 Jan 25 14:15 UTC | 27 Jan 25 14:16 UTC |
| | --wait=true -v=8 | | | | | |
| | --alsologtostderr | | | | | |
| node | list -p multinode-978001 | multinode-978001 | jenkins | v1.35.0 | 27 Jan 25 14:16 UTC | |
| node | multinode-978001 node delete | multinode-978001 | jenkins | v1.35.0 | 27 Jan 25 14:16 UTC | 27 Jan 25 14:17 UTC |
| | m03 | | | | | |
| stop | multinode-978001 stop | multinode-978001 | jenkins | v1.35.0 | 27 Jan 25 14:17 UTC | 27 Jan 25 14:17 UTC |
| start | -p multinode-978001 | multinode-978001 | jenkins | v1.35.0 | 27 Jan 25 14:17 UTC | 27 Jan 25 14:18 UTC |
| | --wait=true -v=8 | | | | | |
| | --alsologtostderr | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| node | list -p multinode-978001 | multinode-978001 | jenkins | v1.35.0 | 27 Jan 25 14:18 UTC | |
| start | -p multinode-978001-m02 | multinode-978001-m02 | jenkins | v1.35.0 | 27 Jan 25 14:18 UTC | |
| | --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| start | -p multinode-978001-m03 | multinode-978001-m03 | jenkins | v1.35.0 | 27 Jan 25 14:18 UTC | 27 Jan 25 14:18 UTC |
| | --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| node | add -p multinode-978001 | multinode-978001 | jenkins | v1.35.0 | 27 Jan 25 14:18 UTC | |
| delete | -p multinode-978001-m03 | multinode-978001-m03 | jenkins | v1.35.0 | 27 Jan 25 14:18 UTC | 27 Jan 25 14:18 UTC |
| delete | -p multinode-978001 | multinode-978001 | jenkins | v1.35.0 | 27 Jan 25 14:18 UTC | 27 Jan 25 14:18 UTC |
| start | -p test-preload-185490 | test-preload-185490 | jenkins | v1.35.0 | 27 Jan 25 14:18 UTC | 27 Jan 25 14:20 UTC |
| | --memory=2200 | | | | | |
| | --alsologtostderr | | | | | |
| | --wait=true --preload=false | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| | --kubernetes-version=v1.24.4 | | | | | |
| image | test-preload-185490 image pull | test-preload-185490 | jenkins | v1.35.0 | 27 Jan 25 14:20 UTC | 27 Jan 25 14:20 UTC |
| | gcr.io/k8s-minikube/busybox | | | | | |
| stop | -p test-preload-185490 | test-preload-185490 | jenkins | v1.35.0 | 27 Jan 25 14:20 UTC | 27 Jan 25 14:20 UTC |
| start | -p test-preload-185490 | test-preload-185490 | jenkins | v1.35.0 | 27 Jan 25 14:20 UTC | 27 Jan 25 14:20 UTC |
| | --memory=2200 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| | --wait=true --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| image | test-preload-185490 image list | test-preload-185490 | jenkins | v1.35.0 | 27 Jan 25 14:20 UTC | 27 Jan 25 14:20 UTC |
| delete | -p test-preload-185490 | test-preload-185490 | jenkins | v1.35.0 | 27 Jan 25 14:20 UTC | 27 Jan 25 14:20 UTC |
| start | -p scheduled-stop-735749 | scheduled-stop-735749 | jenkins | v1.35.0 | 27 Jan 25 14:20 UTC | 27 Jan 25 14:21 UTC |
| | --memory=2048 --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| stop | -p scheduled-stop-735749 | scheduled-stop-735749 | jenkins | v1.35.0 | 27 Jan 25 14:21 UTC | |
| | --schedule 5m | | | | | |
| stop | -p scheduled-stop-735749 | scheduled-stop-735749 | jenkins | v1.35.0 | 27 Jan 25 14:21 UTC | |
| | --schedule 5m | | | | | |
| stop | -p scheduled-stop-735749 | scheduled-stop-735749 | jenkins | v1.35.0 | 27 Jan 25 14:21 UTC | |
| | --schedule 5m | | | | | |
| stop | -p scheduled-stop-735749 | scheduled-stop-735749 | jenkins | v1.35.0 | 27 Jan 25 14:21 UTC | |
| | --schedule 15s | | | | | |
| stop | -p scheduled-stop-735749 | scheduled-stop-735749 | jenkins | v1.35.0 | 27 Jan 25 14:21 UTC | |
| | --schedule 15s | | | | | |
| stop | -p scheduled-stop-735749 | scheduled-stop-735749 | jenkins | v1.35.0 | 27 Jan 25 14:21 UTC | |
| | --schedule 15s | | | | | |
|---------|--------------------------------|-----------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2025/01/27 14:20:41
Running on machine: ip-172-31-21-244
Binary: Built with gc go1.23.4 for linux/arm64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0127 14:20:41.842988 1143413 out.go:345] Setting OutFile to fd 1 ...
I0127 14:20:41.843098 1143413 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0127 14:20:41.843102 1143413 out.go:358] Setting ErrFile to fd 2...
I0127 14:20:41.843106 1143413 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0127 14:20:41.843333 1143413 root.go:338] Updating PATH: /home/jenkins/minikube-integration/20325-991453/.minikube/bin
I0127 14:20:41.843697 1143413 out.go:352] Setting JSON to false
I0127 14:20:41.844577 1143413 start.go:129] hostinfo: {"hostname":"ip-172-31-21-244","uptime":18187,"bootTime":1737969455,"procs":179,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1075-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"da8ac1fd-6236-412a-a346-95873c98230d"}
I0127 14:20:41.844638 1143413 start.go:139] virtualization:
I0127 14:20:41.848582 1143413 out.go:177] * [scheduled-stop-735749] minikube v1.35.0 on Ubuntu 20.04 (arm64)
I0127 14:20:41.853156 1143413 out.go:177] - MINIKUBE_LOCATION=20325
I0127 14:20:41.853323 1143413 notify.go:220] Checking for updates...
I0127 14:20:41.859769 1143413 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0127 14:20:41.863139 1143413 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/20325-991453/kubeconfig
I0127 14:20:41.866331 1143413 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/20325-991453/.minikube
I0127 14:20:41.869552 1143413 out.go:177] - MINIKUBE_BIN=out/minikube-linux-arm64
I0127 14:20:41.872720 1143413 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0127 14:20:41.876091 1143413 driver.go:394] Setting default libvirt URI to qemu:///system
I0127 14:20:41.902963 1143413 docker.go:123] docker version: linux-27.5.1:Docker Engine - Community
I0127 14:20:41.903078 1143413 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0127 14:20:41.961716 1143413 info.go:266] docker info: {ID:5FDH:SA5P:5GCT:NLAS:B73P:SGDQ:PBG5:UBVH:UZY3:RXGO:CI7S:WAIH Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:24 OomKillDisable:true NGoroutines:43 SystemTime:2025-01-27 14:20:41.952700267 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1075-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214831104 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-21-244 Labels:[] ExperimentalBuild:false ServerVersion:27.5.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:bcc810d6b9066471b0b6fa75f557a15a1cbf31bb Expected:bcc810d6b9066471b0b6fa75f557a15a1cbf31bb} RuncCommit:{ID:v1.2.4-0-g6c52b3f Expected:v1.2.4-0-g6c52b3f} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErro
rs:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.20.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.32.4]] Warnings:<nil>}}
I0127 14:20:41.961815 1143413 docker.go:318] overlay module found
I0127 14:20:41.965005 1143413 out.go:177] * Using the docker driver based on user configuration
I0127 14:20:41.967883 1143413 start.go:297] selected driver: docker
I0127 14:20:41.967890 1143413 start.go:901] validating driver "docker" against <nil>
I0127 14:20:41.967902 1143413 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0127 14:20:41.968670 1143413 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0127 14:20:42.024519 1143413 info.go:266] docker info: {ID:5FDH:SA5P:5GCT:NLAS:B73P:SGDQ:PBG5:UBVH:UZY3:RXGO:CI7S:WAIH Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:24 OomKillDisable:true NGoroutines:43 SystemTime:2025-01-27 14:20:42.015175348 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1075-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214831104 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-21-244 Labels:[] ExperimentalBuild:false ServerVersion:27.5.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:bcc810d6b9066471b0b6fa75f557a15a1cbf31bb Expected:bcc810d6b9066471b0b6fa75f557a15a1cbf31bb} RuncCommit:{ID:v1.2.4-0-g6c52b3f Expected:v1.2.4-0-g6c52b3f} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErro
rs:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.20.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.32.4]] Warnings:<nil>}}
I0127 14:20:42.024719 1143413 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0127 14:20:42.024942 1143413 start_flags.go:929] Wait components to verify : map[apiserver:true system_pods:true]
I0127 14:20:42.027937 1143413 out.go:177] * Using Docker driver with root privileges
I0127 14:20:42.030948 1143413 cni.go:84] Creating CNI manager for ""
I0127 14:20:42.031006 1143413 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I0127 14:20:42.031015 1143413 start_flags.go:319] Found "CNI" CNI - setting NetworkPlugin=cni
I0127 14:20:42.031100 1143413 start.go:340] cluster config:
{Name:scheduled-stop-735749 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.32.1 ClusterName:scheduled-stop-735749 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:contain
erd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.32.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0127 14:20:42.036360 1143413 out.go:177] * Starting "scheduled-stop-735749" primary control-plane node in "scheduled-stop-735749" cluster
I0127 14:20:42.039287 1143413 cache.go:121] Beginning downloading kic base image for docker with containerd
I0127 14:20:42.042321 1143413 out.go:177] * Pulling base image v0.0.46 ...
I0127 14:20:42.045265 1143413 preload.go:131] Checking if preload exists for k8s version v1.32.1 and runtime containerd
I0127 14:20:42.045321 1143413 preload.go:146] Found local preload: /home/jenkins/minikube-integration/20325-991453/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.32.1-containerd-overlay2-arm64.tar.lz4
I0127 14:20:42.045330 1143413 cache.go:56] Caching tarball of preloaded images
I0127 14:20:42.045347 1143413 image.go:81] Checking for gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 in local docker daemon
I0127 14:20:42.045436 1143413 preload.go:172] Found /home/jenkins/minikube-integration/20325-991453/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.32.1-containerd-overlay2-arm64.tar.lz4 in cache, skipping download
I0127 14:20:42.045446 1143413 cache.go:59] Finished verifying existence of preloaded tar for v1.32.1 on containerd
I0127 14:20:42.046020 1143413 profile.go:143] Saving config to /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/config.json ...
I0127 14:20:42.046060 1143413 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/config.json: {Name:mk083e2b1cdb9055fd758f7de4ee63f29fca6018 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 14:20:42.064777 1143413 image.go:100] Found gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 in local docker daemon, skipping pull
I0127 14:20:42.064789 1143413 cache.go:145] gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 exists in daemon, skipping load
I0127 14:20:42.064808 1143413 cache.go:227] Successfully downloaded all kic artifacts
I0127 14:20:42.064838 1143413 start.go:360] acquireMachinesLock for scheduled-stop-735749: {Name:mkeb98366adf9bbfc405336b49e4b09261f1b968 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0127 14:20:42.064959 1143413 start.go:364] duration metric: took 106.745µs to acquireMachinesLock for "scheduled-stop-735749"
I0127 14:20:42.064992 1143413 start.go:93] Provisioning new machine with config: &{Name:scheduled-stop-735749 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.32.1 ClusterName:scheduled-stop-735749 Namespace:default APIServerHAVIP: APIServerName:minikubeCA
APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.32.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHA
uthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.32.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
I0127 14:20:42.065074 1143413 start.go:125] createHost starting for "" (driver="docker")
I0127 14:20:42.068626 1143413 out.go:235] * Creating docker container (CPUs=2, Memory=2048MB) ...
I0127 14:20:42.068965 1143413 start.go:159] libmachine.API.Create for "scheduled-stop-735749" (driver="docker")
I0127 14:20:42.068997 1143413 client.go:168] LocalClient.Create starting
I0127 14:20:42.069077 1143413 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/20325-991453/.minikube/certs/ca.pem
I0127 14:20:42.069118 1143413 main.go:141] libmachine: Decoding PEM data...
I0127 14:20:42.069131 1143413 main.go:141] libmachine: Parsing certificate...
I0127 14:20:42.069185 1143413 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/20325-991453/.minikube/certs/cert.pem
I0127 14:20:42.069205 1143413 main.go:141] libmachine: Decoding PEM data...
I0127 14:20:42.069214 1143413 main.go:141] libmachine: Parsing certificate...
I0127 14:20:42.069604 1143413 cli_runner.go:164] Run: docker network inspect scheduled-stop-735749 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0127 14:20:42.087983 1143413 cli_runner.go:211] docker network inspect scheduled-stop-735749 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0127 14:20:42.088069 1143413 network_create.go:284] running [docker network inspect scheduled-stop-735749] to gather additional debugging logs...
I0127 14:20:42.088086 1143413 cli_runner.go:164] Run: docker network inspect scheduled-stop-735749
W0127 14:20:42.106048 1143413 cli_runner.go:211] docker network inspect scheduled-stop-735749 returned with exit code 1
I0127 14:20:42.106078 1143413 network_create.go:287] error running [docker network inspect scheduled-stop-735749]: docker network inspect scheduled-stop-735749: exit status 1
stdout:
[]
stderr:
Error response from daemon: network scheduled-stop-735749 not found
I0127 14:20:42.106091 1143413 network_create.go:289] output of [docker network inspect scheduled-stop-735749]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network scheduled-stop-735749 not found
** /stderr **
I0127 14:20:42.106210 1143413 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0127 14:20:42.126276 1143413 network.go:211] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-1270e5594423 IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:02:42:98:b3:9d:3e} reservation:<nil>}
I0127 14:20:42.126810 1143413 network.go:211] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-3681b4bed2ff IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:02:42:19:d3:92:21} reservation:<nil>}
I0127 14:20:42.127142 1143413 network.go:211] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-35812aa39cda IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:02:42:1c:a5:9b:76} reservation:<nil>}
I0127 14:20:42.127632 1143413 network.go:206] using free private subnet 192.168.76.0/24: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x400198f380}
I0127 14:20:42.127651 1143413 network_create.go:124] attempt to create docker network scheduled-stop-735749 192.168.76.0/24 with gateway 192.168.76.1 and MTU of 1500 ...
I0127 14:20:42.127731 1143413 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.76.0/24 --gateway=192.168.76.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=scheduled-stop-735749 scheduled-stop-735749
I0127 14:20:42.214036 1143413 network_create.go:108] docker network scheduled-stop-735749 192.168.76.0/24 created
I0127 14:20:42.214071 1143413 kic.go:121] calculated static IP "192.168.76.2" for the "scheduled-stop-735749" container
I0127 14:20:42.214151 1143413 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0127 14:20:42.231418 1143413 cli_runner.go:164] Run: docker volume create scheduled-stop-735749 --label name.minikube.sigs.k8s.io=scheduled-stop-735749 --label created_by.minikube.sigs.k8s.io=true
I0127 14:20:42.251958 1143413 oci.go:103] Successfully created a docker volume scheduled-stop-735749
I0127 14:20:42.252036 1143413 cli_runner.go:164] Run: docker run --rm --name scheduled-stop-735749-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-735749 --entrypoint /usr/bin/test -v scheduled-stop-735749:/var gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 -d /var/lib
I0127 14:20:42.841329 1143413 oci.go:107] Successfully prepared a docker volume scheduled-stop-735749
I0127 14:20:42.841369 1143413 preload.go:131] Checking if preload exists for k8s version v1.32.1 and runtime containerd
I0127 14:20:42.841387 1143413 kic.go:194] Starting extracting preloaded images to volume ...
I0127 14:20:42.841463 1143413 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/20325-991453/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.32.1-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-735749:/extractDir gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 -I lz4 -xf /preloaded.tar -C /extractDir
I0127 14:20:47.225757 1143413 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/20325-991453/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.32.1-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-735749:/extractDir gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 -I lz4 -xf /preloaded.tar -C /extractDir: (4.384238872s)
I0127 14:20:47.225777 1143413 kic.go:203] duration metric: took 4.384386602s to extract preloaded images to volume ...
W0127 14:20:47.225912 1143413 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0127 14:20:47.226036 1143413 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0127 14:20:47.278724 1143413 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname scheduled-stop-735749 --name scheduled-stop-735749 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-735749 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=scheduled-stop-735749 --network scheduled-stop-735749 --ip 192.168.76.2 --volume scheduled-stop-735749:/var --security-opt apparmor=unconfined --memory=2048mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279
I0127 14:20:47.604611 1143413 cli_runner.go:164] Run: docker container inspect scheduled-stop-735749 --format={{.State.Running}}
I0127 14:20:47.625185 1143413 cli_runner.go:164] Run: docker container inspect scheduled-stop-735749 --format={{.State.Status}}
I0127 14:20:47.651731 1143413 cli_runner.go:164] Run: docker exec scheduled-stop-735749 stat /var/lib/dpkg/alternatives/iptables
I0127 14:20:47.703876 1143413 oci.go:144] the created container "scheduled-stop-735749" has a running status.
I0127 14:20:47.703899 1143413 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/20325-991453/.minikube/machines/scheduled-stop-735749/id_rsa...
I0127 14:20:48.329573 1143413 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/20325-991453/.minikube/machines/scheduled-stop-735749/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0127 14:20:48.368495 1143413 cli_runner.go:164] Run: docker container inspect scheduled-stop-735749 --format={{.State.Status}}
I0127 14:20:48.394440 1143413 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0127 14:20:48.394451 1143413 kic_runner.go:114] Args: [docker exec --privileged scheduled-stop-735749 chown docker:docker /home/docker/.ssh/authorized_keys]
I0127 14:20:48.446065 1143413 cli_runner.go:164] Run: docker container inspect scheduled-stop-735749 --format={{.State.Status}}
I0127 14:20:48.470665 1143413 machine.go:93] provisionDockerMachine start ...
I0127 14:20:48.470757 1143413 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-735749
I0127 14:20:48.498179 1143413 main.go:141] libmachine: Using SSH client type: native
I0127 14:20:48.498580 1143413 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x4132a0] 0x415ae0 <nil> [] 0s} 127.0.0.1 34096 <nil> <nil>}
I0127 14:20:48.498589 1143413 main.go:141] libmachine: About to run SSH command:
hostname
I0127 14:20:48.630430 1143413 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-735749
I0127 14:20:48.630444 1143413 ubuntu.go:169] provisioning hostname "scheduled-stop-735749"
I0127 14:20:48.630509 1143413 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-735749
I0127 14:20:48.651835 1143413 main.go:141] libmachine: Using SSH client type: native
I0127 14:20:48.652109 1143413 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x4132a0] 0x415ae0 <nil> [] 0s} 127.0.0.1 34096 <nil> <nil>}
I0127 14:20:48.652120 1143413 main.go:141] libmachine: About to run SSH command:
sudo hostname scheduled-stop-735749 && echo "scheduled-stop-735749" | sudo tee /etc/hostname
I0127 14:20:48.795044 1143413 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-735749
I0127 14:20:48.795128 1143413 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-735749
I0127 14:20:48.812532 1143413 main.go:141] libmachine: Using SSH client type: native
I0127 14:20:48.812770 1143413 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x4132a0] 0x415ae0 <nil> [] 0s} 127.0.0.1 34096 <nil> <nil>}
I0127 14:20:48.812785 1143413 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sscheduled-stop-735749' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 scheduled-stop-735749/g' /etc/hosts;
else
echo '127.0.1.1 scheduled-stop-735749' | sudo tee -a /etc/hosts;
fi
fi
I0127 14:20:48.934564 1143413 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0127 14:20:48.934582 1143413 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/20325-991453/.minikube CaCertPath:/home/jenkins/minikube-integration/20325-991453/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/20325-991453/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/20325-991453/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/20325-991453/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/20325-991453/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/20325-991453/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/20325-991453/.minikube}
I0127 14:20:48.934606 1143413 ubuntu.go:177] setting up certificates
I0127 14:20:48.934614 1143413 provision.go:84] configureAuth start
I0127 14:20:48.934677 1143413 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-735749
I0127 14:20:48.952306 1143413 provision.go:143] copyHostCerts
I0127 14:20:48.952375 1143413 exec_runner.go:144] found /home/jenkins/minikube-integration/20325-991453/.minikube/cert.pem, removing ...
I0127 14:20:48.952383 1143413 exec_runner.go:203] rm: /home/jenkins/minikube-integration/20325-991453/.minikube/cert.pem
I0127 14:20:48.952466 1143413 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20325-991453/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/20325-991453/.minikube/cert.pem (1123 bytes)
I0127 14:20:48.952558 1143413 exec_runner.go:144] found /home/jenkins/minikube-integration/20325-991453/.minikube/key.pem, removing ...
I0127 14:20:48.952562 1143413 exec_runner.go:203] rm: /home/jenkins/minikube-integration/20325-991453/.minikube/key.pem
I0127 14:20:48.952596 1143413 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20325-991453/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/20325-991453/.minikube/key.pem (1679 bytes)
I0127 14:20:48.952645 1143413 exec_runner.go:144] found /home/jenkins/minikube-integration/20325-991453/.minikube/ca.pem, removing ...
I0127 14:20:48.952649 1143413 exec_runner.go:203] rm: /home/jenkins/minikube-integration/20325-991453/.minikube/ca.pem
I0127 14:20:48.952671 1143413 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20325-991453/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/20325-991453/.minikube/ca.pem (1078 bytes)
I0127 14:20:48.952717 1143413 provision.go:117] generating server cert: /home/jenkins/minikube-integration/20325-991453/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/20325-991453/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/20325-991453/.minikube/certs/ca-key.pem org=jenkins.scheduled-stop-735749 san=[127.0.0.1 192.168.76.2 localhost minikube scheduled-stop-735749]
I0127 14:20:49.313089 1143413 provision.go:177] copyRemoteCerts
I0127 14:20:49.313149 1143413 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0127 14:20:49.313192 1143413 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-735749
I0127 14:20:49.330142 1143413 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34096 SSHKeyPath:/home/jenkins/minikube-integration/20325-991453/.minikube/machines/scheduled-stop-735749/id_rsa Username:docker}
I0127 14:20:49.423507 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
I0127 14:20:49.448332 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0127 14:20:49.472385 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/machines/server.pem --> /etc/docker/server.pem (1229 bytes)
I0127 14:20:49.500512 1143413 provision.go:87] duration metric: took 565.884425ms to configureAuth
I0127 14:20:49.500528 1143413 ubuntu.go:193] setting minikube options for container-runtime
I0127 14:20:49.500722 1143413 config.go:182] Loaded profile config "scheduled-stop-735749": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.32.1
I0127 14:20:49.500728 1143413 machine.go:96] duration metric: took 1.030053508s to provisionDockerMachine
I0127 14:20:49.500734 1143413 client.go:171] duration metric: took 7.431732103s to LocalClient.Create
I0127 14:20:49.500757 1143413 start.go:167] duration metric: took 7.431796314s to libmachine.API.Create "scheduled-stop-735749"
I0127 14:20:49.500764 1143413 start.go:293] postStartSetup for "scheduled-stop-735749" (driver="docker")
I0127 14:20:49.500772 1143413 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0127 14:20:49.500826 1143413 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0127 14:20:49.500865 1143413 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-735749
I0127 14:20:49.519612 1143413 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34096 SSHKeyPath:/home/jenkins/minikube-integration/20325-991453/.minikube/machines/scheduled-stop-735749/id_rsa Username:docker}
I0127 14:20:49.615817 1143413 ssh_runner.go:195] Run: cat /etc/os-release
I0127 14:20:49.619109 1143413 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0127 14:20:49.619139 1143413 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0127 14:20:49.619149 1143413 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0127 14:20:49.619159 1143413 info.go:137] Remote host: Ubuntu 22.04.5 LTS
I0127 14:20:49.619168 1143413 filesync.go:126] Scanning /home/jenkins/minikube-integration/20325-991453/.minikube/addons for local assets ...
I0127 14:20:49.619237 1143413 filesync.go:126] Scanning /home/jenkins/minikube-integration/20325-991453/.minikube/files for local assets ...
I0127 14:20:49.619321 1143413 filesync.go:149] local asset: /home/jenkins/minikube-integration/20325-991453/.minikube/files/etc/ssl/certs/9968282.pem -> 9968282.pem in /etc/ssl/certs
I0127 14:20:49.619453 1143413 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I0127 14:20:49.627945 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/files/etc/ssl/certs/9968282.pem --> /etc/ssl/certs/9968282.pem (1708 bytes)
I0127 14:20:49.652187 1143413 start.go:296] duration metric: took 151.408795ms for postStartSetup
I0127 14:20:49.652557 1143413 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-735749
I0127 14:20:49.669061 1143413 profile.go:143] Saving config to /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/config.json ...
I0127 14:20:49.669342 1143413 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0127 14:20:49.669384 1143413 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-735749
I0127 14:20:49.685749 1143413 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34096 SSHKeyPath:/home/jenkins/minikube-integration/20325-991453/.minikube/machines/scheduled-stop-735749/id_rsa Username:docker}
I0127 14:20:49.770901 1143413 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0127 14:20:49.775145 1143413 start.go:128] duration metric: took 7.710056284s to createHost
I0127 14:20:49.775159 1143413 start.go:83] releasing machines lock for "scheduled-stop-735749", held for 7.710192141s
I0127 14:20:49.775239 1143413 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-735749
I0127 14:20:49.794000 1143413 ssh_runner.go:195] Run: cat /version.json
I0127 14:20:49.794043 1143413 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-735749
I0127 14:20:49.794278 1143413 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0127 14:20:49.794349 1143413 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-735749
I0127 14:20:49.816990 1143413 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34096 SSHKeyPath:/home/jenkins/minikube-integration/20325-991453/.minikube/machines/scheduled-stop-735749/id_rsa Username:docker}
I0127 14:20:49.824254 1143413 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34096 SSHKeyPath:/home/jenkins/minikube-integration/20325-991453/.minikube/machines/scheduled-stop-735749/id_rsa Username:docker}
I0127 14:20:49.902686 1143413 ssh_runner.go:195] Run: systemctl --version
I0127 14:20:50.050041 1143413 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0127 14:20:50.055745 1143413 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0127 14:20:50.082395 1143413 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0127 14:20:50.082473 1143413 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0127 14:20:50.116070 1143413 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0127 14:20:50.116083 1143413 start.go:495] detecting cgroup driver to use...
I0127 14:20:50.116115 1143413 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0127 14:20:50.116170 1143413 ssh_runner.go:195] Run: sudo systemctl stop -f crio
I0127 14:20:50.129422 1143413 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0127 14:20:50.141826 1143413 docker.go:217] disabling cri-docker service (if available) ...
I0127 14:20:50.141898 1143413 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
I0127 14:20:50.156890 1143413 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
I0127 14:20:50.171921 1143413 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
I0127 14:20:50.259530 1143413 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
I0127 14:20:50.354531 1143413 docker.go:233] disabling docker service ...
I0127 14:20:50.354592 1143413 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
I0127 14:20:50.375875 1143413 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
I0127 14:20:50.388221 1143413 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
I0127 14:20:50.482433 1143413 ssh_runner.go:195] Run: sudo systemctl mask docker.service
I0127 14:20:50.572184 1143413 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
I0127 14:20:50.583716 1143413 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0127 14:20:50.600987 1143413 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0127 14:20:50.611162 1143413 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0127 14:20:50.621098 1143413 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0127 14:20:50.621160 1143413 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0127 14:20:50.631193 1143413 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0127 14:20:50.641091 1143413 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0127 14:20:50.651879 1143413 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0127 14:20:50.661969 1143413 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0127 14:20:50.671414 1143413 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0127 14:20:50.681213 1143413 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0127 14:20:50.691197 1143413 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0127 14:20:50.701359 1143413 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0127 14:20:50.710036 1143413 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0127 14:20:50.718394 1143413 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0127 14:20:50.796511 1143413 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0127 14:20:50.921479 1143413 start.go:542] Will wait 60s for socket path /run/containerd/containerd.sock
I0127 14:20:50.921539 1143413 ssh_runner.go:195] Run: stat /run/containerd/containerd.sock
I0127 14:20:50.925542 1143413 start.go:563] Will wait 60s for crictl version
I0127 14:20:50.925599 1143413 ssh_runner.go:195] Run: which crictl
I0127 14:20:50.929147 1143413 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0127 14:20:50.963914 1143413 start.go:579] Version: 0.1.0
RuntimeName: containerd
RuntimeVersion: 1.7.24
RuntimeApiVersion: v1
I0127 14:20:50.963973 1143413 ssh_runner.go:195] Run: containerd --version
I0127 14:20:50.987275 1143413 ssh_runner.go:195] Run: containerd --version
I0127 14:20:51.020839 1143413 out.go:177] * Preparing Kubernetes v1.32.1 on containerd 1.7.24 ...
I0127 14:20:51.023788 1143413 cli_runner.go:164] Run: docker network inspect scheduled-stop-735749 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0127 14:20:51.040139 1143413 ssh_runner.go:195] Run: grep 192.168.76.1 host.minikube.internal$ /etc/hosts
I0127 14:20:51.043914 1143413 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.76.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0127 14:20:51.055293 1143413 kubeadm.go:883] updating cluster {Name:scheduled-stop-735749 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.32.1 ClusterName:scheduled-stop-735749 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] A
PIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.32.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock
: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0127 14:20:51.055407 1143413 preload.go:131] Checking if preload exists for k8s version v1.32.1 and runtime containerd
I0127 14:20:51.055466 1143413 ssh_runner.go:195] Run: sudo crictl images --output json
I0127 14:20:51.093029 1143413 containerd.go:627] all images are preloaded for containerd runtime.
I0127 14:20:51.093042 1143413 containerd.go:534] Images already preloaded, skipping extraction
I0127 14:20:51.093116 1143413 ssh_runner.go:195] Run: sudo crictl images --output json
I0127 14:20:51.129983 1143413 containerd.go:627] all images are preloaded for containerd runtime.
I0127 14:20:51.129996 1143413 cache_images.go:84] Images are preloaded, skipping loading
I0127 14:20:51.130002 1143413 kubeadm.go:934] updating node { 192.168.76.2 8443 v1.32.1 containerd true true} ...
I0127 14:20:51.130097 1143413 kubeadm.go:946] kubelet [Unit]
Wants=containerd.service
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.32.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=scheduled-stop-735749 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.76.2
[Install]
config:
{KubernetesVersion:v1.32.1 ClusterName:scheduled-stop-735749 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0127 14:20:51.130163 1143413 ssh_runner.go:195] Run: sudo crictl info
I0127 14:20:51.167853 1143413 cni.go:84] Creating CNI manager for ""
I0127 14:20:51.167864 1143413 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I0127 14:20:51.167873 1143413 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0127 14:20:51.167896 1143413 kubeadm.go:189] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.76.2 APIServerPort:8443 KubernetesVersion:v1.32.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:scheduled-stop-735749 NodeName:scheduled-stop-735749 DNSDomain:cluster.local CRISocket:/run/containerd/containerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.76.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.76.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt St
aticPodPath:/etc/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///run/containerd/containerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0127 14:20:51.168020 1143413 kubeadm.go:195] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta4
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.76.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///run/containerd/containerd.sock
name: "scheduled-stop-735749"
kubeletExtraArgs:
- name: "node-ip"
value: "192.168.76.2"
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta4
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.76.2"]
extraArgs:
- name: "enable-admission-plugins"
value: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
- name: "allocate-node-cidrs"
value: "true"
- name: "leader-elect"
value: "false"
scheduler:
extraArgs:
- name: "leader-elect"
value: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
- name: "proxy-refresh-interval"
value: "70000"
kubernetesVersion: v1.32.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///run/containerd/containerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0127 14:20:51.168099 1143413 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.32.1
I0127 14:20:51.177629 1143413 binaries.go:44] Found k8s binaries, skipping transfer
I0127 14:20:51.177711 1143413 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0127 14:20:51.186915 1143413 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (325 bytes)
I0127 14:20:51.206061 1143413 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0127 14:20:51.224821 1143413 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2311 bytes)
I0127 14:20:51.244070 1143413 ssh_runner.go:195] Run: grep 192.168.76.2 control-plane.minikube.internal$ /etc/hosts
I0127 14:20:51.247703 1143413 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.76.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0127 14:20:51.258742 1143413 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0127 14:20:51.349767 1143413 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0127 14:20:51.363995 1143413 certs.go:68] Setting up /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749 for IP: 192.168.76.2
I0127 14:20:51.364006 1143413 certs.go:194] generating shared ca certs ...
I0127 14:20:51.364021 1143413 certs.go:226] acquiring lock for ca certs: {Name:mkb01e5d6af6caa976157830f22ddbd96a9b670f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 14:20:51.364161 1143413 certs.go:235] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/20325-991453/.minikube/ca.key
I0127 14:20:51.364199 1143413 certs.go:235] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/20325-991453/.minikube/proxy-client-ca.key
I0127 14:20:51.364205 1143413 certs.go:256] generating profile certs ...
I0127 14:20:51.364268 1143413 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/client.key
I0127 14:20:51.364285 1143413 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/client.crt with IP's: []
I0127 14:20:51.890430 1143413 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/client.crt ...
I0127 14:20:51.890444 1143413 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/client.crt: {Name:mkaeeaaf2f440661f7fedcc57b0da3d06e6bf4d0 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 14:20:51.890655 1143413 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/client.key ...
I0127 14:20:51.890666 1143413 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/client.key: {Name:mkad8bab00153b37faffc7f5b0d64e361b00722d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 14:20:51.890769 1143413 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/apiserver.key.2bad953c
I0127 14:20:51.890782 1143413 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/apiserver.crt.2bad953c with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.76.2]
I0127 14:20:52.675896 1143413 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/apiserver.crt.2bad953c ...
I0127 14:20:52.675912 1143413 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/apiserver.crt.2bad953c: {Name:mk5d3747630ce2ef476c165a41e0426e50b4c12c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 14:20:52.676118 1143413 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/apiserver.key.2bad953c ...
I0127 14:20:52.676126 1143413 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/apiserver.key.2bad953c: {Name:mkbbbed3f39c777a5a19e30ab326ea1eaf56843c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 14:20:52.676210 1143413 certs.go:381] copying /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/apiserver.crt.2bad953c -> /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/apiserver.crt
I0127 14:20:52.676283 1143413 certs.go:385] copying /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/apiserver.key.2bad953c -> /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/apiserver.key
I0127 14:20:52.676335 1143413 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/proxy-client.key
I0127 14:20:52.676347 1143413 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/proxy-client.crt with IP's: []
I0127 14:20:53.114642 1143413 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/proxy-client.crt ...
I0127 14:20:53.114657 1143413 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/proxy-client.crt: {Name:mk2db7f34e4621f3b8ce7e6186764b66b7435ed8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 14:20:53.114853 1143413 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/proxy-client.key ...
I0127 14:20:53.114861 1143413 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/proxy-client.key: {Name:mkd3089b785f8a328d2212209d3d0a350a5972c2 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 14:20:53.115048 1143413 certs.go:484] found cert: /home/jenkins/minikube-integration/20325-991453/.minikube/certs/996828.pem (1338 bytes)
W0127 14:20:53.115082 1143413 certs.go:480] ignoring /home/jenkins/minikube-integration/20325-991453/.minikube/certs/996828_empty.pem, impossibly tiny 0 bytes
I0127 14:20:53.115090 1143413 certs.go:484] found cert: /home/jenkins/minikube-integration/20325-991453/.minikube/certs/ca-key.pem (1675 bytes)
I0127 14:20:53.115112 1143413 certs.go:484] found cert: /home/jenkins/minikube-integration/20325-991453/.minikube/certs/ca.pem (1078 bytes)
I0127 14:20:53.115136 1143413 certs.go:484] found cert: /home/jenkins/minikube-integration/20325-991453/.minikube/certs/cert.pem (1123 bytes)
I0127 14:20:53.115165 1143413 certs.go:484] found cert: /home/jenkins/minikube-integration/20325-991453/.minikube/certs/key.pem (1679 bytes)
I0127 14:20:53.115213 1143413 certs.go:484] found cert: /home/jenkins/minikube-integration/20325-991453/.minikube/files/etc/ssl/certs/9968282.pem (1708 bytes)
I0127 14:20:53.115908 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0127 14:20:53.141411 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I0127 14:20:53.167155 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0127 14:20:53.192346 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I0127 14:20:53.217818 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1432 bytes)
I0127 14:20:53.242832 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
I0127 14:20:53.268643 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0127 14:20:53.293640 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/profiles/scheduled-stop-735749/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0127 14:20:53.318792 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0127 14:20:53.344550 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/certs/996828.pem --> /usr/share/ca-certificates/996828.pem (1338 bytes)
I0127 14:20:53.370165 1143413 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20325-991453/.minikube/files/etc/ssl/certs/9968282.pem --> /usr/share/ca-certificates/9968282.pem (1708 bytes)
I0127 14:20:53.395959 1143413 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0127 14:20:53.414608 1143413 ssh_runner.go:195] Run: openssl version
I0127 14:20:53.420396 1143413 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0127 14:20:53.430264 1143413 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0127 14:20:53.434092 1143413 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Jan 27 13:47 /usr/share/ca-certificates/minikubeCA.pem
I0127 14:20:53.434153 1143413 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0127 14:20:53.441564 1143413 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0127 14:20:53.451175 1143413 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/996828.pem && ln -fs /usr/share/ca-certificates/996828.pem /etc/ssl/certs/996828.pem"
I0127 14:20:53.460960 1143413 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/996828.pem
I0127 14:20:53.464991 1143413 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 Jan 27 13:55 /usr/share/ca-certificates/996828.pem
I0127 14:20:53.465049 1143413 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/996828.pem
I0127 14:20:53.472787 1143413 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/996828.pem /etc/ssl/certs/51391683.0"
I0127 14:20:53.482521 1143413 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/9968282.pem && ln -fs /usr/share/ca-certificates/9968282.pem /etc/ssl/certs/9968282.pem"
I0127 14:20:53.491991 1143413 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/9968282.pem
I0127 14:20:53.495810 1143413 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 Jan 27 13:55 /usr/share/ca-certificates/9968282.pem
I0127 14:20:53.495867 1143413 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/9968282.pem
I0127 14:20:53.502904 1143413 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/9968282.pem /etc/ssl/certs/3ec20f2e.0"
I0127 14:20:53.512703 1143413 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0127 14:20:53.516109 1143413 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0127 14:20:53.516155 1143413 kubeadm.go:392] StartCluster: {Name:scheduled-stop-735749 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.32.1 ClusterName:scheduled-stop-735749 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIS
erverIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.32.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: S
SHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0127 14:20:53.516229 1143413 cri.go:54] listing CRI containers in root /run/containerd/runc/k8s.io: {State:paused Name: Namespaces:[kube-system]}
I0127 14:20:53.516287 1143413 ssh_runner.go:195] Run: sudo -s eval "crictl ps -a --quiet --label io.kubernetes.pod.namespace=kube-system"
I0127 14:20:53.553321 1143413 cri.go:89] found id: ""
I0127 14:20:53.553384 1143413 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0127 14:20:53.562464 1143413 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0127 14:20:53.571494 1143413 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0127 14:20:53.571551 1143413 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0127 14:20:53.580605 1143413 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0127 14:20:53.580615 1143413 kubeadm.go:157] found existing configuration files:
I0127 14:20:53.580667 1143413 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0127 14:20:53.589617 1143413 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0127 14:20:53.589672 1143413 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0127 14:20:53.598428 1143413 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0127 14:20:53.607656 1143413 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0127 14:20:53.607715 1143413 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0127 14:20:53.616591 1143413 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0127 14:20:53.625954 1143413 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0127 14:20:53.626011 1143413 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0127 14:20:53.635649 1143413 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0127 14:20:53.644624 1143413 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0127 14:20:53.644682 1143413 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0127 14:20:53.653454 1143413 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.32.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0127 14:20:53.695422 1143413 kubeadm.go:310] [init] Using Kubernetes version: v1.32.1
I0127 14:20:53.695472 1143413 kubeadm.go:310] [preflight] Running pre-flight checks
I0127 14:20:53.722721 1143413 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0127 14:20:53.722786 1143413 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1075-aws[0m
I0127 14:20:53.722834 1143413 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0127 14:20:53.722879 1143413 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0127 14:20:53.722930 1143413 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0127 14:20:53.722976 1143413 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0127 14:20:53.723024 1143413 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0127 14:20:53.723071 1143413 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0127 14:20:53.723118 1143413 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0127 14:20:53.723162 1143413 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0127 14:20:53.723209 1143413 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0127 14:20:53.723254 1143413 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0127 14:20:53.791870 1143413 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0127 14:20:53.791973 1143413 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0127 14:20:53.792062 1143413 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0127 14:20:53.797362 1143413 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0127 14:20:53.803553 1143413 out.go:235] - Generating certificates and keys ...
I0127 14:20:53.803736 1143413 kubeadm.go:310] [certs] Using existing ca certificate authority
I0127 14:20:53.803811 1143413 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0127 14:20:54.283461 1143413 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0127 14:20:54.990547 1143413 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0127 14:20:55.265596 1143413 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0127 14:20:56.494817 1143413 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0127 14:20:56.820400 1143413 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0127 14:20:56.820731 1143413 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [localhost scheduled-stop-735749] and IPs [192.168.76.2 127.0.0.1 ::1]
I0127 14:20:57.493224 1143413 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0127 14:20:57.493542 1143413 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [localhost scheduled-stop-735749] and IPs [192.168.76.2 127.0.0.1 ::1]
I0127 14:20:58.374730 1143413 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0127 14:20:59.763695 1143413 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0127 14:21:00.210193 1143413 kubeadm.go:310] [certs] Generating "sa" key and public key
I0127 14:21:00.210259 1143413 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0127 14:21:00.739199 1143413 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0127 14:21:01.505227 1143413 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0127 14:21:02.484359 1143413 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0127 14:21:02.629834 1143413 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0127 14:21:03.736711 1143413 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0127 14:21:03.737546 1143413 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0127 14:21:03.740452 1143413 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0127 14:21:03.744089 1143413 out.go:235] - Booting up control plane ...
I0127 14:21:03.744200 1143413 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0127 14:21:03.744292 1143413 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0127 14:21:03.744374 1143413 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0127 14:21:03.754777 1143413 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0127 14:21:03.761446 1143413 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0127 14:21:03.761653 1143413 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0127 14:21:03.856416 1143413 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0127 14:21:03.856527 1143413 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0127 14:21:04.856711 1143413 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 1.001609184s
I0127 14:21:04.856790 1143413 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0127 14:21:10.858666 1143413 kubeadm.go:310] [api-check] The API server is healthy after 6.001942913s
I0127 14:21:10.879023 1143413 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0127 14:21:10.896024 1143413 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0127 14:21:10.932111 1143413 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0127 14:21:10.932306 1143413 kubeadm.go:310] [mark-control-plane] Marking the node scheduled-stop-735749 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0127 14:21:10.949734 1143413 kubeadm.go:310] [bootstrap-token] Using token: 2z1caa.xxn9wfnju1welqhz
I0127 14:21:10.952659 1143413 out.go:235] - Configuring RBAC rules ...
I0127 14:21:10.952782 1143413 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0127 14:21:10.962870 1143413 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0127 14:21:10.974675 1143413 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0127 14:21:10.978780 1143413 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0127 14:21:10.983138 1143413 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0127 14:21:10.987950 1143413 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0127 14:21:11.265616 1143413 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0127 14:21:11.695938 1143413 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0127 14:21:12.267598 1143413 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0127 14:21:12.268821 1143413 kubeadm.go:310]
I0127 14:21:12.268885 1143413 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0127 14:21:12.268889 1143413 kubeadm.go:310]
I0127 14:21:12.268981 1143413 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0127 14:21:12.268992 1143413 kubeadm.go:310]
I0127 14:21:12.269017 1143413 kubeadm.go:310] mkdir -p $HOME/.kube
I0127 14:21:12.269092 1143413 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0127 14:21:12.269142 1143413 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0127 14:21:12.269145 1143413 kubeadm.go:310]
I0127 14:21:12.269197 1143413 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0127 14:21:12.269201 1143413 kubeadm.go:310]
I0127 14:21:12.269247 1143413 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0127 14:21:12.269250 1143413 kubeadm.go:310]
I0127 14:21:12.269301 1143413 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0127 14:21:12.269374 1143413 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0127 14:21:12.269441 1143413 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0127 14:21:12.269445 1143413 kubeadm.go:310]
I0127 14:21:12.269528 1143413 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0127 14:21:12.269603 1143413 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0127 14:21:12.269606 1143413 kubeadm.go:310]
I0127 14:21:12.269689 1143413 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 2z1caa.xxn9wfnju1welqhz \
I0127 14:21:12.269791 1143413 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:f34590ae1ee4943696d2681a1698aa06942b53edb55aebb706b7fa1cb74640f2 \
I0127 14:21:12.269810 1143413 kubeadm.go:310] --control-plane
I0127 14:21:12.269814 1143413 kubeadm.go:310]
I0127 14:21:12.269897 1143413 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0127 14:21:12.269900 1143413 kubeadm.go:310]
I0127 14:21:12.269980 1143413 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 2z1caa.xxn9wfnju1welqhz \
I0127 14:21:12.270080 1143413 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:f34590ae1ee4943696d2681a1698aa06942b53edb55aebb706b7fa1cb74640f2
I0127 14:21:12.273391 1143413 kubeadm.go:310] [WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2
I0127 14:21:12.273608 1143413 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1075-aws\n", err: exit status 1
I0127 14:21:12.273711 1143413 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0127 14:21:12.273729 1143413 cni.go:84] Creating CNI manager for ""
I0127 14:21:12.273736 1143413 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I0127 14:21:12.276825 1143413 out.go:177] * Configuring CNI (Container Networking Interface) ...
I0127 14:21:12.280026 1143413 ssh_runner.go:195] Run: stat /opt/cni/bin/portmap
I0127 14:21:12.283934 1143413 cni.go:182] applying CNI manifest using /var/lib/minikube/binaries/v1.32.1/kubectl ...
I0127 14:21:12.283946 1143413 ssh_runner.go:362] scp memory --> /var/tmp/minikube/cni.yaml (2601 bytes)
I0127 14:21:12.302663 1143413 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.32.1/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml
I0127 14:21:12.591882 1143413 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0127 14:21:12.592030 1143413 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.32.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0127 14:21:12.592115 1143413 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.32.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes scheduled-stop-735749 minikube.k8s.io/updated_at=2025_01_27T14_21_12_0700 minikube.k8s.io/version=v1.35.0 minikube.k8s.io/commit=6a5089c94d5c3e26f81a121b7614c4f7f440f9c0 minikube.k8s.io/name=scheduled-stop-735749 minikube.k8s.io/primary=true
I0127 14:21:12.830459 1143413 ops.go:34] apiserver oom_adj: -16
I0127 14:21:12.830477 1143413 kubeadm.go:1113] duration metric: took 238.50513ms to wait for elevateKubeSystemPrivileges
I0127 14:21:12.830500 1143413 kubeadm.go:394] duration metric: took 19.314341384s to StartCluster
I0127 14:21:12.830516 1143413 settings.go:142] acquiring lock: {Name:mk3848c56ae240dfd12dbfb5c16262cb59f58032 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 14:21:12.830574 1143413 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/20325-991453/kubeconfig
I0127 14:21:12.831748 1143413 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20325-991453/kubeconfig: {Name:mkea3415af2050ac5f7446b7fe73df41bf02ce17 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 14:21:12.832309 1143413 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.32.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
I0127 14:21:12.833212 1143413 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.32.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0127 14:21:12.833518 1143413 config.go:182] Loaded profile config "scheduled-stop-735749": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.32.1
I0127 14:21:12.833567 1143413 addons.go:511] enable addons start: toEnable=map[ambassador:false amd-gpu-device-plugin:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:false volcano:false volumesnapshots:false yakd:false]
I0127 14:21:12.833706 1143413 addons.go:69] Setting storage-provisioner=true in profile "scheduled-stop-735749"
I0127 14:21:12.833728 1143413 addons.go:238] Setting addon storage-provisioner=true in "scheduled-stop-735749"
I0127 14:21:12.833752 1143413 host.go:66] Checking if "scheduled-stop-735749" exists ...
I0127 14:21:12.834517 1143413 addons.go:69] Setting default-storageclass=true in profile "scheduled-stop-735749"
I0127 14:21:12.834533 1143413 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "scheduled-stop-735749"
I0127 14:21:12.834928 1143413 cli_runner.go:164] Run: docker container inspect scheduled-stop-735749 --format={{.State.Status}}
I0127 14:21:12.835729 1143413 cli_runner.go:164] Run: docker container inspect scheduled-stop-735749 --format={{.State.Status}}
I0127 14:21:12.844110 1143413 out.go:177] * Verifying Kubernetes components...
I0127 14:21:12.847117 1143413 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0127 14:21:12.871904 1143413 addons.go:238] Setting addon default-storageclass=true in "scheduled-stop-735749"
I0127 14:21:12.871932 1143413 host.go:66] Checking if "scheduled-stop-735749" exists ...
I0127 14:21:12.872824 1143413 cli_runner.go:164] Run: docker container inspect scheduled-stop-735749 --format={{.State.Status}}
I0127 14:21:12.892404 1143413 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0127 14:21:12.893272 1143413 addons.go:435] installing /etc/kubernetes/addons/storageclass.yaml
I0127 14:21:12.893283 1143413 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0127 14:21:12.893360 1143413 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-735749
I0127 14:21:12.895622 1143413 addons.go:435] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0127 14:21:12.895633 1143413 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0127 14:21:12.895698 1143413 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-735749
I0127 14:21:12.931629 1143413 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34096 SSHKeyPath:/home/jenkins/minikube-integration/20325-991453/.minikube/machines/scheduled-stop-735749/id_rsa Username:docker}
I0127 14:21:12.953410 1143413 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:34096 SSHKeyPath:/home/jenkins/minikube-integration/20325-991453/.minikube/machines/scheduled-stop-735749/id_rsa Username:docker}
I0127 14:21:13.082140 1143413 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.32.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.76.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.32.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0127 14:21:13.082269 1143413 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0127 14:21:13.087564 1143413 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.32.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0127 14:21:13.118797 1143413 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.32.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0127 14:21:13.474948 1143413 api_server.go:52] waiting for apiserver process to appear ...
I0127 14:21:13.475000 1143413 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0127 14:21:13.475068 1143413 start.go:971] {"host.minikube.internal": 192.168.76.1} host record injected into CoreDNS's ConfigMap
I0127 14:21:13.698502 1143413 api_server.go:72] duration metric: took 866.166663ms to wait for apiserver process to appear ...
I0127 14:21:13.698514 1143413 api_server.go:88] waiting for apiserver healthz status ...
I0127 14:21:13.698530 1143413 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
I0127 14:21:13.703311 1143413 out.go:177] * Enabled addons: default-storageclass, storage-provisioner
I0127 14:21:13.707692 1143413 addons.go:514] duration metric: took 874.120107ms for enable addons: enabled=[default-storageclass storage-provisioner]
I0127 14:21:13.708552 1143413 api_server.go:279] https://192.168.76.2:8443/healthz returned 200:
ok
I0127 14:21:13.709545 1143413 api_server.go:141] control plane version: v1.32.1
I0127 14:21:13.709558 1143413 api_server.go:131] duration metric: took 11.039879ms to wait for apiserver health ...
I0127 14:21:13.709565 1143413 system_pods.go:43] waiting for kube-system pods to appear ...
I0127 14:21:13.717513 1143413 system_pods.go:59] 5 kube-system pods found
I0127 14:21:13.717534 1143413 system_pods.go:61] "etcd-scheduled-stop-735749" [4abacb68-80d2-45d7-97fd-3f25a84b012f] Running / Ready:ContainersNotReady (containers with unready status: [etcd]) / ContainersReady:ContainersNotReady (containers with unready status: [etcd])
I0127 14:21:13.717541 1143413 system_pods.go:61] "kube-apiserver-scheduled-stop-735749" [174e95e1-fd38-434c-a61c-9617339b2f34] Running / Ready:ContainersNotReady (containers with unready status: [kube-apiserver]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-apiserver])
I0127 14:21:13.717549 1143413 system_pods.go:61] "kube-controller-manager-scheduled-stop-735749" [bbcd382d-8c87-437e-853d-a5e85ef6aa30] Running / Ready:ContainersNotReady (containers with unready status: [kube-controller-manager]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-controller-manager])
I0127 14:21:13.717555 1143413 system_pods.go:61] "kube-scheduler-scheduled-stop-735749" [804b9a6c-73e5-44c3-9dac-cb76e16e4e6c] Running / Ready:ContainersNotReady (containers with unready status: [kube-scheduler]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-scheduler])
I0127 14:21:13.717562 1143413 system_pods.go:61] "storage-provisioner" [441b63a4-fb56-4eaa-a223-1dcf19c2bcaf] Pending: PodScheduled:Unschedulable (0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.)
I0127 14:21:13.717567 1143413 system_pods.go:74] duration metric: took 7.997373ms to wait for pod list to return data ...
I0127 14:21:13.717580 1143413 kubeadm.go:582] duration metric: took 885.248895ms to wait for: map[apiserver:true system_pods:true]
I0127 14:21:13.717591 1143413 node_conditions.go:102] verifying NodePressure condition ...
I0127 14:21:13.721026 1143413 node_conditions.go:122] node storage ephemeral capacity is 203034800Ki
I0127 14:21:13.721044 1143413 node_conditions.go:123] node cpu capacity is 2
I0127 14:21:13.721054 1143413 node_conditions.go:105] duration metric: took 3.459529ms to run NodePressure ...
I0127 14:21:13.721067 1143413 start.go:241] waiting for startup goroutines ...
I0127 14:21:13.979007 1143413 kapi.go:214] "coredns" deployment in "kube-system" namespace and "scheduled-stop-735749" context rescaled to 1 replicas
I0127 14:21:13.979035 1143413 start.go:246] waiting for cluster config update ...
I0127 14:21:13.979045 1143413 start.go:255] writing updated cluster config ...
I0127 14:21:13.979333 1143413 ssh_runner.go:195] Run: rm -f paused
I0127 14:21:14.048685 1143413 start.go:600] kubectl: 1.32.1, cluster: 1.32.1 (minor skew: 0)
I0127 14:21:14.052050 1143413 out.go:177] * Done! kubectl is now configured to use "scheduled-stop-735749" cluster and "default" namespace by default
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
5ff22105c9846 7fc9d4aa817aa 10 seconds ago Running etcd 0 29f54e456d5b1 etcd-scheduled-stop-735749
a4b8c16a2fbc9 ddb38cac617cb 10 seconds ago Running kube-scheduler 0 9744d282f22e1 kube-scheduler-scheduled-stop-735749
746e960627509 265c2dedf28ab 10 seconds ago Running kube-apiserver 0 4ef955ccf5455 kube-apiserver-scheduled-stop-735749
dca8ceecd1b46 2933761aa7ada 10 seconds ago Running kube-controller-manager 0 a0e5e2e8995e1 kube-controller-manager-scheduled-stop-735749
==> containerd <==
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.066953964Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.089922454Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.090274945Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.090473472Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.094623141Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.161773416Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-scheduled-stop-735749,Uid:dd0e3ecc14dc0c85c5b5b7c3092493ef,Namespace:kube-system,Attempt:0,} returns sandbox id \"4ef955ccf545587854d8e698935d1f4d8b62a65525164aa039d663a92bfdf19d\""
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.163853592Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-scheduled-stop-735749,Uid:e0dacb0c468f59571819daff087d0add,Namespace:kube-system,Attempt:0,} returns sandbox id \"a0e5e2e8995e15ef0551672ed7f182a58f3501c3bbf130d2cefaf87365852884\""
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.184155895Z" level=info msg="CreateContainer within sandbox \"a0e5e2e8995e15ef0551672ed7f182a58f3501c3bbf130d2cefaf87365852884\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}"
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.185297666Z" level=info msg="CreateContainer within sandbox \"4ef955ccf545587854d8e698935d1f4d8b62a65525164aa039d663a92bfdf19d\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}"
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.215614455Z" level=info msg="CreateContainer within sandbox \"a0e5e2e8995e15ef0551672ed7f182a58f3501c3bbf130d2cefaf87365852884\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"dca8ceecd1b467dc0ac0b59d2fed5eab15114ac667bd66bd10aeb3e648638a11\""
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.216588961Z" level=info msg="StartContainer for \"dca8ceecd1b467dc0ac0b59d2fed5eab15114ac667bd66bd10aeb3e648638a11\""
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.222606488Z" level=info msg="CreateContainer within sandbox \"4ef955ccf545587854d8e698935d1f4d8b62a65525164aa039d663a92bfdf19d\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"746e960627509f5bfda79aa49891d2e8e78b7402b0df7a9c25f4d4a625eee1bb\""
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.223998083Z" level=info msg="StartContainer for \"746e960627509f5bfda79aa49891d2e8e78b7402b0df7a9c25f4d4a625eee1bb\""
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.225507992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-scheduled-stop-735749,Uid:0b8cdda7cd67e20aac922785dec37ab5,Namespace:kube-system,Attempt:0,} returns sandbox id \"9744d282f22e1d4e8b43f69d0003e32b876f2bf14bc144e7084a6816a9f63bad\""
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.228286727Z" level=info msg="CreateContainer within sandbox \"9744d282f22e1d4e8b43f69d0003e32b876f2bf14bc144e7084a6816a9f63bad\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}"
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.245802886Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:etcd-scheduled-stop-735749,Uid:f63237b1ef419fabfb53f84a99be94af,Namespace:kube-system,Attempt:0,} returns sandbox id \"29f54e456d5b167a9994fbf37bdf83ec695600dc214bf4caac75d46663fbc141\""
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.248726831Z" level=info msg="CreateContainer within sandbox \"29f54e456d5b167a9994fbf37bdf83ec695600dc214bf4caac75d46663fbc141\" for container &ContainerMetadata{Name:etcd,Attempt:0,}"
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.254337492Z" level=info msg="CreateContainer within sandbox \"9744d282f22e1d4e8b43f69d0003e32b876f2bf14bc144e7084a6816a9f63bad\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"a4b8c16a2fbc9deb1170f798cdce9f51a243def2048a6a682eb426c29ca4fc90\""
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.272632027Z" level=info msg="StartContainer for \"a4b8c16a2fbc9deb1170f798cdce9f51a243def2048a6a682eb426c29ca4fc90\""
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.291902628Z" level=info msg="CreateContainer within sandbox \"29f54e456d5b167a9994fbf37bdf83ec695600dc214bf4caac75d46663fbc141\" for &ContainerMetadata{Name:etcd,Attempt:0,} returns container id \"5ff22105c9846a8945350c1c0dc8c48ac6f10fb928d9d2723cf2b3c67ab10fe0\""
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.298155881Z" level=info msg="StartContainer for \"5ff22105c9846a8945350c1c0dc8c48ac6f10fb928d9d2723cf2b3c67ab10fe0\""
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.389680904Z" level=info msg="StartContainer for \"dca8ceecd1b467dc0ac0b59d2fed5eab15114ac667bd66bd10aeb3e648638a11\" returns successfully"
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.390010307Z" level=info msg="StartContainer for \"746e960627509f5bfda79aa49891d2e8e78b7402b0df7a9c25f4d4a625eee1bb\" returns successfully"
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.423533806Z" level=info msg="StartContainer for \"a4b8c16a2fbc9deb1170f798cdce9f51a243def2048a6a682eb426c29ca4fc90\" returns successfully"
Jan 27 14:21:05 scheduled-stop-735749 containerd[827]: time="2025-01-27T14:21:05.530419721Z" level=info msg="StartContainer for \"5ff22105c9846a8945350c1c0dc8c48ac6f10fb928d9d2723cf2b3c67ab10fe0\" returns successfully"
==> describe nodes <==
Name: scheduled-stop-735749
Roles: control-plane
Labels: beta.kubernetes.io/arch=arm64
beta.kubernetes.io/os=linux
kubernetes.io/arch=arm64
kubernetes.io/hostname=scheduled-stop-735749
kubernetes.io/os=linux
minikube.k8s.io/commit=6a5089c94d5c3e26f81a121b7614c4f7f440f9c0
minikube.k8s.io/name=scheduled-stop-735749
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2025_01_27T14_21_12_0700
minikube.k8s.io/version=v1.35.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///run/containerd/containerd.sock
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Mon, 27 Jan 2025 14:21:08 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: scheduled-stop-735749
AcquireTime: <unset>
RenewTime: Mon, 27 Jan 2025 14:21:11 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Mon, 27 Jan 2025 14:21:09 +0000 Mon, 27 Jan 2025 14:21:06 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Mon, 27 Jan 2025 14:21:09 +0000 Mon, 27 Jan 2025 14:21:06 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Mon, 27 Jan 2025 14:21:09 +0000 Mon, 27 Jan 2025 14:21:06 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Mon, 27 Jan 2025 14:21:09 +0000 Mon, 27 Jan 2025 14:21:09 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.76.2
Hostname: scheduled-stop-735749
Capacity:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022296Ki
pods: 110
Allocatable:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022296Ki
pods: 110
System Info:
Machine ID: f51cb11673b54d7b8372a1b05b3c1134
System UUID: 2a5e378d-9d2b-4ae1-ab99-5ffbd0a4b846
Boot ID: a1371510-1059-4556-8252-e6af77302cb7
Kernel Version: 5.15.0-1075-aws
OS Image: Ubuntu 22.04.5 LTS
Operating System: linux
Architecture: arm64
Container Runtime Version: containerd://1.7.24
Kubelet Version: v1.32.1
Kube-Proxy Version: v1.32.1
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (5 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
kube-system etcd-scheduled-stop-735749 100m (5%) 0 (0%) 100Mi (1%) 0 (0%) 4s
kube-system kube-apiserver-scheduled-stop-735749 250m (12%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-controller-manager-scheduled-stop-735749 200m (10%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-scheduler-scheduled-stop-735749 100m (5%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 2s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 650m (32%) 0 (0%)
memory 100Mi (1%) 0 (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
hugepages-32Mi 0 (0%) 0 (0%)
hugepages-64Ki 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal NodeHasSufficientMemory 11s (x8 over 11s) kubelet Node scheduled-stop-735749 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 11s (x8 over 11s) kubelet Node scheduled-stop-735749 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 11s (x7 over 11s) kubelet Node scheduled-stop-735749 status is now: NodeHasSufficientPID
Normal NodeAllocatableEnforced 11s kubelet Updated Node Allocatable limit across pods
Normal Starting 4s kubelet Starting kubelet.
Warning CgroupV1 4s kubelet cgroup v1 support is in maintenance mode, please migrate to cgroup v2
Normal NodeAllocatableEnforced 4s kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 4s kubelet Node scheduled-stop-735749 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 4s kubelet Node scheduled-stop-735749 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 4s kubelet Node scheduled-stop-735749 status is now: NodeHasSufficientPID
Normal RegisteredNode 0s node-controller Node scheduled-stop-735749 event: Registered Node scheduled-stop-735749 in Controller
==> dmesg <==
[Jan27 13:18] overlayfs: '/var/lib/containers/storage/overlay/l/Q2QJNMTVZL6GMULS36RA5ZJGSA' not a directory
==> etcd [5ff22105c9846a8945350c1c0dc8c48ac6f10fb928d9d2723cf2b3c67ab10fe0] <==
{"level":"info","ts":"2025-01-27T14:21:05.698146Z","caller":"embed/etcd.go:729","msg":"starting with client TLS","tls-info":"cert = /var/lib/minikube/certs/etcd/server.crt, key = /var/lib/minikube/certs/etcd/server.key, client-cert=, client-key=, trusted-ca = /var/lib/minikube/certs/etcd/ca.crt, client-cert-auth = true, crl-file = ","cipher-suites":[]}
{"level":"info","ts":"2025-01-27T14:21:05.702533Z","caller":"embed/etcd.go:600","msg":"serving peer traffic","address":"192.168.76.2:2380"}
{"level":"info","ts":"2025-01-27T14:21:05.703787Z","caller":"embed/etcd.go:572","msg":"cmux::serve","address":"192.168.76.2:2380"}
{"level":"info","ts":"2025-01-27T14:21:05.703021Z","caller":"embed/etcd.go:280","msg":"now serving peer/client/metrics","local-member-id":"ea7e25599daad906","initial-advertise-peer-urls":["https://192.168.76.2:2380"],"listen-peer-urls":["https://192.168.76.2:2380"],"advertise-client-urls":["https://192.168.76.2:2379"],"listen-client-urls":["https://127.0.0.1:2379","https://192.168.76.2:2379"],"listen-metrics-urls":["http://127.0.0.1:2381"]}
{"level":"info","ts":"2025-01-27T14:21:05.703047Z","caller":"embed/etcd.go:871","msg":"serving metrics","address":"http://127.0.0.1:2381"}
{"level":"info","ts":"2025-01-27T14:21:05.748578Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 is starting a new election at term 1"}
{"level":"info","ts":"2025-01-27T14:21:05.748803Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 became pre-candidate at term 1"}
{"level":"info","ts":"2025-01-27T14:21:05.748929Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 received MsgPreVoteResp from ea7e25599daad906 at term 1"}
{"level":"info","ts":"2025-01-27T14:21:05.749043Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 became candidate at term 2"}
{"level":"info","ts":"2025-01-27T14:21:05.749148Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 received MsgVoteResp from ea7e25599daad906 at term 2"}
{"level":"info","ts":"2025-01-27T14:21:05.749245Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 became leader at term 2"}
{"level":"info","ts":"2025-01-27T14:21:05.749337Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: ea7e25599daad906 elected leader ea7e25599daad906 at term 2"}
{"level":"info","ts":"2025-01-27T14:21:05.751340Z","caller":"etcdserver/server.go:2140","msg":"published local member to cluster through raft","local-member-id":"ea7e25599daad906","local-member-attributes":"{Name:scheduled-stop-735749 ClientURLs:[https://192.168.76.2:2379]}","request-path":"/0/members/ea7e25599daad906/attributes","cluster-id":"6f20f2c4b2fb5f8a","publish-timeout":"7s"}
{"level":"info","ts":"2025-01-27T14:21:05.751602Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2025-01-27T14:21:05.751977Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2025-01-27T14:21:05.753881Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2025-01-27T14:21:05.759927Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2025-01-27T14:21:05.754525Z","caller":"etcdserver/server.go:2651","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2025-01-27T14:21:05.755029Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2025-01-27T14:21:05.758448Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2025-01-27T14:21:05.794723Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2025-01-27T14:21:05.806428Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.76.2:2379"}
{"level":"info","ts":"2025-01-27T14:21:05.822442Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"6f20f2c4b2fb5f8a","local-member-id":"ea7e25599daad906","cluster-version":"3.5"}
{"level":"info","ts":"2025-01-27T14:21:05.822603Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2025-01-27T14:21:05.822668Z","caller":"etcdserver/server.go:2675","msg":"cluster version is updated","cluster-version":"3.5"}
==> kernel <==
14:21:16 up 5:03, 0 users, load average: 1.36, 1.67, 2.19
Linux scheduled-stop-735749 5.15.0-1075-aws #82~20.04.1-Ubuntu SMP Thu Dec 19 05:23:06 UTC 2024 aarch64 aarch64 aarch64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.5 LTS"
==> kube-apiserver [746e960627509f5bfda79aa49891d2e8e78b7402b0df7a9c25f4d4a625eee1bb] <==
I0127 14:21:08.870080 1 cache.go:39] Caches are synced for autoregister controller
I0127 14:21:08.878917 1 cache.go:39] Caches are synced for LocalAvailability controller
I0127 14:21:08.879171 1 apf_controller.go:382] Running API Priority and Fairness config worker
I0127 14:21:08.879317 1 apf_controller.go:385] Running API Priority and Fairness periodic rebalancing process
I0127 14:21:08.880043 1 cache.go:39] Caches are synced for RemoteAvailability controller
I0127 14:21:08.893087 1 cache.go:39] Caches are synced for APIServiceRegistrationController controller
I0127 14:21:08.893368 1 handler_discovery.go:451] Starting ResourceDiscoveryManager
E0127 14:21:08.910069 1 controller.go:145] "Failed to ensure lease exists, will retry" err="namespaces \"kube-system\" not found" interval="200ms"
I0127 14:21:08.911546 1 shared_informer.go:320] Caches are synced for *generic.policySource[*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicy,*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicyBinding,k8s.io/apiserver/pkg/admission/plugin/policy/validating.Validator]
I0127 14:21:08.911772 1 policy_source.go:240] refreshing policies
I0127 14:21:08.934033 1 controller.go:615] quota admission added evaluator for: namespaces
I0127 14:21:09.121733 1 controller.go:615] quota admission added evaluator for: leases.coordination.k8s.io
I0127 14:21:09.698120 1 storage_scheduling.go:95] created PriorityClass system-node-critical with value 2000001000
I0127 14:21:09.706842 1 storage_scheduling.go:95] created PriorityClass system-cluster-critical with value 2000000000
I0127 14:21:09.706866 1 storage_scheduling.go:111] all system priority classes are created successfully or already exist.
I0127 14:21:10.339841 1 controller.go:615] quota admission added evaluator for: roles.rbac.authorization.k8s.io
I0127 14:21:10.394267 1 controller.go:615] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
I0127 14:21:10.505659 1 alloc.go:330] "allocated clusterIPs" service="default/kubernetes" clusterIPs={"IPv4":"10.96.0.1"}
W0127 14:21:10.513039 1 lease.go:265] Resetting endpoints for master service "kubernetes" to [192.168.76.2]
I0127 14:21:10.514171 1 controller.go:615] quota admission added evaluator for: endpoints
I0127 14:21:10.520302 1 controller.go:615] quota admission added evaluator for: endpointslices.discovery.k8s.io
I0127 14:21:10.812395 1 controller.go:615] quota admission added evaluator for: serviceaccounts
I0127 14:21:11.680242 1 controller.go:615] quota admission added evaluator for: deployments.apps
I0127 14:21:11.694203 1 alloc.go:330] "allocated clusterIPs" service="kube-system/kube-dns" clusterIPs={"IPv4":"10.96.0.10"}
I0127 14:21:11.705549 1 controller.go:615] quota admission added evaluator for: daemonsets.apps
==> kube-controller-manager [dca8ceecd1b467dc0ac0b59d2fed5eab15114ac667bd66bd10aeb3e648638a11] <==
I0127 14:21:15.366919 1 shared_informer.go:320] Caches are synced for PV protection
I0127 14:21:15.367304 1 shared_informer.go:320] Caches are synced for crt configmap
I0127 14:21:15.367629 1 shared_informer.go:320] Caches are synced for expand
I0127 14:21:15.367787 1 shared_informer.go:320] Caches are synced for bootstrap_signer
I0127 14:21:15.372631 1 shared_informer.go:320] Caches are synced for namespace
I0127 14:21:15.382131 1 shared_informer.go:320] Caches are synced for TTL after finished
I0127 14:21:15.383289 1 shared_informer.go:320] Caches are synced for resource quota
I0127 14:21:15.402755 1 shared_informer.go:320] Caches are synced for garbage collector
I0127 14:21:15.409513 1 shared_informer.go:320] Caches are synced for taint-eviction-controller
I0127 14:21:15.409581 1 shared_informer.go:320] Caches are synced for disruption
I0127 14:21:15.409760 1 shared_informer.go:320] Caches are synced for garbage collector
I0127 14:21:15.409771 1 garbagecollector.go:154] "Garbage collector: all resource monitors have synced" logger="garbage-collector-controller"
I0127 14:21:15.409779 1 garbagecollector.go:157] "Proceeding to collect garbage" logger="garbage-collector-controller"
I0127 14:21:15.411386 1 shared_informer.go:320] Caches are synced for endpoint_slice_mirroring
I0127 14:21:15.411501 1 shared_informer.go:320] Caches are synced for PVC protection
I0127 14:21:15.411543 1 shared_informer.go:320] Caches are synced for TTL
I0127 14:21:15.412207 1 shared_informer.go:320] Caches are synced for daemon sets
I0127 14:21:15.414439 1 shared_informer.go:320] Caches are synced for ClusterRoleAggregator
I0127 14:21:15.414521 1 shared_informer.go:320] Caches are synced for deployment
I0127 14:21:15.414959 1 shared_informer.go:320] Caches are synced for resource quota
I0127 14:21:15.416487 1 shared_informer.go:320] Caches are synced for ReplicationController
I0127 14:21:15.417588 1 shared_informer.go:320] Caches are synced for job
I0127 14:21:15.421926 1 shared_informer.go:320] Caches are synced for ephemeral
I0127 14:21:15.532208 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="scheduled-stop-735749"
I0127 14:21:16.222341 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="scheduled-stop-735749"
==> kube-scheduler [a4b8c16a2fbc9deb1170f798cdce9f51a243def2048a6a682eb426c29ca4fc90] <==
W0127 14:21:09.144202 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
E0127 14:21:09.144248 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0127 14:21:09.144554 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0127 14:21:09.144836 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0127 14:21:09.144726 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0127 14:21:09.144925 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0127 14:21:09.144767 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E0127 14:21:09.145084 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0127 14:21:09.144802 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0127 14:21:09.145247 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0127 14:21:10.017717 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0127 14:21:10.018024 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0127 14:21:10.072477 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
E0127 14:21:10.072518 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0127 14:21:10.079496 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0127 14:21:10.079540 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0127 14:21:10.087979 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0127 14:21:10.088025 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0127 14:21:10.114676 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.VolumeAttachment: volumeattachments.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "volumeattachments" in API group "storage.k8s.io" at the cluster scope
E0127 14:21:10.114885 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.VolumeAttachment: failed to list *v1.VolumeAttachment: volumeattachments.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"volumeattachments\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0127 14:21:10.117155 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0127 14:21:10.117210 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
W0127 14:21:10.376978 1 reflector.go:569] runtime/asm_arm64.s:1223: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0127 14:21:10.377105 1 reflector.go:166] "Unhandled Error" err="runtime/asm_arm64.s:1223: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
I0127 14:21:13.434894 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Jan 27 14:21:11 scheduled-stop-735749 kubelet[1552]: I0127 14:21:11.968931 1552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e0dacb0c468f59571819daff087d0add-etc-ca-certificates\") pod \"kube-controller-manager-scheduled-stop-735749\" (UID: \"e0dacb0c468f59571819daff087d0add\") " pod="kube-system/kube-controller-manager-scheduled-stop-735749"
Jan 27 14:21:12 scheduled-stop-735749 kubelet[1552]: I0127 14:21:12.525523 1552 apiserver.go:52] "Watching apiserver"
Jan 27 14:21:12 scheduled-stop-735749 kubelet[1552]: I0127 14:21:12.562758 1552 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world"
Jan 27 14:21:12 scheduled-stop-735749 kubelet[1552]: I0127 14:21:12.638645 1552 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-scheduled-stop-735749"
Jan 27 14:21:12 scheduled-stop-735749 kubelet[1552]: I0127 14:21:12.643428 1552 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/etcd-scheduled-stop-735749"
Jan 27 14:21:12 scheduled-stop-735749 kubelet[1552]: E0127 14:21:12.652495 1552 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-scheduled-stop-735749\" already exists" pod="kube-system/kube-apiserver-scheduled-stop-735749"
Jan 27 14:21:12 scheduled-stop-735749 kubelet[1552]: E0127 14:21:12.660923 1552 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"etcd-scheduled-stop-735749\" already exists" pod="kube-system/etcd-scheduled-stop-735749"
Jan 27 14:21:12 scheduled-stop-735749 kubelet[1552]: I0127 14:21:12.785470 1552 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-scheduled-stop-735749" podStartSLOduration=1.7854360599999999 podStartE2EDuration="1.78543606s" podCreationTimestamp="2025-01-27 14:21:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-27 14:21:12.767367554 +0000 UTC m=+1.322078293" watchObservedRunningTime="2025-01-27 14:21:12.78543606 +0000 UTC m=+1.340146799"
Jan 27 14:21:12 scheduled-stop-735749 kubelet[1552]: I0127 14:21:12.804496 1552 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/etcd-scheduled-stop-735749" podStartSLOduration=1.8044758810000001 podStartE2EDuration="1.804475881s" podCreationTimestamp="2025-01-27 14:21:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-27 14:21:12.785882752 +0000 UTC m=+1.340593499" watchObservedRunningTime="2025-01-27 14:21:12.804475881 +0000 UTC m=+1.359186619"
Jan 27 14:21:12 scheduled-stop-735749 kubelet[1552]: I0127 14:21:12.824443 1552 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-scheduled-stop-735749" podStartSLOduration=1.8244166499999999 podStartE2EDuration="1.82441665s" podCreationTimestamp="2025-01-27 14:21:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-27 14:21:12.805001209 +0000 UTC m=+1.359711964" watchObservedRunningTime="2025-01-27 14:21:12.82441665 +0000 UTC m=+1.379127389"
Jan 27 14:21:15 scheduled-stop-735749 kubelet[1552]: I0127 14:21:15.536405 1552 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-scheduled-stop-735749" podStartSLOduration=4.536383775 podStartE2EDuration="4.536383775s" podCreationTimestamp="2025-01-27 14:21:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-27 14:21:12.824240884 +0000 UTC m=+1.378951640" watchObservedRunningTime="2025-01-27 14:21:15.536383775 +0000 UTC m=+4.091094514"
Jan 27 14:21:15 scheduled-stop-735749 kubelet[1552]: I0127 14:21:15.701110 1552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/host-path/441b63a4-fb56-4eaa-a223-1dcf19c2bcaf-tmp\") pod \"storage-provisioner\" (UID: \"441b63a4-fb56-4eaa-a223-1dcf19c2bcaf\") " pod="kube-system/storage-provisioner"
Jan 27 14:21:15 scheduled-stop-735749 kubelet[1552]: I0127 14:21:15.701272 1552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scwsb\" (UniqueName: \"kubernetes.io/projected/441b63a4-fb56-4eaa-a223-1dcf19c2bcaf-kube-api-access-scwsb\") pod \"storage-provisioner\" (UID: \"441b63a4-fb56-4eaa-a223-1dcf19c2bcaf\") " pod="kube-system/storage-provisioner"
Jan 27 14:21:15 scheduled-stop-735749 kubelet[1552]: E0127 14:21:15.811953 1552 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found
Jan 27 14:21:15 scheduled-stop-735749 kubelet[1552]: E0127 14:21:15.812004 1552 projected.go:194] Error preparing data for projected volume kube-api-access-scwsb for pod kube-system/storage-provisioner: configmap "kube-root-ca.crt" not found
Jan 27 14:21:15 scheduled-stop-735749 kubelet[1552]: E0127 14:21:15.812095 1552 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/441b63a4-fb56-4eaa-a223-1dcf19c2bcaf-kube-api-access-scwsb podName:441b63a4-fb56-4eaa-a223-1dcf19c2bcaf nodeName:}" failed. No retries permitted until 2025-01-27 14:21:16.312056594 +0000 UTC m=+4.866767341 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-scwsb" (UniqueName: "kubernetes.io/projected/441b63a4-fb56-4eaa-a223-1dcf19c2bcaf-kube-api-access-scwsb") pod "storage-provisioner" (UID: "441b63a4-fb56-4eaa-a223-1dcf19c2bcaf") : configmap "kube-root-ca.crt" not found
Jan 27 14:21:16 scheduled-stop-735749 kubelet[1552]: I0127 14:21:16.322077 1552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/127f907b-0f52-457d-901a-3c595e5cc1fc-lib-modules\") pod \"kindnet-t92vm\" (UID: \"127f907b-0f52-457d-901a-3c595e5cc1fc\") " pod="kube-system/kindnet-t92vm"
Jan 27 14:21:16 scheduled-stop-735749 kubelet[1552]: I0127 14:21:16.322348 1552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cf4de18e-8369-4392-bf76-4657c7df1a52-lib-modules\") pod \"kube-proxy-dmb8w\" (UID: \"cf4de18e-8369-4392-bf76-4657c7df1a52\") " pod="kube-system/kube-proxy-dmb8w"
Jan 27 14:21:16 scheduled-stop-735749 kubelet[1552]: I0127 14:21:16.322409 1552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jnvq\" (UniqueName: \"kubernetes.io/projected/cf4de18e-8369-4392-bf76-4657c7df1a52-kube-api-access-6jnvq\") pod \"kube-proxy-dmb8w\" (UID: \"cf4de18e-8369-4392-bf76-4657c7df1a52\") " pod="kube-system/kube-proxy-dmb8w"
Jan 27 14:21:16 scheduled-stop-735749 kubelet[1552]: I0127 14:21:16.322459 1552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/127f907b-0f52-457d-901a-3c595e5cc1fc-xtables-lock\") pod \"kindnet-t92vm\" (UID: \"127f907b-0f52-457d-901a-3c595e5cc1fc\") " pod="kube-system/kindnet-t92vm"
Jan 27 14:21:16 scheduled-stop-735749 kubelet[1552]: I0127 14:21:16.322479 1552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h879\" (UniqueName: \"kubernetes.io/projected/127f907b-0f52-457d-901a-3c595e5cc1fc-kube-api-access-4h879\") pod \"kindnet-t92vm\" (UID: \"127f907b-0f52-457d-901a-3c595e5cc1fc\") " pod="kube-system/kindnet-t92vm"
Jan 27 14:21:16 scheduled-stop-735749 kubelet[1552]: I0127 14:21:16.322500 1552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/cf4de18e-8369-4392-bf76-4657c7df1a52-kube-proxy\") pod \"kube-proxy-dmb8w\" (UID: \"cf4de18e-8369-4392-bf76-4657c7df1a52\") " pod="kube-system/kube-proxy-dmb8w"
Jan 27 14:21:16 scheduled-stop-735749 kubelet[1552]: I0127 14:21:16.322517 1552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/cf4de18e-8369-4392-bf76-4657c7df1a52-xtables-lock\") pod \"kube-proxy-dmb8w\" (UID: \"cf4de18e-8369-4392-bf76-4657c7df1a52\") " pod="kube-system/kube-proxy-dmb8w"
Jan 27 14:21:16 scheduled-stop-735749 kubelet[1552]: I0127 14:21:16.322538 1552 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-cfg\" (UniqueName: \"kubernetes.io/host-path/127f907b-0f52-457d-901a-3c595e5cc1fc-cni-cfg\") pod \"kindnet-t92vm\" (UID: \"127f907b-0f52-457d-901a-3c595e5cc1fc\") " pod="kube-system/kindnet-t92vm"
Jan 27 14:21:16 scheduled-stop-735749 kubelet[1552]: I0127 14:21:16.322927 1552 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory"
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-arm64 status --format={{.APIServer}} -p scheduled-stop-735749 -n scheduled-stop-735749
helpers_test.go:261: (dbg) Run: kubectl --context scheduled-stop-735749 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: coredns-668d6bf9bc-fjnt9 kindnet-t92vm kube-proxy-dmb8w storage-provisioner
helpers_test.go:274: ======> post-mortem[TestScheduledStopUnix]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context scheduled-stop-735749 describe pod coredns-668d6bf9bc-fjnt9 kindnet-t92vm kube-proxy-dmb8w storage-provisioner
helpers_test.go:277: (dbg) Non-zero exit: kubectl --context scheduled-stop-735749 describe pod coredns-668d6bf9bc-fjnt9 kindnet-t92vm kube-proxy-dmb8w storage-provisioner: exit status 1 (112.266886ms)
** stderr **
Error from server (NotFound): pods "coredns-668d6bf9bc-fjnt9" not found
Error from server (NotFound): pods "kindnet-t92vm" not found
Error from server (NotFound): pods "kube-proxy-dmb8w" not found
Error from server (NotFound): pods "storage-provisioner" not found
** /stderr **
helpers_test.go:279: kubectl --context scheduled-stop-735749 describe pod coredns-668d6bf9bc-fjnt9 kindnet-t92vm kube-proxy-dmb8w storage-provisioner: exit status 1
helpers_test.go:175: Cleaning up "scheduled-stop-735749" profile ...
helpers_test.go:178: (dbg) Run: out/minikube-linux-arm64 delete -p scheduled-stop-735749
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p scheduled-stop-735749: (1.962843658s)
--- FAIL: TestScheduledStopUnix (37.32s)