=== RUN TestScheduledStopUnix
scheduled_stop_test.go:128: (dbg) Run: out/minikube-linux-arm64 start -p scheduled-stop-403142 --memory=2048 --driver=docker --container-runtime=containerd
E0127 02:42:06.565397 3586800 cert_rotation.go:171] "Unhandled Error" err="key failed with : open /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/addons-791589/client.crt: no such file or directory" logger="UnhandledError"
scheduled_stop_test.go:128: (dbg) Done: out/minikube-linux-arm64 start -p scheduled-stop-403142 --memory=2048 --driver=docker --container-runtime=containerd: (33.411837461s)
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-403142 --schedule 5m
scheduled_stop_test.go:191: (dbg) Run: out/minikube-linux-arm64 status --format={{.TimeToStop}} -p scheduled-stop-403142 -n scheduled-stop-403142
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:137: (dbg) Run: out/minikube-linux-arm64 stop -p scheduled-stop-403142 --schedule 15s
scheduled_stop_test.go:169: signal error was: <nil>
scheduled_stop_test.go:98: process 3737384 running but should have been killed on reschedule of stop
panic.go:629: *** TestScheduledStopUnix FAILED at 2025-01-27 02:42:26.704111092 +0000 UTC m=+2112.324694908
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestScheduledStopUnix]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect scheduled-stop-403142
helpers_test.go:235: (dbg) docker inspect scheduled-stop-403142:
-- stdout --
[
{
"Id": "9ed9f87e1d5d08074c62f163f5255e997ce686d01b3d3d0c1c0d2d4b6fdd600b",
"Created": "2025-01-27T02:41:58.183616506Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 3735450,
"ExitCode": 0,
"Error": "",
"StartedAt": "2025-01-27T02:41:58.347203991Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:0434cf58b6dbace281e5de753aa4b2e3fe33dc9a3be53021531403743c3f155a",
"ResolvConfPath": "/var/lib/docker/containers/9ed9f87e1d5d08074c62f163f5255e997ce686d01b3d3d0c1c0d2d4b6fdd600b/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/9ed9f87e1d5d08074c62f163f5255e997ce686d01b3d3d0c1c0d2d4b6fdd600b/hostname",
"HostsPath": "/var/lib/docker/containers/9ed9f87e1d5d08074c62f163f5255e997ce686d01b3d3d0c1c0d2d4b6fdd600b/hosts",
"LogPath": "/var/lib/docker/containers/9ed9f87e1d5d08074c62f163f5255e997ce686d01b3d3d0c1c0d2d4b6fdd600b/9ed9f87e1d5d08074c62f163f5255e997ce686d01b3d3d0c1c0d2d4b6fdd600b-json.log",
"Name": "/scheduled-stop-403142",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"scheduled-stop-403142:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {}
},
"NetworkMode": "scheduled-stop-403142",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 2147483648,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 4294967296,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/15bae22734b53d27129ce653c37ac409debbe222a5efa8f33abbe880d70ec95d-init/diff:/var/lib/docker/overlay2/5296668a0a30b38feb9159e191c47d5587ed9f36bb9a48e894c12f88095e8aab/diff",
"MergedDir": "/var/lib/docker/overlay2/15bae22734b53d27129ce653c37ac409debbe222a5efa8f33abbe880d70ec95d/merged",
"UpperDir": "/var/lib/docker/overlay2/15bae22734b53d27129ce653c37ac409debbe222a5efa8f33abbe880d70ec95d/diff",
"WorkDir": "/var/lib/docker/overlay2/15bae22734b53d27129ce653c37ac409debbe222a5efa8f33abbe880d70ec95d/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "scheduled-stop-403142",
"Source": "/var/lib/docker/volumes/scheduled-stop-403142/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "scheduled-stop-403142",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "scheduled-stop-403142",
"name.minikube.sigs.k8s.io": "scheduled-stop-403142",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "5babefdb6db16583dd57f5fc3adb4f18904ad77ee7f07b2779596fa22bab403e",
"SandboxKey": "/var/run/docker/netns/5babefdb6db1",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "37686"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "37687"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "37690"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "37688"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "37689"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"scheduled-stop-403142": {
"IPAMConfig": {
"IPv4Address": "192.168.76.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:4c:02",
"DriverOpts": null,
"NetworkID": "d4f735f82bda2544cc2a505f6f918bac8bb135054ef3ef665f568475efd5db1b",
"EndpointID": "65cd271e7591678f8062713a6d9868aa87c825102187ec522d70441f57ed8f2a",
"Gateway": "192.168.76.1",
"IPAddress": "192.168.76.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"scheduled-stop-403142",
"9ed9f87e1d5d"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-arm64 status --format={{.Host}} -p scheduled-stop-403142 -n scheduled-stop-403142
helpers_test.go:244: <<< TestScheduledStopUnix FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestScheduledStopUnix]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-arm64 -p scheduled-stop-403142 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-arm64 -p scheduled-stop-403142 logs -n 25: (1.394027868s)
helpers_test.go:252: TestScheduledStopUnix logs:
-- stdout --
==> Audit <==
|---------|--------------------------------|-----------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|--------------------------------|-----------------------|---------|---------|---------------------|---------------------|
| stop | -p multinode-136196 | multinode-136196 | jenkins | v1.35.0 | 27 Jan 25 02:35 UTC | 27 Jan 25 02:36 UTC |
| start | -p multinode-136196 | multinode-136196 | jenkins | v1.35.0 | 27 Jan 25 02:36 UTC | 27 Jan 25 02:37 UTC |
| | --wait=true -v=8 | | | | | |
| | --alsologtostderr | | | | | |
| node | list -p multinode-136196 | multinode-136196 | jenkins | v1.35.0 | 27 Jan 25 02:37 UTC | |
| node | multinode-136196 node delete | multinode-136196 | jenkins | v1.35.0 | 27 Jan 25 02:37 UTC | 27 Jan 25 02:38 UTC |
| | m03 | | | | | |
| stop | multinode-136196 stop | multinode-136196 | jenkins | v1.35.0 | 27 Jan 25 02:38 UTC | 27 Jan 25 02:38 UTC |
| start | -p multinode-136196 | multinode-136196 | jenkins | v1.35.0 | 27 Jan 25 02:38 UTC | 27 Jan 25 02:39 UTC |
| | --wait=true -v=8 | | | | | |
| | --alsologtostderr | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| node | list -p multinode-136196 | multinode-136196 | jenkins | v1.35.0 | 27 Jan 25 02:39 UTC | |
| start | -p multinode-136196-m02 | multinode-136196-m02 | jenkins | v1.35.0 | 27 Jan 25 02:39 UTC | |
| | --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| start | -p multinode-136196-m03 | multinode-136196-m03 | jenkins | v1.35.0 | 27 Jan 25 02:39 UTC | 27 Jan 25 02:39 UTC |
| | --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| node | add -p multinode-136196 | multinode-136196 | jenkins | v1.35.0 | 27 Jan 25 02:39 UTC | |
| delete | -p multinode-136196-m03 | multinode-136196-m03 | jenkins | v1.35.0 | 27 Jan 25 02:39 UTC | 27 Jan 25 02:39 UTC |
| delete | -p multinode-136196 | multinode-136196 | jenkins | v1.35.0 | 27 Jan 25 02:39 UTC | 27 Jan 25 02:40 UTC |
| start | -p test-preload-965764 | test-preload-965764 | jenkins | v1.35.0 | 27 Jan 25 02:40 UTC | 27 Jan 25 02:41 UTC |
| | --memory=2200 | | | | | |
| | --alsologtostderr | | | | | |
| | --wait=true --preload=false | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| | --kubernetes-version=v1.24.4 | | | | | |
| image | test-preload-965764 image pull | test-preload-965764 | jenkins | v1.35.0 | 27 Jan 25 02:41 UTC | 27 Jan 25 02:41 UTC |
| | gcr.io/k8s-minikube/busybox | | | | | |
| stop | -p test-preload-965764 | test-preload-965764 | jenkins | v1.35.0 | 27 Jan 25 02:41 UTC | 27 Jan 25 02:41 UTC |
| start | -p test-preload-965764 | test-preload-965764 | jenkins | v1.35.0 | 27 Jan 25 02:41 UTC | 27 Jan 25 02:41 UTC |
| | --memory=2200 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| | --wait=true --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| image | test-preload-965764 image list | test-preload-965764 | jenkins | v1.35.0 | 27 Jan 25 02:41 UTC | 27 Jan 25 02:41 UTC |
| delete | -p test-preload-965764 | test-preload-965764 | jenkins | v1.35.0 | 27 Jan 25 02:41 UTC | 27 Jan 25 02:41 UTC |
| start | -p scheduled-stop-403142 | scheduled-stop-403142 | jenkins | v1.35.0 | 27 Jan 25 02:41 UTC | 27 Jan 25 02:42 UTC |
| | --memory=2048 --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| stop | -p scheduled-stop-403142 | scheduled-stop-403142 | jenkins | v1.35.0 | 27 Jan 25 02:42 UTC | |
| | --schedule 5m | | | | | |
| stop | -p scheduled-stop-403142 | scheduled-stop-403142 | jenkins | v1.35.0 | 27 Jan 25 02:42 UTC | |
| | --schedule 5m | | | | | |
| stop | -p scheduled-stop-403142 | scheduled-stop-403142 | jenkins | v1.35.0 | 27 Jan 25 02:42 UTC | |
| | --schedule 5m | | | | | |
| stop | -p scheduled-stop-403142 | scheduled-stop-403142 | jenkins | v1.35.0 | 27 Jan 25 02:42 UTC | |
| | --schedule 15s | | | | | |
| stop | -p scheduled-stop-403142 | scheduled-stop-403142 | jenkins | v1.35.0 | 27 Jan 25 02:42 UTC | |
| | --schedule 15s | | | | | |
| stop | -p scheduled-stop-403142 | scheduled-stop-403142 | jenkins | v1.35.0 | 27 Jan 25 02:42 UTC | |
| | --schedule 15s | | | | | |
|---------|--------------------------------|-----------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2025/01/27 02:41:52
Running on machine: ip-172-31-21-244
Binary: Built with gc go1.23.4 for linux/arm64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0127 02:41:52.825280 3734956 out.go:345] Setting OutFile to fd 1 ...
I0127 02:41:52.825419 3734956 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0127 02:41:52.825423 3734956 out.go:358] Setting ErrFile to fd 2...
I0127 02:41:52.825427 3734956 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0127 02:41:52.825665 3734956 root.go:338] Updating PATH: /home/jenkins/minikube-integration/20316-3581420/.minikube/bin
I0127 02:41:52.826032 3734956 out.go:352] Setting JSON to false
I0127 02:41:52.826930 3734956 start.go:129] hostinfo: {"hostname":"ip-172-31-21-244","uptime":91457,"bootTime":1737854256,"procs":164,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1075-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"da8ac1fd-6236-412a-a346-95873c98230d"}
I0127 02:41:52.826986 3734956 start.go:139] virtualization:
I0127 02:41:52.830948 3734956 out.go:177] * [scheduled-stop-403142] minikube v1.35.0 on Ubuntu 20.04 (arm64)
I0127 02:41:52.835463 3734956 out.go:177] - MINIKUBE_LOCATION=20316
I0127 02:41:52.835589 3734956 notify.go:220] Checking for updates...
I0127 02:41:52.842162 3734956 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0127 02:41:52.845402 3734956 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/20316-3581420/kubeconfig
I0127 02:41:52.848471 3734956 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/20316-3581420/.minikube
I0127 02:41:52.851698 3734956 out.go:177] - MINIKUBE_BIN=out/minikube-linux-arm64
I0127 02:41:52.854817 3734956 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0127 02:41:52.858269 3734956 driver.go:394] Setting default libvirt URI to qemu:///system
I0127 02:41:52.883689 3734956 docker.go:123] docker version: linux-27.5.1:Docker Engine - Community
I0127 02:41:52.883832 3734956 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0127 02:41:52.941852 3734956 info.go:266] docker info: {ID:5FDH:SA5P:5GCT:NLAS:B73P:SGDQ:PBG5:UBVH:UZY3:RXGO:CI7S:WAIH Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:24 OomKillDisable:true NGoroutines:43 SystemTime:2025-01-27 02:41:52.932358168 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1075-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214835200 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-21-244 Labels:[] ExperimentalBuild:false ServerVersion:27.5.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:bcc810d6b9066471b0b6fa75f557a15a1cbf31bb Expected:bcc810d6b9066471b0b6fa75f557a15a1cbf31bb} RuncCommit:{ID:v1.2.4-0-g6c52b3f Expected:v1.2.4-0-g6c52b3f} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErro
rs:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.20.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.32.4]] Warnings:<nil>}}
I0127 02:41:52.941948 3734956 docker.go:318] overlay module found
I0127 02:41:52.947128 3734956 out.go:177] * Using the docker driver based on user configuration
I0127 02:41:52.949986 3734956 start.go:297] selected driver: docker
I0127 02:41:52.950001 3734956 start.go:901] validating driver "docker" against <nil>
I0127 02:41:52.950014 3734956 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0127 02:41:52.950874 3734956 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0127 02:41:53.009850 3734956 info.go:266] docker info: {ID:5FDH:SA5P:5GCT:NLAS:B73P:SGDQ:PBG5:UBVH:UZY3:RXGO:CI7S:WAIH Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:3 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:24 OomKillDisable:true NGoroutines:43 SystemTime:2025-01-27 02:41:52.999947196 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1075-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:a
arch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214835200 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-21-244 Labels:[] ExperimentalBuild:false ServerVersion:27.5.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:bcc810d6b9066471b0b6fa75f557a15a1cbf31bb Expected:bcc810d6b9066471b0b6fa75f557a15a1cbf31bb} RuncCommit:{ID:v1.2.4-0-g6c52b3f Expected:v1.2.4-0-g6c52b3f} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErro
rs:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.20.0] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.32.4]] Warnings:<nil>}}
I0127 02:41:53.010061 3734956 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0127 02:41:53.010387 3734956 start_flags.go:929] Wait components to verify : map[apiserver:true system_pods:true]
I0127 02:41:53.013443 3734956 out.go:177] * Using Docker driver with root privileges
I0127 02:41:53.016429 3734956 cni.go:84] Creating CNI manager for ""
I0127 02:41:53.016499 3734956 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I0127 02:41:53.016506 3734956 start_flags.go:319] Found "CNI" CNI - setting NetworkPlugin=cni
I0127 02:41:53.016599 3734956 start.go:340] cluster config:
{Name:scheduled-stop-403142 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.32.1 ClusterName:scheduled-stop-403142 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:contain
erd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.32.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0127 02:41:53.019930 3734956 out.go:177] * Starting "scheduled-stop-403142" primary control-plane node in "scheduled-stop-403142" cluster
I0127 02:41:53.022903 3734956 cache.go:121] Beginning downloading kic base image for docker with containerd
I0127 02:41:53.025933 3734956 out.go:177] * Pulling base image v0.0.46 ...
I0127 02:41:53.028797 3734956 preload.go:131] Checking if preload exists for k8s version v1.32.1 and runtime containerd
I0127 02:41:53.028848 3734956 preload.go:146] Found local preload: /home/jenkins/minikube-integration/20316-3581420/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.32.1-containerd-overlay2-arm64.tar.lz4
I0127 02:41:53.028856 3734956 cache.go:56] Caching tarball of preloaded images
I0127 02:41:53.028882 3734956 image.go:81] Checking for gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 in local docker daemon
I0127 02:41:53.028967 3734956 preload.go:172] Found /home/jenkins/minikube-integration/20316-3581420/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.32.1-containerd-overlay2-arm64.tar.lz4 in cache, skipping download
I0127 02:41:53.028977 3734956 cache.go:59] Finished verifying existence of preloaded tar for v1.32.1 on containerd
I0127 02:41:53.029362 3734956 profile.go:143] Saving config to /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/config.json ...
I0127 02:41:53.029383 3734956 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/config.json: {Name:mk78408a5703760627a4cabe26efc9b49c20c124 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 02:41:53.050301 3734956 image.go:100] Found gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 in local docker daemon, skipping pull
I0127 02:41:53.050313 3734956 cache.go:145] gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 exists in daemon, skipping load
I0127 02:41:53.050325 3734956 cache.go:230] Successfully downloaded all kic artifacts
I0127 02:41:53.050358 3734956 start.go:360] acquireMachinesLock for scheduled-stop-403142: {Name:mk82b25d60d641edf186439411207ad6c043b1cb Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0127 02:41:53.050464 3734956 start.go:364] duration metric: took 92.051µs to acquireMachinesLock for "scheduled-stop-403142"
I0127 02:41:53.050488 3734956 start.go:93] Provisioning new machine with config: &{Name:scheduled-stop-403142 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.32.1 ClusterName:scheduled-stop-403142 Namespace:default APIServerHAVIP: APIServerName:minikubeCA
APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.32.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHA
uthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.32.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
I0127 02:41:53.050555 3734956 start.go:125] createHost starting for "" (driver="docker")
I0127 02:41:53.055794 3734956 out.go:235] * Creating docker container (CPUs=2, Memory=2048MB) ...
I0127 02:41:53.056055 3734956 start.go:159] libmachine.API.Create for "scheduled-stop-403142" (driver="docker")
I0127 02:41:53.056086 3734956 client.go:168] LocalClient.Create starting
I0127 02:41:53.056174 3734956 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/20316-3581420/.minikube/certs/ca.pem
I0127 02:41:53.056207 3734956 main.go:141] libmachine: Decoding PEM data...
I0127 02:41:53.056223 3734956 main.go:141] libmachine: Parsing certificate...
I0127 02:41:53.056273 3734956 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/20316-3581420/.minikube/certs/cert.pem
I0127 02:41:53.056292 3734956 main.go:141] libmachine: Decoding PEM data...
I0127 02:41:53.056301 3734956 main.go:141] libmachine: Parsing certificate...
I0127 02:41:53.056668 3734956 cli_runner.go:164] Run: docker network inspect scheduled-stop-403142 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0127 02:41:53.073313 3734956 cli_runner.go:211] docker network inspect scheduled-stop-403142 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0127 02:41:53.073394 3734956 network_create.go:284] running [docker network inspect scheduled-stop-403142] to gather additional debugging logs...
I0127 02:41:53.073409 3734956 cli_runner.go:164] Run: docker network inspect scheduled-stop-403142
W0127 02:41:53.090682 3734956 cli_runner.go:211] docker network inspect scheduled-stop-403142 returned with exit code 1
I0127 02:41:53.090709 3734956 network_create.go:287] error running [docker network inspect scheduled-stop-403142]: docker network inspect scheduled-stop-403142: exit status 1
stdout:
[]
stderr:
Error response from daemon: network scheduled-stop-403142 not found
I0127 02:41:53.090723 3734956 network_create.go:289] output of [docker network inspect scheduled-stop-403142]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network scheduled-stop-403142 not found
** /stderr **
I0127 02:41:53.090829 3734956 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0127 02:41:53.111882 3734956 network.go:211] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-20c6b9faf740 IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:02:42:a5:84:e8:b3} reservation:<nil>}
I0127 02:41:53.112395 3734956 network.go:211] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-ed55a6afcd29 IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:02:42:ae:45:09:f0} reservation:<nil>}
I0127 02:41:53.112940 3734956 network.go:211] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-6d1bfb053f15 IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:02:42:0f:00:a9:30} reservation:<nil>}
I0127 02:41:53.113541 3734956 network.go:206] using free private subnet 192.168.76.0/24: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x4001949760}
I0127 02:41:53.113569 3734956 network_create.go:124] attempt to create docker network scheduled-stop-403142 192.168.76.0/24 with gateway 192.168.76.1 and MTU of 1500 ...
I0127 02:41:53.113668 3734956 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.76.0/24 --gateway=192.168.76.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=scheduled-stop-403142 scheduled-stop-403142
I0127 02:41:53.189392 3734956 network_create.go:108] docker network scheduled-stop-403142 192.168.76.0/24 created
I0127 02:41:53.189416 3734956 kic.go:121] calculated static IP "192.168.76.2" for the "scheduled-stop-403142" container
I0127 02:41:53.189500 3734956 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0127 02:41:53.205242 3734956 cli_runner.go:164] Run: docker volume create scheduled-stop-403142 --label name.minikube.sigs.k8s.io=scheduled-stop-403142 --label created_by.minikube.sigs.k8s.io=true
I0127 02:41:53.223595 3734956 oci.go:103] Successfully created a docker volume scheduled-stop-403142
I0127 02:41:53.223685 3734956 cli_runner.go:164] Run: docker run --rm --name scheduled-stop-403142-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-403142 --entrypoint /usr/bin/test -v scheduled-stop-403142:/var gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 -d /var/lib
I0127 02:41:53.768580 3734956 oci.go:107] Successfully prepared a docker volume scheduled-stop-403142
I0127 02:41:53.768626 3734956 preload.go:131] Checking if preload exists for k8s version v1.32.1 and runtime containerd
I0127 02:41:53.768643 3734956 kic.go:194] Starting extracting preloaded images to volume ...
I0127 02:41:53.768713 3734956 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/20316-3581420/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.32.1-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-403142:/extractDir gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 -I lz4 -xf /preloaded.tar -C /extractDir
I0127 02:41:58.116009 3734956 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/20316-3581420/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.32.1-containerd-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v scheduled-stop-403142:/extractDir gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 -I lz4 -xf /preloaded.tar -C /extractDir: (4.347199391s)
I0127 02:41:58.116030 3734956 kic.go:203] duration metric: took 4.347383623s to extract preloaded images to volume ...
W0127 02:41:58.116177 3734956 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0127 02:41:58.116287 3734956 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0127 02:41:58.168820 3734956 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname scheduled-stop-403142 --name scheduled-stop-403142 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=scheduled-stop-403142 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=scheduled-stop-403142 --network scheduled-stop-403142 --ip 192.168.76.2 --volume scheduled-stop-403142:/var --security-opt apparmor=unconfined --memory=2048mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279
I0127 02:41:58.507913 3734956 cli_runner.go:164] Run: docker container inspect scheduled-stop-403142 --format={{.State.Running}}
I0127 02:41:58.534091 3734956 cli_runner.go:164] Run: docker container inspect scheduled-stop-403142 --format={{.State.Status}}
I0127 02:41:58.559793 3734956 cli_runner.go:164] Run: docker exec scheduled-stop-403142 stat /var/lib/dpkg/alternatives/iptables
I0127 02:41:58.619804 3734956 oci.go:144] the created container "scheduled-stop-403142" has a running status.
I0127 02:41:58.619825 3734956 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/20316-3581420/.minikube/machines/scheduled-stop-403142/id_rsa...
I0127 02:41:58.843080 3734956 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/20316-3581420/.minikube/machines/scheduled-stop-403142/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0127 02:41:58.880257 3734956 cli_runner.go:164] Run: docker container inspect scheduled-stop-403142 --format={{.State.Status}}
I0127 02:41:58.911997 3734956 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0127 02:41:58.912008 3734956 kic_runner.go:114] Args: [docker exec --privileged scheduled-stop-403142 chown docker:docker /home/docker/.ssh/authorized_keys]
I0127 02:41:58.977955 3734956 cli_runner.go:164] Run: docker container inspect scheduled-stop-403142 --format={{.State.Status}}
I0127 02:41:59.002352 3734956 machine.go:93] provisionDockerMachine start ...
I0127 02:41:59.002441 3734956 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-403142
I0127 02:41:59.025266 3734956 main.go:141] libmachine: Using SSH client type: native
I0127 02:41:59.025554 3734956 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x4132a0] 0x415ae0 <nil> [] 0s} 127.0.0.1 37686 <nil> <nil>}
I0127 02:41:59.025562 3734956 main.go:141] libmachine: About to run SSH command:
hostname
I0127 02:41:59.028502 3734956 main.go:141] libmachine: Error dialing TCP: ssh: handshake failed: EOF
I0127 02:42:02.154300 3734956 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-403142
I0127 02:42:02.154316 3734956 ubuntu.go:169] provisioning hostname "scheduled-stop-403142"
I0127 02:42:02.154396 3734956 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-403142
I0127 02:42:02.172015 3734956 main.go:141] libmachine: Using SSH client type: native
I0127 02:42:02.172262 3734956 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x4132a0] 0x415ae0 <nil> [] 0s} 127.0.0.1 37686 <nil> <nil>}
I0127 02:42:02.172272 3734956 main.go:141] libmachine: About to run SSH command:
sudo hostname scheduled-stop-403142 && echo "scheduled-stop-403142" | sudo tee /etc/hostname
I0127 02:42:02.309934 3734956 main.go:141] libmachine: SSH cmd err, output: <nil>: scheduled-stop-403142
I0127 02:42:02.310005 3734956 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-403142
I0127 02:42:02.328224 3734956 main.go:141] libmachine: Using SSH client type: native
I0127 02:42:02.328472 3734956 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x4132a0] 0x415ae0 <nil> [] 0s} 127.0.0.1 37686 <nil> <nil>}
I0127 02:42:02.328487 3734956 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sscheduled-stop-403142' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 scheduled-stop-403142/g' /etc/hosts;
else
echo '127.0.1.1 scheduled-stop-403142' | sudo tee -a /etc/hosts;
fi
fi
I0127 02:42:02.454082 3734956 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0127 02:42:02.454120 3734956 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/20316-3581420/.minikube CaCertPath:/home/jenkins/minikube-integration/20316-3581420/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/20316-3581420/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/20316-3581420/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/20316-3581420/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/20316-3581420/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/20316-3581420/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/20316-3581420/.minikube}
I0127 02:42:02.454151 3734956 ubuntu.go:177] setting up certificates
I0127 02:42:02.454159 3734956 provision.go:84] configureAuth start
I0127 02:42:02.454216 3734956 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-403142
I0127 02:42:02.473775 3734956 provision.go:143] copyHostCerts
I0127 02:42:02.473835 3734956 exec_runner.go:144] found /home/jenkins/minikube-integration/20316-3581420/.minikube/ca.pem, removing ...
I0127 02:42:02.473842 3734956 exec_runner.go:203] rm: /home/jenkins/minikube-integration/20316-3581420/.minikube/ca.pem
I0127 02:42:02.473919 3734956 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20316-3581420/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/20316-3581420/.minikube/ca.pem (1078 bytes)
I0127 02:42:02.474021 3734956 exec_runner.go:144] found /home/jenkins/minikube-integration/20316-3581420/.minikube/cert.pem, removing ...
I0127 02:42:02.474025 3734956 exec_runner.go:203] rm: /home/jenkins/minikube-integration/20316-3581420/.minikube/cert.pem
I0127 02:42:02.474050 3734956 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20316-3581420/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/20316-3581420/.minikube/cert.pem (1123 bytes)
I0127 02:42:02.474255 3734956 exec_runner.go:144] found /home/jenkins/minikube-integration/20316-3581420/.minikube/key.pem, removing ...
I0127 02:42:02.474260 3734956 exec_runner.go:203] rm: /home/jenkins/minikube-integration/20316-3581420/.minikube/key.pem
I0127 02:42:02.474289 3734956 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20316-3581420/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/20316-3581420/.minikube/key.pem (1679 bytes)
I0127 02:42:02.474364 3734956 provision.go:117] generating server cert: /home/jenkins/minikube-integration/20316-3581420/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/20316-3581420/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/20316-3581420/.minikube/certs/ca-key.pem org=jenkins.scheduled-stop-403142 san=[127.0.0.1 192.168.76.2 localhost minikube scheduled-stop-403142]
I0127 02:42:02.965230 3734956 provision.go:177] copyRemoteCerts
I0127 02:42:02.965289 3734956 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0127 02:42:02.965345 3734956 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-403142
I0127 02:42:02.982267 3734956 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:37686 SSHKeyPath:/home/jenkins/minikube-integration/20316-3581420/.minikube/machines/scheduled-stop-403142/id_rsa Username:docker}
I0127 02:42:03.075330 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0127 02:42:03.100142 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/machines/server.pem --> /etc/docker/server.pem (1229 bytes)
I0127 02:42:03.125607 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
I0127 02:42:03.152824 3734956 provision.go:87] duration metric: took 698.652707ms to configureAuth
I0127 02:42:03.152845 3734956 ubuntu.go:193] setting minikube options for container-runtime
I0127 02:42:03.153041 3734956 config.go:182] Loaded profile config "scheduled-stop-403142": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.32.1
I0127 02:42:03.153047 3734956 machine.go:96] duration metric: took 4.150684598s to provisionDockerMachine
I0127 02:42:03.153052 3734956 client.go:171] duration metric: took 10.096961148s to LocalClient.Create
I0127 02:42:03.153066 3734956 start.go:167] duration metric: took 10.097012486s to libmachine.API.Create "scheduled-stop-403142"
I0127 02:42:03.153072 3734956 start.go:293] postStartSetup for "scheduled-stop-403142" (driver="docker")
I0127 02:42:03.153080 3734956 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0127 02:42:03.153138 3734956 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0127 02:42:03.153182 3734956 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-403142
I0127 02:42:03.173590 3734956 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:37686 SSHKeyPath:/home/jenkins/minikube-integration/20316-3581420/.minikube/machines/scheduled-stop-403142/id_rsa Username:docker}
I0127 02:42:03.263129 3734956 ssh_runner.go:195] Run: cat /etc/os-release
I0127 02:42:03.266384 3734956 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0127 02:42:03.266410 3734956 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0127 02:42:03.266420 3734956 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0127 02:42:03.266426 3734956 info.go:137] Remote host: Ubuntu 22.04.5 LTS
I0127 02:42:03.266436 3734956 filesync.go:126] Scanning /home/jenkins/minikube-integration/20316-3581420/.minikube/addons for local assets ...
I0127 02:42:03.266494 3734956 filesync.go:126] Scanning /home/jenkins/minikube-integration/20316-3581420/.minikube/files for local assets ...
I0127 02:42:03.266586 3734956 filesync.go:149] local asset: /home/jenkins/minikube-integration/20316-3581420/.minikube/files/etc/ssl/certs/35868002.pem -> 35868002.pem in /etc/ssl/certs
I0127 02:42:03.266719 3734956 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I0127 02:42:03.275007 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/files/etc/ssl/certs/35868002.pem --> /etc/ssl/certs/35868002.pem (1708 bytes)
I0127 02:42:03.298634 3734956 start.go:296] duration metric: took 145.549397ms for postStartSetup
I0127 02:42:03.299002 3734956 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-403142
I0127 02:42:03.316592 3734956 profile.go:143] Saving config to /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/config.json ...
I0127 02:42:03.316926 3734956 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0127 02:42:03.316981 3734956 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-403142
I0127 02:42:03.335042 3734956 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:37686 SSHKeyPath:/home/jenkins/minikube-integration/20316-3581420/.minikube/machines/scheduled-stop-403142/id_rsa Username:docker}
I0127 02:42:03.423288 3734956 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0127 02:42:03.427786 3734956 start.go:128] duration metric: took 10.377215852s to createHost
I0127 02:42:03.427802 3734956 start.go:83] releasing machines lock for "scheduled-stop-403142", held for 10.377330983s
I0127 02:42:03.427880 3734956 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" scheduled-stop-403142
I0127 02:42:03.444967 3734956 ssh_runner.go:195] Run: cat /version.json
I0127 02:42:03.445013 3734956 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-403142
I0127 02:42:03.445025 3734956 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0127 02:42:03.445075 3734956 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-403142
I0127 02:42:03.467597 3734956 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:37686 SSHKeyPath:/home/jenkins/minikube-integration/20316-3581420/.minikube/machines/scheduled-stop-403142/id_rsa Username:docker}
I0127 02:42:03.470479 3734956 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:37686 SSHKeyPath:/home/jenkins/minikube-integration/20316-3581420/.minikube/machines/scheduled-stop-403142/id_rsa Username:docker}
I0127 02:42:03.692235 3734956 ssh_runner.go:195] Run: systemctl --version
I0127 02:42:03.696495 3734956 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0127 02:42:03.700663 3734956 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0127 02:42:03.724587 3734956 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0127 02:42:03.724656 3734956 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0127 02:42:03.755982 3734956 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0127 02:42:03.755995 3734956 start.go:495] detecting cgroup driver to use...
I0127 02:42:03.756027 3734956 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0127 02:42:03.756076 3734956 ssh_runner.go:195] Run: sudo systemctl stop -f crio
I0127 02:42:03.768656 3734956 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0127 02:42:03.780319 3734956 docker.go:217] disabling cri-docker service (if available) ...
I0127 02:42:03.780372 3734956 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
I0127 02:42:03.794525 3734956 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
I0127 02:42:03.809242 3734956 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
I0127 02:42:03.903374 3734956 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
I0127 02:42:03.998798 3734956 docker.go:233] disabling docker service ...
I0127 02:42:03.998889 3734956 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
I0127 02:42:04.023209 3734956 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
I0127 02:42:04.036170 3734956 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
I0127 02:42:04.130949 3734956 ssh_runner.go:195] Run: sudo systemctl mask docker.service
I0127 02:42:04.217343 3734956 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
I0127 02:42:04.228915 3734956 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0127 02:42:04.245417 3734956 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0127 02:42:04.256430 3734956 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0127 02:42:04.266294 3734956 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0127 02:42:04.266354 3734956 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0127 02:42:04.276613 3734956 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0127 02:42:04.286452 3734956 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0127 02:42:04.297109 3734956 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0127 02:42:04.306931 3734956 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0127 02:42:04.315970 3734956 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0127 02:42:04.325452 3734956 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0127 02:42:04.335325 3734956 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0127 02:42:04.344732 3734956 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0127 02:42:04.353456 3734956 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0127 02:42:04.362236 3734956 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0127 02:42:04.451378 3734956 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0127 02:42:04.582239 3734956 start.go:542] Will wait 60s for socket path /run/containerd/containerd.sock
I0127 02:42:04.582303 3734956 ssh_runner.go:195] Run: stat /run/containerd/containerd.sock
I0127 02:42:04.585908 3734956 start.go:563] Will wait 60s for crictl version
I0127 02:42:04.585972 3734956 ssh_runner.go:195] Run: which crictl
I0127 02:42:04.589268 3734956 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0127 02:42:04.625645 3734956 start.go:579] Version: 0.1.0
RuntimeName: containerd
RuntimeVersion: 1.7.24
RuntimeApiVersion: v1
I0127 02:42:04.625704 3734956 ssh_runner.go:195] Run: containerd --version
I0127 02:42:04.647295 3734956 ssh_runner.go:195] Run: containerd --version
I0127 02:42:04.674729 3734956 out.go:177] * Preparing Kubernetes v1.32.1 on containerd 1.7.24 ...
I0127 02:42:04.680170 3734956 cli_runner.go:164] Run: docker network inspect scheduled-stop-403142 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0127 02:42:04.696520 3734956 ssh_runner.go:195] Run: grep 192.168.76.1 host.minikube.internal$ /etc/hosts
I0127 02:42:04.700097 3734956 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.76.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0127 02:42:04.711158 3734956 kubeadm.go:883] updating cluster {Name:scheduled-stop-403142 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.32.1 ClusterName:scheduled-stop-403142 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] A
PIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.32.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock
: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0127 02:42:04.711265 3734956 preload.go:131] Checking if preload exists for k8s version v1.32.1 and runtime containerd
I0127 02:42:04.711321 3734956 ssh_runner.go:195] Run: sudo crictl images --output json
I0127 02:42:04.745824 3734956 containerd.go:627] all images are preloaded for containerd runtime.
I0127 02:42:04.745836 3734956 containerd.go:534] Images already preloaded, skipping extraction
I0127 02:42:04.745896 3734956 ssh_runner.go:195] Run: sudo crictl images --output json
I0127 02:42:04.779709 3734956 containerd.go:627] all images are preloaded for containerd runtime.
I0127 02:42:04.779721 3734956 cache_images.go:84] Images are preloaded, skipping loading
I0127 02:42:04.779728 3734956 kubeadm.go:934] updating node { 192.168.76.2 8443 v1.32.1 containerd true true} ...
I0127 02:42:04.779820 3734956 kubeadm.go:946] kubelet [Unit]
Wants=containerd.service
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.32.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=scheduled-stop-403142 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.76.2
[Install]
config:
{KubernetesVersion:v1.32.1 ClusterName:scheduled-stop-403142 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0127 02:42:04.779892 3734956 ssh_runner.go:195] Run: sudo crictl info
I0127 02:42:04.819147 3734956 cni.go:84] Creating CNI manager for ""
I0127 02:42:04.819159 3734956 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I0127 02:42:04.819166 3734956 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0127 02:42:04.819190 3734956 kubeadm.go:189] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.76.2 APIServerPort:8443 KubernetesVersion:v1.32.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:scheduled-stop-403142 NodeName:scheduled-stop-403142 DNSDomain:cluster.local CRISocket:/run/containerd/containerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.76.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.76.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt St
aticPodPath:/etc/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///run/containerd/containerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0127 02:42:04.819309 3734956 kubeadm.go:195] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta4
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.76.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///run/containerd/containerd.sock
name: "scheduled-stop-403142"
kubeletExtraArgs:
- name: "node-ip"
value: "192.168.76.2"
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta4
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.76.2"]
extraArgs:
- name: "enable-admission-plugins"
value: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
- name: "allocate-node-cidrs"
value: "true"
- name: "leader-elect"
value: "false"
scheduler:
extraArgs:
- name: "leader-elect"
value: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
- name: "proxy-refresh-interval"
value: "70000"
kubernetesVersion: v1.32.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///run/containerd/containerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0127 02:42:04.819382 3734956 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.32.1
I0127 02:42:04.828236 3734956 binaries.go:44] Found k8s binaries, skipping transfer
I0127 02:42:04.828302 3734956 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0127 02:42:04.836998 3734956 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (325 bytes)
I0127 02:42:04.855156 3734956 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0127 02:42:04.873482 3734956 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2311 bytes)
I0127 02:42:04.892216 3734956 ssh_runner.go:195] Run: grep 192.168.76.2 control-plane.minikube.internal$ /etc/hosts
I0127 02:42:04.895722 3734956 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.76.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0127 02:42:04.906608 3734956 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0127 02:42:04.995593 3734956 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0127 02:42:05.012262 3734956 certs.go:68] Setting up /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142 for IP: 192.168.76.2
I0127 02:42:05.012278 3734956 certs.go:194] generating shared ca certs ...
I0127 02:42:05.012298 3734956 certs.go:226] acquiring lock for ca certs: {Name:mk1bae14ef6af74439063c8478bc03213541b880 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 02:42:05.012452 3734956 certs.go:235] skipping valid "minikubeCA" ca cert: /home/jenkins/minikube-integration/20316-3581420/.minikube/ca.key
I0127 02:42:05.012494 3734956 certs.go:235] skipping valid "proxyClientCA" ca cert: /home/jenkins/minikube-integration/20316-3581420/.minikube/proxy-client-ca.key
I0127 02:42:05.012500 3734956 certs.go:256] generating profile certs ...
I0127 02:42:05.012557 3734956 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/client.key
I0127 02:42:05.012578 3734956 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/client.crt with IP's: []
I0127 02:42:05.311072 3734956 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/client.crt ...
I0127 02:42:05.311088 3734956 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/client.crt: {Name:mk758880b9f27137c5825329bacf49448c5cad41 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 02:42:05.311295 3734956 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/client.key ...
I0127 02:42:05.311304 3734956 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/client.key: {Name:mk879725613c34bf0afdb4f6b5fb32b1571d8bc3 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 02:42:05.311399 3734956 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/apiserver.key.017333d5
I0127 02:42:05.311414 3734956 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/apiserver.crt.017333d5 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.76.2]
I0127 02:42:05.663186 3734956 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/apiserver.crt.017333d5 ...
I0127 02:42:05.663207 3734956 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/apiserver.crt.017333d5: {Name:mk4ddb281a24b78e4c524508551aade5f3117495 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 02:42:05.663399 3734956 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/apiserver.key.017333d5 ...
I0127 02:42:05.663408 3734956 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/apiserver.key.017333d5: {Name:mkf5223095261c2104865219194fdb2ac4add9c7 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 02:42:05.663487 3734956 certs.go:381] copying /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/apiserver.crt.017333d5 -> /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/apiserver.crt
I0127 02:42:05.663559 3734956 certs.go:385] copying /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/apiserver.key.017333d5 -> /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/apiserver.key
I0127 02:42:05.663608 3734956 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/proxy-client.key
I0127 02:42:05.663620 3734956 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/proxy-client.crt with IP's: []
I0127 02:42:06.082450 3734956 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/proxy-client.crt ...
I0127 02:42:06.082465 3734956 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/proxy-client.crt: {Name:mk367cb4694c7a90f6f0ab861c6ee8f7454d4167 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 02:42:06.082653 3734956 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/proxy-client.key ...
I0127 02:42:06.082661 3734956 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/proxy-client.key: {Name:mkc646a039146abb6bb4b51e8763ecab3439ae4a Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 02:42:06.082871 3734956 certs.go:484] found cert: /home/jenkins/minikube-integration/20316-3581420/.minikube/certs/3586800.pem (1338 bytes)
W0127 02:42:06.082906 3734956 certs.go:480] ignoring /home/jenkins/minikube-integration/20316-3581420/.minikube/certs/3586800_empty.pem, impossibly tiny 0 bytes
I0127 02:42:06.082913 3734956 certs.go:484] found cert: /home/jenkins/minikube-integration/20316-3581420/.minikube/certs/ca-key.pem (1675 bytes)
I0127 02:42:06.082939 3734956 certs.go:484] found cert: /home/jenkins/minikube-integration/20316-3581420/.minikube/certs/ca.pem (1078 bytes)
I0127 02:42:06.082961 3734956 certs.go:484] found cert: /home/jenkins/minikube-integration/20316-3581420/.minikube/certs/cert.pem (1123 bytes)
I0127 02:42:06.082983 3734956 certs.go:484] found cert: /home/jenkins/minikube-integration/20316-3581420/.minikube/certs/key.pem (1679 bytes)
I0127 02:42:06.083025 3734956 certs.go:484] found cert: /home/jenkins/minikube-integration/20316-3581420/.minikube/files/etc/ssl/certs/35868002.pem (1708 bytes)
I0127 02:42:06.083647 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0127 02:42:06.108966 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I0127 02:42:06.133394 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0127 02:42:06.157386 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I0127 02:42:06.181005 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1432 bytes)
I0127 02:42:06.205492 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0127 02:42:06.229240 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0127 02:42:06.252121 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/profiles/scheduled-stop-403142/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
I0127 02:42:06.275707 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/files/etc/ssl/certs/35868002.pem --> /usr/share/ca-certificates/35868002.pem (1708 bytes)
I0127 02:42:06.300687 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0127 02:42:06.324800 3734956 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20316-3581420/.minikube/certs/3586800.pem --> /usr/share/ca-certificates/3586800.pem (1338 bytes)
I0127 02:42:06.348835 3734956 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0127 02:42:06.366508 3734956 ssh_runner.go:195] Run: openssl version
I0127 02:42:06.372043 3734956 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/35868002.pem && ln -fs /usr/share/ca-certificates/35868002.pem /etc/ssl/certs/35868002.pem"
I0127 02:42:06.381488 3734956 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/35868002.pem
I0127 02:42:06.385148 3734956 certs.go:528] hashing: -rw-r--r-- 1 root root 1708 Jan 27 02:16 /usr/share/ca-certificates/35868002.pem
I0127 02:42:06.385204 3734956 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/35868002.pem
I0127 02:42:06.392107 3734956 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/3ec20f2e.0 || ln -fs /etc/ssl/certs/35868002.pem /etc/ssl/certs/3ec20f2e.0"
I0127 02:42:06.401880 3734956 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0127 02:42:06.411447 3734956 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0127 02:42:06.415062 3734956 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Jan 27 02:08 /usr/share/ca-certificates/minikubeCA.pem
I0127 02:42:06.415131 3734956 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0127 02:42:06.422033 3734956 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0127 02:42:06.431508 3734956 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/3586800.pem && ln -fs /usr/share/ca-certificates/3586800.pem /etc/ssl/certs/3586800.pem"
I0127 02:42:06.440704 3734956 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/3586800.pem
I0127 02:42:06.444205 3734956 certs.go:528] hashing: -rw-r--r-- 1 root root 1338 Jan 27 02:16 /usr/share/ca-certificates/3586800.pem
I0127 02:42:06.444263 3734956 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/3586800.pem
I0127 02:42:06.451258 3734956 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/51391683.0 || ln -fs /etc/ssl/certs/3586800.pem /etc/ssl/certs/51391683.0"
I0127 02:42:06.460514 3734956 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0127 02:42:06.463730 3734956 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0127 02:42:06.463772 3734956 kubeadm.go:392] StartCluster: {Name:scheduled-stop-403142 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.46@sha256:fd2d445ddcc33ebc5c6b68a17e6219ea207ce63c005095ea1525296da2d1a279 Memory:2048 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.32.1 ClusterName:scheduled-stop-403142 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIS
erverIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.32.1 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: S
SHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0127 02:42:06.463846 3734956 cri.go:54] listing CRI containers in root /run/containerd/runc/k8s.io: {State:paused Name: Namespaces:[kube-system]}
I0127 02:42:06.463899 3734956 ssh_runner.go:195] Run: sudo -s eval "crictl ps -a --quiet --label io.kubernetes.pod.namespace=kube-system"
I0127 02:42:06.500185 3734956 cri.go:89] found id: ""
I0127 02:42:06.500250 3734956 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0127 02:42:06.509364 3734956 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0127 02:42:06.518369 3734956 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0127 02:42:06.518434 3734956 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0127 02:42:06.528373 3734956 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0127 02:42:06.528392 3734956 kubeadm.go:157] found existing configuration files:
I0127 02:42:06.528445 3734956 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0127 02:42:06.537161 3734956 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0127 02:42:06.537216 3734956 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0127 02:42:06.545446 3734956 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0127 02:42:06.554189 3734956 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0127 02:42:06.554247 3734956 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0127 02:42:06.563080 3734956 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0127 02:42:06.572096 3734956 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0127 02:42:06.572156 3734956 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0127 02:42:06.580961 3734956 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0127 02:42:06.591494 3734956 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0127 02:42:06.591550 3734956 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0127 02:42:06.601325 3734956 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.32.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0127 02:42:06.648428 3734956 kubeadm.go:310] [init] Using Kubernetes version: v1.32.1
I0127 02:42:06.648479 3734956 kubeadm.go:310] [preflight] Running pre-flight checks
I0127 02:42:06.671520 3734956 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0127 02:42:06.671587 3734956 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1075-aws[0m
I0127 02:42:06.671621 3734956 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0127 02:42:06.671666 3734956 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0127 02:42:06.671714 3734956 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0127 02:42:06.671760 3734956 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0127 02:42:06.671808 3734956 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0127 02:42:06.671856 3734956 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0127 02:42:06.671903 3734956 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0127 02:42:06.671947 3734956 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0127 02:42:06.672008 3734956 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0127 02:42:06.672062 3734956 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0127 02:42:06.734408 3734956 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0127 02:42:06.734513 3734956 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0127 02:42:06.734604 3734956 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0127 02:42:06.746024 3734956 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0127 02:42:06.752361 3734956 out.go:235] - Generating certificates and keys ...
I0127 02:42:06.752536 3734956 kubeadm.go:310] [certs] Using existing ca certificate authority
I0127 02:42:06.752606 3734956 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0127 02:42:06.971360 3734956 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0127 02:42:07.194436 3734956 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0127 02:42:08.143305 3734956 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0127 02:42:08.813266 3734956 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0127 02:42:09.288011 3734956 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0127 02:42:09.288291 3734956 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [localhost scheduled-stop-403142] and IPs [192.168.76.2 127.0.0.1 ::1]
I0127 02:42:09.603835 3734956 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0127 02:42:09.604106 3734956 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [localhost scheduled-stop-403142] and IPs [192.168.76.2 127.0.0.1 ::1]
I0127 02:42:10.445654 3734956 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0127 02:42:11.208312 3734956 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0127 02:42:11.949736 3734956 kubeadm.go:310] [certs] Generating "sa" key and public key
I0127 02:42:11.949802 3734956 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0127 02:42:12.379384 3734956 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0127 02:42:13.177641 3734956 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0127 02:42:13.809051 3734956 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0127 02:42:14.092681 3734956 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0127 02:42:14.420653 3734956 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0127 02:42:14.421245 3734956 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0127 02:42:14.424139 3734956 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0127 02:42:14.427713 3734956 out.go:235] - Booting up control plane ...
I0127 02:42:14.427829 3734956 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0127 02:42:14.427904 3734956 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0127 02:42:14.427971 3734956 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0127 02:42:14.437630 3734956 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0127 02:42:14.443881 3734956 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0127 02:42:14.443929 3734956 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0127 02:42:14.550575 3734956 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0127 02:42:14.550693 3734956 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0127 02:42:16.543064 3734956 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 2.00095755s
I0127 02:42:16.543164 3734956 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0127 02:42:23.044893 3734956 kubeadm.go:310] [api-check] The API server is healthy after 6.501864409s
I0127 02:42:23.068902 3734956 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0127 02:42:23.091744 3734956 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0127 02:42:23.123632 3734956 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0127 02:42:23.123852 3734956 kubeadm.go:310] [mark-control-plane] Marking the node scheduled-stop-403142 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0127 02:42:23.135477 3734956 kubeadm.go:310] [bootstrap-token] Using token: qta9rx.je3htja9l90ga7qv
I0127 02:42:23.138388 3734956 out.go:235] - Configuring RBAC rules ...
I0127 02:42:23.138525 3734956 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0127 02:42:23.143559 3734956 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0127 02:42:23.155266 3734956 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0127 02:42:23.159658 3734956 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0127 02:42:23.164380 3734956 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0127 02:42:23.171458 3734956 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0127 02:42:23.452636 3734956 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0127 02:42:23.887839 3734956 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0127 02:42:24.452237 3734956 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0127 02:42:24.453414 3734956 kubeadm.go:310]
I0127 02:42:24.453481 3734956 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0127 02:42:24.453486 3734956 kubeadm.go:310]
I0127 02:42:24.453562 3734956 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0127 02:42:24.453566 3734956 kubeadm.go:310]
I0127 02:42:24.453590 3734956 kubeadm.go:310] mkdir -p $HOME/.kube
I0127 02:42:24.453648 3734956 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0127 02:42:24.453698 3734956 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0127 02:42:24.453701 3734956 kubeadm.go:310]
I0127 02:42:24.453754 3734956 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0127 02:42:24.453758 3734956 kubeadm.go:310]
I0127 02:42:24.453804 3734956 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0127 02:42:24.453807 3734956 kubeadm.go:310]
I0127 02:42:24.453858 3734956 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0127 02:42:24.453933 3734956 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0127 02:42:24.454016 3734956 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0127 02:42:24.454020 3734956 kubeadm.go:310]
I0127 02:42:24.454153 3734956 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0127 02:42:24.454243 3734956 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0127 02:42:24.454248 3734956 kubeadm.go:310]
I0127 02:42:24.454330 3734956 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token qta9rx.je3htja9l90ga7qv \
I0127 02:42:24.454445 3734956 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:83891a1b2b837c79fabbfd6fe62cd9786dc4221059a44014b5acb94babe950cd \
I0127 02:42:24.454465 3734956 kubeadm.go:310] --control-plane
I0127 02:42:24.454468 3734956 kubeadm.go:310]
I0127 02:42:24.454552 3734956 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0127 02:42:24.454555 3734956 kubeadm.go:310]
I0127 02:42:24.454642 3734956 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token qta9rx.je3htja9l90ga7qv \
I0127 02:42:24.454748 3734956 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:83891a1b2b837c79fabbfd6fe62cd9786dc4221059a44014b5acb94babe950cd
I0127 02:42:24.459506 3734956 kubeadm.go:310] [WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2
I0127 02:42:24.459781 3734956 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1075-aws\n", err: exit status 1
I0127 02:42:24.459905 3734956 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0127 02:42:24.459965 3734956 cni.go:84] Creating CNI manager for ""
I0127 02:42:24.459974 3734956 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I0127 02:42:24.465005 3734956 out.go:177] * Configuring CNI (Container Networking Interface) ...
I0127 02:42:24.467894 3734956 ssh_runner.go:195] Run: stat /opt/cni/bin/portmap
I0127 02:42:24.471890 3734956 cni.go:182] applying CNI manifest using /var/lib/minikube/binaries/v1.32.1/kubectl ...
I0127 02:42:24.471901 3734956 ssh_runner.go:362] scp memory --> /var/tmp/minikube/cni.yaml (2601 bytes)
I0127 02:42:24.491615 3734956 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.32.1/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml
I0127 02:42:24.784055 3734956 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0127 02:42:24.784183 3734956 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.32.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0127 02:42:24.784260 3734956 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.32.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes scheduled-stop-403142 minikube.k8s.io/updated_at=2025_01_27T02_42_24_0700 minikube.k8s.io/version=v1.35.0 minikube.k8s.io/commit=6bb462d349d93b9bf1c5a4f87817e5e9ea11cc95 minikube.k8s.io/name=scheduled-stop-403142 minikube.k8s.io/primary=true
I0127 02:42:24.975276 3734956 ops.go:34] apiserver oom_adj: -16
I0127 02:42:24.975299 3734956 kubeadm.go:1113] duration metric: took 191.170998ms to wait for elevateKubeSystemPrivileges
I0127 02:42:24.975336 3734956 kubeadm.go:394] duration metric: took 18.511566323s to StartCluster
I0127 02:42:24.975354 3734956 settings.go:142] acquiring lock: {Name:mk735c76882f337c2ca62b3dd2d1bbcced4c92cb Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 02:42:24.975439 3734956 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/20316-3581420/kubeconfig
I0127 02:42:24.976116 3734956 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20316-3581420/kubeconfig: {Name:mkc8ad8c78feebc7c27d31aea066c6fc5e1767bd Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0127 02:42:24.976340 3734956 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.32.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0127 02:42:24.976343 3734956 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.76.2 Port:8443 KubernetesVersion:v1.32.1 ContainerRuntime:containerd ControlPlane:true Worker:true}
I0127 02:42:24.976620 3734956 config.go:182] Loaded profile config "scheduled-stop-403142": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.32.1
I0127 02:42:24.976676 3734956 addons.go:511] enable addons start: toEnable=map[ambassador:false amd-gpu-device-plugin:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:false volcano:false volumesnapshots:false yakd:false]
I0127 02:42:24.976754 3734956 addons.go:69] Setting storage-provisioner=true in profile "scheduled-stop-403142"
I0127 02:42:24.976772 3734956 addons.go:238] Setting addon storage-provisioner=true in "scheduled-stop-403142"
I0127 02:42:24.976794 3734956 host.go:66] Checking if "scheduled-stop-403142" exists ...
I0127 02:42:24.976822 3734956 addons.go:69] Setting default-storageclass=true in profile "scheduled-stop-403142"
I0127 02:42:24.976836 3734956 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "scheduled-stop-403142"
I0127 02:42:24.977154 3734956 cli_runner.go:164] Run: docker container inspect scheduled-stop-403142 --format={{.State.Status}}
I0127 02:42:24.977348 3734956 cli_runner.go:164] Run: docker container inspect scheduled-stop-403142 --format={{.State.Status}}
I0127 02:42:24.979529 3734956 out.go:177] * Verifying Kubernetes components...
I0127 02:42:24.986218 3734956 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0127 02:42:25.016047 3734956 addons.go:238] Setting addon default-storageclass=true in "scheduled-stop-403142"
I0127 02:42:25.016080 3734956 host.go:66] Checking if "scheduled-stop-403142" exists ...
I0127 02:42:25.017253 3734956 cli_runner.go:164] Run: docker container inspect scheduled-stop-403142 --format={{.State.Status}}
I0127 02:42:25.025409 3734956 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0127 02:42:25.028462 3734956 addons.go:435] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0127 02:42:25.028476 3734956 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0127 02:42:25.028550 3734956 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-403142
I0127 02:42:25.055991 3734956 addons.go:435] installing /etc/kubernetes/addons/storageclass.yaml
I0127 02:42:25.056005 3734956 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0127 02:42:25.056085 3734956 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" scheduled-stop-403142
I0127 02:42:25.075120 3734956 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:37686 SSHKeyPath:/home/jenkins/minikube-integration/20316-3581420/.minikube/machines/scheduled-stop-403142/id_rsa Username:docker}
I0127 02:42:25.103561 3734956 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:37686 SSHKeyPath:/home/jenkins/minikube-integration/20316-3581420/.minikube/machines/scheduled-stop-403142/id_rsa Username:docker}
I0127 02:42:25.228061 3734956 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.32.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.76.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.32.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0127 02:42:25.228164 3734956 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0127 02:42:25.332141 3734956 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.32.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0127 02:42:25.333416 3734956 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.32.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0127 02:42:25.588091 3734956 api_server.go:52] waiting for apiserver process to appear ...
I0127 02:42:25.588142 3734956 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0127 02:42:25.588248 3734956 start.go:971] {"host.minikube.internal": 192.168.76.1} host record injected into CoreDNS's ConfigMap
I0127 02:42:25.872065 3734956 api_server.go:72] duration metric: took 895.699269ms to wait for apiserver process to appear ...
I0127 02:42:25.872074 3734956 api_server.go:88] waiting for apiserver healthz status ...
I0127 02:42:25.872090 3734956 api_server.go:253] Checking apiserver healthz at https://192.168.76.2:8443/healthz ...
I0127 02:42:25.882574 3734956 api_server.go:279] https://192.168.76.2:8443/healthz returned 200:
ok
I0127 02:42:25.883790 3734956 api_server.go:141] control plane version: v1.32.1
I0127 02:42:25.883805 3734956 api_server.go:131] duration metric: took 11.725518ms to wait for apiserver health ...
I0127 02:42:25.883811 3734956 system_pods.go:43] waiting for kube-system pods to appear ...
I0127 02:42:25.890151 3734956 system_pods.go:59] 5 kube-system pods found
I0127 02:42:25.890162 3734956 out.go:177] * Enabled addons: storage-provisioner, default-storageclass
I0127 02:42:25.890175 3734956 system_pods.go:61] "etcd-scheduled-stop-403142" [402ba61e-1886-4358-a49a-8475e6884554] Running / Ready:ContainersNotReady (containers with unready status: [etcd]) / ContainersReady:ContainersNotReady (containers with unready status: [etcd])
I0127 02:42:25.890182 3734956 system_pods.go:61] "kube-apiserver-scheduled-stop-403142" [b0c16e5c-c202-48c4-a5c6-0218b6873bbf] Running / Ready:ContainersNotReady (containers with unready status: [kube-apiserver]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-apiserver])
I0127 02:42:25.890190 3734956 system_pods.go:61] "kube-controller-manager-scheduled-stop-403142" [77ef8281-a875-4b3c-b3ec-0e261121189f] Running / Ready:ContainersNotReady (containers with unready status: [kube-controller-manager]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-controller-manager])
I0127 02:42:25.890197 3734956 system_pods.go:61] "kube-scheduler-scheduled-stop-403142" [2d2f4811-7ad1-469c-b111-d59f62e7cd06] Running / Ready:ContainersNotReady (containers with unready status: [kube-scheduler]) / ContainersReady:ContainersNotReady (containers with unready status: [kube-scheduler])
I0127 02:42:25.890201 3734956 system_pods.go:61] "storage-provisioner" [029f640c-de68-47fd-96dd-88130384396c] Pending: PodScheduled:Unschedulable (0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.)
I0127 02:42:25.890212 3734956 system_pods.go:74] duration metric: took 6.38978ms to wait for pod list to return data ...
I0127 02:42:25.890223 3734956 kubeadm.go:582] duration metric: took 913.860883ms to wait for: map[apiserver:true system_pods:true]
I0127 02:42:25.890287 3734956 node_conditions.go:102] verifying NodePressure condition ...
I0127 02:42:25.893530 3734956 node_conditions.go:122] node storage ephemeral capacity is 203034800Ki
I0127 02:42:25.893547 3734956 node_conditions.go:123] node cpu capacity is 2
I0127 02:42:25.893557 3734956 node_conditions.go:105] duration metric: took 3.265592ms to run NodePressure ...
I0127 02:42:25.893567 3734956 start.go:241] waiting for startup goroutines ...
I0127 02:42:25.893577 3734956 addons.go:514] duration metric: took 916.905329ms for enable addons: enabled=[storage-provisioner default-storageclass]
I0127 02:42:26.092439 3734956 kapi.go:214] "coredns" deployment in "kube-system" namespace and "scheduled-stop-403142" context rescaled to 1 replicas
I0127 02:42:26.092462 3734956 start.go:246] waiting for cluster config update ...
I0127 02:42:26.092472 3734956 start.go:255] writing updated cluster config ...
I0127 02:42:26.092761 3734956 ssh_runner.go:195] Run: rm -f paused
I0127 02:42:26.163306 3734956 start.go:600] kubectl: 1.32.1, cluster: 1.32.1 (minor skew: 0)
I0127 02:42:26.166553 3734956 out.go:177] * Done! kubectl is now configured to use "scheduled-stop-403142" cluster and "default" namespace by default
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
a44fa8c758070 7fc9d4aa817aa 11 seconds ago Running etcd 0 535344203a34f etcd-scheduled-stop-403142
c0bfa19c4f045 265c2dedf28ab 11 seconds ago Running kube-apiserver 0 744b50547c1f7 kube-apiserver-scheduled-stop-403142
f849e424becdd ddb38cac617cb 11 seconds ago Running kube-scheduler 0 2b92de8c36c84 kube-scheduler-scheduled-stop-403142
578769dc997b6 2933761aa7ada 11 seconds ago Running kube-controller-manager 0 9162c2291407b kube-controller-manager-scheduled-stop-403142
==> containerd <==
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.715142417Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.724002217Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.724169621Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.724204451Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.724442762Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.801181947Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-scheduled-stop-403142,Uid:1f780fd35d1d7676ed22895cddf3a275,Namespace:kube-system,Attempt:0,} returns sandbox id \"9162c2291407b6bc71814b1dc949db1f44a8954e877d19a9fb01c10160d44cda\""
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.818595135Z" level=info msg="CreateContainer within sandbox \"9162c2291407b6bc71814b1dc949db1f44a8954e877d19a9fb01c10160d44cda\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}"
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.832172958Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-scheduled-stop-403142,Uid:3568a00adda03d4c146e1649e4f29ad0,Namespace:kube-system,Attempt:0,} returns sandbox id \"744b50547c1f7b8f49e4cf2f035b0135e1d78312345520451b2c630de8131e54\""
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.832520961Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-scheduled-stop-403142,Uid:db2cc2783bcbf55452c6c1530655d8b0,Namespace:kube-system,Attempt:0,} returns sandbox id \"2b92de8c36c84677c4dd531bfe0cfdd01ef657b2e81676db11e92fd67531d9da\""
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.842300399Z" level=info msg="CreateContainer within sandbox \"744b50547c1f7b8f49e4cf2f035b0135e1d78312345520451b2c630de8131e54\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}"
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.844780748Z" level=info msg="CreateContainer within sandbox \"2b92de8c36c84677c4dd531bfe0cfdd01ef657b2e81676db11e92fd67531d9da\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}"
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.849404259Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:etcd-scheduled-stop-403142,Uid:39bcb589b15eebd4a33b2c34b9e7c266,Namespace:kube-system,Attempt:0,} returns sandbox id \"535344203a34f01855afbed02c886a93ba26e4787707b14288e2041c48376d31\""
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.852117021Z" level=info msg="CreateContainer within sandbox \"535344203a34f01855afbed02c886a93ba26e4787707b14288e2041c48376d31\" for container &ContainerMetadata{Name:etcd,Attempt:0,}"
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.866488923Z" level=info msg="CreateContainer within sandbox \"9162c2291407b6bc71814b1dc949db1f44a8954e877d19a9fb01c10160d44cda\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"578769dc997b65c220226475f737b39eb343eb078715c90bdf342d323fbc7154\""
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.867469819Z" level=info msg="StartContainer for \"578769dc997b65c220226475f737b39eb343eb078715c90bdf342d323fbc7154\""
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.884265260Z" level=info msg="CreateContainer within sandbox \"2b92de8c36c84677c4dd531bfe0cfdd01ef657b2e81676db11e92fd67531d9da\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"f849e424becdd41a7dd850de9d77d34397cf4bec89568a4deba210a9cd74df63\""
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.885081934Z" level=info msg="StartContainer for \"f849e424becdd41a7dd850de9d77d34397cf4bec89568a4deba210a9cd74df63\""
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.887572130Z" level=info msg="CreateContainer within sandbox \"744b50547c1f7b8f49e4cf2f035b0135e1d78312345520451b2c630de8131e54\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"c0bfa19c4f0453ba29cb932bd5aeb70ce08db42774743283f95a3087d7536aef\""
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.888998347Z" level=info msg="StartContainer for \"c0bfa19c4f0453ba29cb932bd5aeb70ce08db42774743283f95a3087d7536aef\""
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.919641413Z" level=info msg="CreateContainer within sandbox \"535344203a34f01855afbed02c886a93ba26e4787707b14288e2041c48376d31\" for &ContainerMetadata{Name:etcd,Attempt:0,} returns container id \"a44fa8c758070cb77953e752b550835540f36e6dbeee0f7955466f54260280bf\""
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.927948489Z" level=info msg="StartContainer for \"a44fa8c758070cb77953e752b550835540f36e6dbeee0f7955466f54260280bf\""
Jan 27 02:42:16 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:16.982361605Z" level=info msg="StartContainer for \"578769dc997b65c220226475f737b39eb343eb078715c90bdf342d323fbc7154\" returns successfully"
Jan 27 02:42:17 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:17.052658268Z" level=info msg="StartContainer for \"c0bfa19c4f0453ba29cb932bd5aeb70ce08db42774743283f95a3087d7536aef\" returns successfully"
Jan 27 02:42:17 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:17.083769661Z" level=info msg="StartContainer for \"f849e424becdd41a7dd850de9d77d34397cf4bec89568a4deba210a9cd74df63\" returns successfully"
Jan 27 02:42:17 scheduled-stop-403142 containerd[830]: time="2025-01-27T02:42:17.176056207Z" level=info msg="StartContainer for \"a44fa8c758070cb77953e752b550835540f36e6dbeee0f7955466f54260280bf\" returns successfully"
==> describe nodes <==
Name: scheduled-stop-403142
Roles: control-plane
Labels: beta.kubernetes.io/arch=arm64
beta.kubernetes.io/os=linux
kubernetes.io/arch=arm64
kubernetes.io/hostname=scheduled-stop-403142
kubernetes.io/os=linux
minikube.k8s.io/commit=6bb462d349d93b9bf1c5a4f87817e5e9ea11cc95
minikube.k8s.io/name=scheduled-stop-403142
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2025_01_27T02_42_24_0700
minikube.k8s.io/version=v1.35.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///run/containerd/containerd.sock
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Mon, 27 Jan 2025 02:42:20 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: scheduled-stop-403142
AcquireTime: <unset>
RenewTime: Mon, 27 Jan 2025 02:42:23 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Mon, 27 Jan 2025 02:42:21 +0000 Mon, 27 Jan 2025 02:42:17 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Mon, 27 Jan 2025 02:42:21 +0000 Mon, 27 Jan 2025 02:42:17 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Mon, 27 Jan 2025 02:42:21 +0000 Mon, 27 Jan 2025 02:42:17 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Mon, 27 Jan 2025 02:42:21 +0000 Mon, 27 Jan 2025 02:42:21 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.76.2
Hostname: scheduled-stop-403142
Capacity:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022300Ki
pods: 110
Allocatable:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022300Ki
pods: 110
System Info:
Machine ID: c2e3a3fe730a4a1382563dc958dd4700
System UUID: c9f2a712-e800-42ec-968f-c1f87c3a5dc4
Boot ID: ed5e2339-9d7b-4ad8-ab13-7fed1ac53390
Kernel Version: 5.15.0-1075-aws
OS Image: Ubuntu 22.04.5 LTS
Operating System: linux
Architecture: arm64
Container Runtime Version: containerd://1.7.24
Kubelet Version: v1.32.1
Kube-Proxy Version: v1.32.1
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (5 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
kube-system etcd-scheduled-stop-403142 100m (5%) 0 (0%) 100Mi (1%) 0 (0%) 4s
kube-system kube-apiserver-scheduled-stop-403142 250m (12%) 0 (0%) 0 (0%) 0 (0%) 6s
kube-system kube-controller-manager-scheduled-stop-403142 200m (10%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system kube-scheduler-scheduled-stop-403142 100m (5%) 0 (0%) 0 (0%) 0 (0%) 4s
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 3s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 650m (32%) 0 (0%)
memory 100Mi (1%) 0 (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
hugepages-32Mi 0 (0%) 0 (0%)
hugepages-64Ki 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal NodeHasSufficientMemory 12s (x8 over 12s) kubelet Node scheduled-stop-403142 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12s (x8 over 12s) kubelet Node scheduled-stop-403142 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12s (x7 over 12s) kubelet Node scheduled-stop-403142 status is now: NodeHasSufficientPID
Normal NodeAllocatableEnforced 12s kubelet Updated Node Allocatable limit across pods
Normal Starting 5s kubelet Starting kubelet.
Warning CgroupV1 5s kubelet cgroup v1 support is in maintenance mode, please migrate to cgroup v2
Normal NodeAllocatableEnforced 5s kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 4s kubelet Node scheduled-stop-403142 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 4s kubelet Node scheduled-stop-403142 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 4s kubelet Node scheduled-stop-403142 status is now: NodeHasSufficientPID
Normal RegisteredNode 1s node-controller Node scheduled-stop-403142 event: Registered Node scheduled-stop-403142 in Controller
==> dmesg <==
[Jan27 01:33] systemd-journald[221]: Failed to send WATCHDOG=1 notification message: Connection refused
[Jan27 01:42] overlayfs: failed to resolve '/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs/snapshots/28/fs': -2
==> etcd [a44fa8c758070cb77953e752b550835540f36e6dbeee0f7955466f54260280bf] <==
{"level":"info","ts":"2025-01-27T02:42:17.294061Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 switched to configuration voters=(16896983918768216326)"}
{"level":"info","ts":"2025-01-27T02:42:17.294977Z","caller":"membership/cluster.go:421","msg":"added member","cluster-id":"6f20f2c4b2fb5f8a","local-member-id":"ea7e25599daad906","added-peer-id":"ea7e25599daad906","added-peer-peer-urls":["https://192.168.76.2:2380"]}
{"level":"info","ts":"2025-01-27T02:42:17.294466Z","caller":"embed/etcd.go:871","msg":"serving metrics","address":"http://127.0.0.1:2381"}
{"level":"info","ts":"2025-01-27T02:42:17.295287Z","caller":"embed/etcd.go:600","msg":"serving peer traffic","address":"192.168.76.2:2380"}
{"level":"info","ts":"2025-01-27T02:42:17.295387Z","caller":"embed/etcd.go:572","msg":"cmux::serve","address":"192.168.76.2:2380"}
{"level":"info","ts":"2025-01-27T02:42:17.446135Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 is starting a new election at term 1"}
{"level":"info","ts":"2025-01-27T02:42:17.446350Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 became pre-candidate at term 1"}
{"level":"info","ts":"2025-01-27T02:42:17.446503Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 received MsgPreVoteResp from ea7e25599daad906 at term 1"}
{"level":"info","ts":"2025-01-27T02:42:17.446586Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 became candidate at term 2"}
{"level":"info","ts":"2025-01-27T02:42:17.446669Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 received MsgVoteResp from ea7e25599daad906 at term 2"}
{"level":"info","ts":"2025-01-27T02:42:17.446753Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"ea7e25599daad906 became leader at term 2"}
{"level":"info","ts":"2025-01-27T02:42:17.446840Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: ea7e25599daad906 elected leader ea7e25599daad906 at term 2"}
{"level":"info","ts":"2025-01-27T02:42:17.450255Z","caller":"etcdserver/server.go:2651","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2025-01-27T02:42:17.454371Z","caller":"etcdserver/server.go:2140","msg":"published local member to cluster through raft","local-member-id":"ea7e25599daad906","local-member-attributes":"{Name:scheduled-stop-403142 ClientURLs:[https://192.168.76.2:2379]}","request-path":"/0/members/ea7e25599daad906/attributes","cluster-id":"6f20f2c4b2fb5f8a","publish-timeout":"7s"}
{"level":"info","ts":"2025-01-27T02:42:17.454551Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2025-01-27T02:42:17.455129Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2025-01-27T02:42:17.456142Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2025-01-27T02:42:17.457141Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2025-01-27T02:42:17.457392Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"6f20f2c4b2fb5f8a","local-member-id":"ea7e25599daad906","cluster-version":"3.5"}
{"level":"info","ts":"2025-01-27T02:42:17.457630Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2025-01-27T02:42:17.457737Z","caller":"etcdserver/server.go:2675","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2025-01-27T02:42:17.458564Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2025-01-27T02:42:17.459237Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.76.2:2379"}
{"level":"info","ts":"2025-01-27T02:42:17.459387Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2025-01-27T02:42:17.470425Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
==> kernel <==
02:42:28 up 1 day, 1:24, 0 users, load average: 2.58, 2.33, 2.51
Linux scheduled-stop-403142 5.15.0-1075-aws #82~20.04.1-Ubuntu SMP Thu Dec 19 05:23:06 UTC 2024 aarch64 aarch64 aarch64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.5 LTS"
==> kube-apiserver [c0bfa19c4f0453ba29cb932bd5aeb70ce08db42774743283f95a3087d7536aef] <==
I0127 02:42:20.970667 1 cache.go:32] Waiting for caches to sync for autoregister controller
I0127 02:42:20.970756 1 cache.go:39] Caches are synced for autoregister controller
I0127 02:42:21.016160 1 shared_informer.go:320] Caches are synced for node_authorizer
I0127 02:42:21.021918 1 shared_informer.go:320] Caches are synced for cluster_authentication_trust_controller
E0127 02:42:21.050191 1 controller.go:145] "Failed to ensure lease exists, will retry" err="namespaces \"kube-system\" not found" interval="200ms"
I0127 02:42:21.052145 1 shared_informer.go:320] Caches are synced for *generic.policySource[*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicy,*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicyBinding,k8s.io/apiserver/pkg/admission/plugin/policy/validating.Validator]
I0127 02:42:21.052865 1 policy_source.go:240] refreshing policies
I0127 02:42:21.082220 1 shared_informer.go:320] Caches are synced for configmaps
I0127 02:42:21.092166 1 controller.go:615] quota admission added evaluator for: namespaces
E0127 02:42:21.145833 1 controller.go:148] "Unhandled Error" err="while syncing ConfigMap \"kube-system/kube-apiserver-legacy-service-account-token-tracking\", err: namespaces \"kube-system\" not found" logger="UnhandledError"
I0127 02:42:21.268333 1 controller.go:615] quota admission added evaluator for: leases.coordination.k8s.io
I0127 02:42:21.813785 1 storage_scheduling.go:95] created PriorityClass system-node-critical with value 2000001000
I0127 02:42:21.821164 1 storage_scheduling.go:95] created PriorityClass system-cluster-critical with value 2000000000
I0127 02:42:21.822117 1 storage_scheduling.go:111] all system priority classes are created successfully or already exist.
I0127 02:42:22.593763 1 controller.go:615] quota admission added evaluator for: roles.rbac.authorization.k8s.io
I0127 02:42:22.647523 1 controller.go:615] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io
I0127 02:42:22.787865 1 alloc.go:330] "allocated clusterIPs" service="default/kubernetes" clusterIPs={"IPv4":"10.96.0.1"}
W0127 02:42:22.795371 1 lease.go:265] Resetting endpoints for master service "kubernetes" to [192.168.76.2]
I0127 02:42:22.796520 1 controller.go:615] quota admission added evaluator for: endpoints
I0127 02:42:22.801692 1 controller.go:615] quota admission added evaluator for: endpointslices.discovery.k8s.io
I0127 02:42:22.963271 1 controller.go:615] quota admission added evaluator for: serviceaccounts
I0127 02:42:23.867661 1 controller.go:615] quota admission added evaluator for: deployments.apps
I0127 02:42:23.885966 1 alloc.go:330] "allocated clusterIPs" service="kube-system/kube-dns" clusterIPs={"IPv4":"10.96.0.10"}
I0127 02:42:23.901227 1 controller.go:615] quota admission added evaluator for: daemonsets.apps
I0127 02:42:28.217835 1 controller.go:615] quota admission added evaluator for: controllerrevisions.apps
==> kube-controller-manager [578769dc997b65c220226475f737b39eb343eb078715c90bdf342d323fbc7154] <==
I0127 02:42:27.561828 1 shared_informer.go:320] Caches are synced for endpoint_slice
I0127 02:42:27.562006 1 shared_informer.go:320] Caches are synced for certificate-csrapproving
I0127 02:42:27.562705 1 shared_informer.go:320] Caches are synced for HPA
I0127 02:42:27.563100 1 shared_informer.go:320] Caches are synced for job
I0127 02:42:27.563300 1 shared_informer.go:320] Caches are synced for ClusterRoleAggregator
I0127 02:42:27.563894 1 shared_informer.go:320] Caches are synced for TTL after finished
I0127 02:42:27.563391 1 shared_informer.go:320] Caches are synced for crt configmap
I0127 02:42:27.563423 1 shared_informer.go:320] Caches are synced for deployment
I0127 02:42:27.563437 1 shared_informer.go:320] Caches are synced for ephemeral
I0127 02:42:27.563454 1 shared_informer.go:320] Caches are synced for PVC protection
I0127 02:42:27.563476 1 shared_informer.go:320] Caches are synced for ReplicationController
I0127 02:42:27.570829 1 shared_informer.go:320] Caches are synced for garbage collector
I0127 02:42:27.571328 1 garbagecollector.go:154] "Garbage collector: all resource monitors have synced" logger="garbage-collector-controller"
I0127 02:42:27.571467 1 garbagecollector.go:157] "Proceeding to collect garbage" logger="garbage-collector-controller"
I0127 02:42:27.571283 1 shared_informer.go:320] Caches are synced for node
I0127 02:42:27.572055 1 range_allocator.go:177] "Sending events to api server" logger="node-ipam-controller"
I0127 02:42:27.572198 1 range_allocator.go:183] "Starting range CIDR allocator" logger="node-ipam-controller"
I0127 02:42:27.572294 1 shared_informer.go:313] Waiting for caches to sync for cidrallocator
I0127 02:42:27.572384 1 shared_informer.go:320] Caches are synced for cidrallocator
I0127 02:42:27.594712 1 shared_informer.go:320] Caches are synced for disruption
I0127 02:42:27.594847 1 shared_informer.go:320] Caches are synced for resource quota
I0127 02:42:27.607228 1 range_allocator.go:428] "Set node PodCIDR" logger="node-ipam-controller" node="scheduled-stop-403142" podCIDRs=["10.244.0.0/24"]
I0127 02:42:27.607262 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="scheduled-stop-403142"
I0127 02:42:27.607500 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="scheduled-stop-403142"
I0127 02:42:28.174569 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="scheduled-stop-403142"
==> kube-scheduler [f849e424becdd41a7dd850de9d77d34397cf4bec89568a4deba210a9cd74df63] <==
W0127 02:42:22.053852 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0127 02:42:22.053916 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0127 02:42:22.054039 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0127 02:42:22.054122 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0127 02:42:22.054217 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0127 02:42:22.054278 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0127 02:42:22.054369 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E0127 02:42:22.054416 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0127 02:42:22.054523 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0127 02:42:22.054576 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0127 02:42:22.054649 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0127 02:42:22.054701 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0127 02:42:22.054780 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E0127 02:42:22.054833 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0127 02:42:22.055006 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
E0127 02:42:22.055060 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0127 02:42:22.055145 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0127 02:42:22.055192 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0127 02:42:22.055328 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0127 02:42:22.055393 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0127 02:42:22.055428 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0127 02:42:22.055482 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0127 02:42:22.055719 1 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0127 02:42:22.055785 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
I0127 02:42:23.143190 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Jan 27 02:42:24 scheduled-stop-403142 kubelet[1537]: I0127 02:42:24.787798 1537 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world"
Jan 27 02:42:24 scheduled-stop-403142 kubelet[1537]: I0127 02:42:24.850920 1537 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/etcd-scheduled-stop-403142"
Jan 27 02:42:24 scheduled-stop-403142 kubelet[1537]: I0127 02:42:24.851604 1537 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-scheduled-stop-403142"
Jan 27 02:42:24 scheduled-stop-403142 kubelet[1537]: E0127 02:42:24.883823 1537 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-scheduled-stop-403142\" already exists" pod="kube-system/kube-scheduler-scheduled-stop-403142"
Jan 27 02:42:24 scheduled-stop-403142 kubelet[1537]: E0127 02:42:24.891999 1537 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"etcd-scheduled-stop-403142\" already exists" pod="kube-system/etcd-scheduled-stop-403142"
Jan 27 02:42:24 scheduled-stop-403142 kubelet[1537]: I0127 02:42:24.902510 1537 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-scheduled-stop-403142" podStartSLOduration=0.902475084 podStartE2EDuration="902.475084ms" podCreationTimestamp="2025-01-27 02:42:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-27 02:42:24.8875931 +0000 UTC m=+1.220565241" watchObservedRunningTime="2025-01-27 02:42:24.902475084 +0000 UTC m=+1.235447226"
Jan 27 02:42:24 scheduled-stop-403142 kubelet[1537]: I0127 02:42:24.918346 1537 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-scheduled-stop-403142" podStartSLOduration=0.918326329 podStartE2EDuration="918.326329ms" podCreationTimestamp="2025-01-27 02:42:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-27 02:42:24.90352865 +0000 UTC m=+1.236500800" watchObservedRunningTime="2025-01-27 02:42:24.918326329 +0000 UTC m=+1.251298479"
Jan 27 02:42:24 scheduled-stop-403142 kubelet[1537]: I0127 02:42:24.940813 1537 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/etcd-scheduled-stop-403142" podStartSLOduration=0.940789265 podStartE2EDuration="940.789265ms" podCreationTimestamp="2025-01-27 02:42:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-27 02:42:24.919162499 +0000 UTC m=+1.252134641" watchObservedRunningTime="2025-01-27 02:42:24.940789265 +0000 UTC m=+1.273761415"
Jan 27 02:42:24 scheduled-stop-403142 kubelet[1537]: I0127 02:42:24.941151 1537 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-scheduled-stop-403142" podStartSLOduration=2.9411435360000002 podStartE2EDuration="2.941143536s" podCreationTimestamp="2025-01-27 02:42:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-27 02:42:24.937402134 +0000 UTC m=+1.270374275" watchObservedRunningTime="2025-01-27 02:42:24.941143536 +0000 UTC m=+1.274115678"
Jan 27 02:42:27 scheduled-stop-403142 kubelet[1537]: I0127 02:42:27.713037 1537 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/host-path/029f640c-de68-47fd-96dd-88130384396c-tmp\") pod \"storage-provisioner\" (UID: \"029f640c-de68-47fd-96dd-88130384396c\") " pod="kube-system/storage-provisioner"
Jan 27 02:42:27 scheduled-stop-403142 kubelet[1537]: I0127 02:42:27.713479 1537 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqbl2\" (UniqueName: \"kubernetes.io/projected/029f640c-de68-47fd-96dd-88130384396c-kube-api-access-zqbl2\") pod \"storage-provisioner\" (UID: \"029f640c-de68-47fd-96dd-88130384396c\") " pod="kube-system/storage-provisioner"
Jan 27 02:42:27 scheduled-stop-403142 kubelet[1537]: E0127 02:42:27.823878 1537 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found
Jan 27 02:42:27 scheduled-stop-403142 kubelet[1537]: E0127 02:42:27.823916 1537 projected.go:194] Error preparing data for projected volume kube-api-access-zqbl2 for pod kube-system/storage-provisioner: configmap "kube-root-ca.crt" not found
Jan 27 02:42:27 scheduled-stop-403142 kubelet[1537]: E0127 02:42:27.823986 1537 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/029f640c-de68-47fd-96dd-88130384396c-kube-api-access-zqbl2 podName:029f640c-de68-47fd-96dd-88130384396c nodeName:}" failed. No retries permitted until 2025-01-27 02:42:28.323960838 +0000 UTC m=+4.656932980 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-zqbl2" (UniqueName: "kubernetes.io/projected/029f640c-de68-47fd-96dd-88130384396c-kube-api-access-zqbl2") pod "storage-provisioner" (UID: "029f640c-de68-47fd-96dd-88130384396c") : configmap "kube-root-ca.crt" not found
Jan 27 02:42:28 scheduled-stop-403142 kubelet[1537]: I0127 02:42:28.420405 1537 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c3a9e3be-b238-49fc-848d-fa32ddb2cdb2-xtables-lock\") pod \"kube-proxy-t5wqc\" (UID: \"c3a9e3be-b238-49fc-848d-fa32ddb2cdb2\") " pod="kube-system/kube-proxy-t5wqc"
Jan 27 02:42:28 scheduled-stop-403142 kubelet[1537]: I0127 02:42:28.420461 1537 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/c3a9e3be-b238-49fc-848d-fa32ddb2cdb2-kube-proxy\") pod \"kube-proxy-t5wqc\" (UID: \"c3a9e3be-b238-49fc-848d-fa32ddb2cdb2\") " pod="kube-system/kube-proxy-t5wqc"
Jan 27 02:42:28 scheduled-stop-403142 kubelet[1537]: I0127 02:42:28.420482 1537 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6d8c4fd9-9ac5-4b5b-87b0-4aa8906d614e-lib-modules\") pod \"kindnet-lbl6c\" (UID: \"6d8c4fd9-9ac5-4b5b-87b0-4aa8906d614e\") " pod="kube-system/kindnet-lbl6c"
Jan 27 02:42:28 scheduled-stop-403142 kubelet[1537]: I0127 02:42:28.420522 1537 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c3a9e3be-b238-49fc-848d-fa32ddb2cdb2-lib-modules\") pod \"kube-proxy-t5wqc\" (UID: \"c3a9e3be-b238-49fc-848d-fa32ddb2cdb2\") " pod="kube-system/kube-proxy-t5wqc"
Jan 27 02:42:28 scheduled-stop-403142 kubelet[1537]: I0127 02:42:28.420543 1537 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-cfg\" (UniqueName: \"kubernetes.io/host-path/6d8c4fd9-9ac5-4b5b-87b0-4aa8906d614e-cni-cfg\") pod \"kindnet-lbl6c\" (UID: \"6d8c4fd9-9ac5-4b5b-87b0-4aa8906d614e\") " pod="kube-system/kindnet-lbl6c"
Jan 27 02:42:28 scheduled-stop-403142 kubelet[1537]: I0127 02:42:28.420561 1537 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6d8c4fd9-9ac5-4b5b-87b0-4aa8906d614e-xtables-lock\") pod \"kindnet-lbl6c\" (UID: \"6d8c4fd9-9ac5-4b5b-87b0-4aa8906d614e\") " pod="kube-system/kindnet-lbl6c"
Jan 27 02:42:28 scheduled-stop-403142 kubelet[1537]: I0127 02:42:28.420580 1537 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m78gx\" (UniqueName: \"kubernetes.io/projected/c3a9e3be-b238-49fc-848d-fa32ddb2cdb2-kube-api-access-m78gx\") pod \"kube-proxy-t5wqc\" (UID: \"c3a9e3be-b238-49fc-848d-fa32ddb2cdb2\") " pod="kube-system/kube-proxy-t5wqc"
Jan 27 02:42:28 scheduled-stop-403142 kubelet[1537]: I0127 02:42:28.420600 1537 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8cfp\" (UniqueName: \"kubernetes.io/projected/6d8c4fd9-9ac5-4b5b-87b0-4aa8906d614e-kube-api-access-z8cfp\") pod \"kindnet-lbl6c\" (UID: \"6d8c4fd9-9ac5-4b5b-87b0-4aa8906d614e\") " pod="kube-system/kindnet-lbl6c"
Jan 27 02:42:28 scheduled-stop-403142 kubelet[1537]: E0127 02:42:28.420748 1537 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found
Jan 27 02:42:28 scheduled-stop-403142 kubelet[1537]: E0127 02:42:28.420770 1537 projected.go:194] Error preparing data for projected volume kube-api-access-zqbl2 for pod kube-system/storage-provisioner: configmap "kube-root-ca.crt" not found
Jan 27 02:42:28 scheduled-stop-403142 kubelet[1537]: E0127 02:42:28.420819 1537 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/029f640c-de68-47fd-96dd-88130384396c-kube-api-access-zqbl2 podName:029f640c-de68-47fd-96dd-88130384396c nodeName:}" failed. No retries permitted until 2025-01-27 02:42:29.420795526 +0000 UTC m=+5.753767667 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-zqbl2" (UniqueName: "kubernetes.io/projected/029f640c-de68-47fd-96dd-88130384396c-kube-api-access-zqbl2") pod "storage-provisioner" (UID: "029f640c-de68-47fd-96dd-88130384396c") : configmap "kube-root-ca.crt" not found
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-arm64 status --format={{.APIServer}} -p scheduled-stop-403142 -n scheduled-stop-403142
helpers_test.go:261: (dbg) Run: kubectl --context scheduled-stop-403142 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: coredns-668d6bf9bc-pc54p kindnet-lbl6c kube-proxy-t5wqc storage-provisioner
helpers_test.go:274: ======> post-mortem[TestScheduledStopUnix]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context scheduled-stop-403142 describe pod coredns-668d6bf9bc-pc54p kindnet-lbl6c kube-proxy-t5wqc storage-provisioner
helpers_test.go:277: (dbg) Non-zero exit: kubectl --context scheduled-stop-403142 describe pod coredns-668d6bf9bc-pc54p kindnet-lbl6c kube-proxy-t5wqc storage-provisioner: exit status 1 (105.252834ms)
** stderr **
Error from server (NotFound): pods "coredns-668d6bf9bc-pc54p" not found
Error from server (NotFound): pods "kindnet-lbl6c" not found
Error from server (NotFound): pods "kube-proxy-t5wqc" not found
Error from server (NotFound): pods "storage-provisioner" not found
** /stderr **
helpers_test.go:279: kubectl --context scheduled-stop-403142 describe pod coredns-668d6bf9bc-pc54p kindnet-lbl6c kube-proxy-t5wqc storage-provisioner: exit status 1
helpers_test.go:175: Cleaning up "scheduled-stop-403142" profile ...
helpers_test.go:178: (dbg) Run: out/minikube-linux-arm64 delete -p scheduled-stop-403142
helpers_test.go:178: (dbg) Done: out/minikube-linux-arm64 delete -p scheduled-stop-403142: (2.022319439s)
--- FAIL: TestScheduledStopUnix (38.59s)