=== RUN TestNoKubernetes/serial/VerifyNok8sNoK8sDownloads
no_kubernetes_test.go:91: Checking cache directory: /home/jenkins/minikube-integration/21139-11202/.minikube/cache/linux/amd64/v0.0.0
no_kubernetes_test.go:100: Cache directory exists but is empty
no_kubernetes_test.go:102: Cache directory /home/jenkins/minikube-integration/21139-11202/.minikube/cache/linux/amd64/v0.0.0 should not exist when using --no-kubernetes
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:223: ======> post-mortem[TestNoKubernetes/serial/VerifyNok8sNoK8sDownloads]: network settings <======
helpers_test.go:230: HOST ENV snapshots: PROXY env: HTTP_PROXY="<empty>" HTTPS_PROXY="<empty>" NO_PROXY="<empty>"
helpers_test.go:238: ======> post-mortem[TestNoKubernetes/serial/VerifyNok8sNoK8sDownloads]: docker inspect <======
helpers_test.go:239: (dbg) Run: docker inspect NoKubernetes-711902
helpers_test.go:243: (dbg) docker inspect NoKubernetes-711902:
-- stdout --
[
{
"Id": "fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab",
"Created": "2025-10-09T18:33:10.603053192Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 281743,
"ExitCode": 0,
"Error": "",
"StartedAt": "2025-10-09T18:33:10.645086458Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:c6fde2176fdc734a0b1cf5396bccb3dc7d4299b26808035c9aa3b16b26946dbd",
"ResolvConfPath": "/var/lib/docker/containers/fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab/hostname",
"HostsPath": "/var/lib/docker/containers/fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab/hosts",
"LogPath": "/var/lib/docker/containers/fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab/fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab-json.log",
"Name": "/NoKubernetes-711902",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"NoKubernetes-711902:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "NoKubernetes-711902",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "private",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 3221225472,
"NanoCpus": 0,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 6442450944,
"MemorySwappiness": null,
"OomKillDisable": null,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"ID": "fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab",
"LowerDir": "/var/lib/docker/overlay2/dd5a88ca62f066a91467d9e879b36568f1b8b0752c4e706dff8c6e0c841d1ab8-init/diff:/var/lib/docker/overlay2/ef1c51768be2d023b7ddbd111b70e98dd2519a049dcc4d492c339f9603c55e42/diff",
"MergedDir": "/var/lib/docker/overlay2/dd5a88ca62f066a91467d9e879b36568f1b8b0752c4e706dff8c6e0c841d1ab8/merged",
"UpperDir": "/var/lib/docker/overlay2/dd5a88ca62f066a91467d9e879b36568f1b8b0752c4e706dff8c6e0c841d1ab8/diff",
"WorkDir": "/var/lib/docker/overlay2/dd5a88ca62f066a91467d9e879b36568f1b8b0752c4e706dff8c6e0c841d1ab8/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "NoKubernetes-711902",
"Source": "/var/lib/docker/volumes/NoKubernetes-711902/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "NoKubernetes-711902",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "NoKubernetes-711902",
"name.minikube.sigs.k8s.io": "NoKubernetes-711902",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "61282bb5119f90b697bdf1852d26e94baa57332f950b1822e4b6a969012bcd83",
"SandboxKey": "/var/run/docker/netns/61282bb5119f",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33003"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33004"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33007"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33005"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33006"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"NoKubernetes-711902": {
"IPAMConfig": {
"IPv4Address": "192.168.94.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "5a:5b:d0:94:11:ee",
"DriverOpts": null,
"GwPriority": 0,
"NetworkID": "d94f619ba0be9fc76aa2ff0b29d465f2b02b4a1aab78b268d3bf882cd9d17b28",
"EndpointID": "f089f78d38806d2c8fec53b7b6b12215669293e7f04d09926489230674a9aea7",
"Gateway": "192.168.94.1",
"IPAddress": "192.168.94.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"NoKubernetes-711902",
"fdb0ba874a7d"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p NoKubernetes-711902 -n NoKubernetes-711902
helpers_test.go:247: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.Host}} -p NoKubernetes-711902 -n NoKubernetes-711902: exit status 6 (298.156721ms)
-- stdout --
Running
WARNING: Your kubectl is pointing to stale minikube-vm.
To fix the kubectl context, run `minikube update-context`
-- /stdout --
** stderr **
E1009 18:33:17.212517 284193 status.go:458] kubeconfig endpoint: get endpoint: "NoKubernetes-711902" does not appear in /home/jenkins/minikube-integration/21139-11202/kubeconfig
** /stderr **
helpers_test.go:247: status error: exit status 6 (may be ok)
helpers_test.go:252: <<< TestNoKubernetes/serial/VerifyNok8sNoK8sDownloads FAILED: start of post-mortem logs <<<
helpers_test.go:253: ======> post-mortem[TestNoKubernetes/serial/VerifyNok8sNoK8sDownloads]: minikube logs <======
helpers_test.go:255: (dbg) Run: out/minikube-linux-amd64 -p NoKubernetes-711902 logs -n 25
helpers_test.go:260: TestNoKubernetes/serial/VerifyNok8sNoK8sDownloads logs:
-- stdout --
==> Audit <==
┌────────────┬─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┬─────────────────────────────┬──────────┬─────────┬─────────────────────┬─────────────────────┐
│ COMMAND │ ARGS │ PROFILE │ USER │ VERSION │ START TIME │ END TIME │
├────────────┼─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┼─────────────────────────────┼──────────┼─────────┼─────────────────────┼─────────────────────┤
│ stop │ -p scheduled-stop-803375 --schedule 5m │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 5m │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 5m │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 15s │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 15s │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 15s │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --cancel-scheduled │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ 09 Oct 25 18:29 UTC │
│ stop │ -p scheduled-stop-803375 --schedule 15s │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 15s │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 15s │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ 09 Oct 25 18:30 UTC │
│ delete │ -p scheduled-stop-803375 │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:30 UTC │ 09 Oct 25 18:30 UTC │
│ start │ -p skaffold-381243 --memory=3072 --driver=docker --container-runtime=docker │ skaffold-381243 │ jenkins │ v1.37.0 │ 09 Oct 25 18:30 UTC │ 09 Oct 25 18:31 UTC │
│ docker-env │ --shell none -p skaffold-381243 --user=skaffold │ skaffold-381243 │ skaffold │ v1.37.0 │ 09 Oct 25 18:31 UTC │ 09 Oct 25 18:31 UTC │
│ delete │ -p skaffold-381243 │ skaffold-381243 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ 09 Oct 25 18:32 UTC │
│ start │ -p insufficient-storage-290618 --memory=3072 --output=json --wait=true --driver=docker --container-runtime=docker │ insufficient-storage-290618 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ │
│ delete │ -p insufficient-storage-290618 │ insufficient-storage-290618 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ 09 Oct 25 18:32 UTC │
│ start │ -p pause-261319 --memory=3072 --install-addons=false --wait=all --driver=docker --container-runtime=docker │ pause-261319 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ │
│ start │ -p NoKubernetes-711902 --no-kubernetes --kubernetes-version=v1.28.0 --driver=docker --container-runtime=docker │ NoKubernetes-711902 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ │
│ start │ -p offline-docker-695286 --alsologtostderr -v=1 --memory=3072 --wait=true --driver=docker --container-runtime=docker │ offline-docker-695286 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ │
│ start │ -p NoKubernetes-711902 --memory=3072 --alsologtostderr -v=5 --driver=docker --container-runtime=docker │ NoKubernetes-711902 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ 09 Oct 25 18:32 UTC │
│ start │ -p stopped-upgrade-731100 --memory=3072 --vm-driver=docker --container-runtime=docker │ stopped-upgrade-731100 │ jenkins │ v1.32.0 │ 09 Oct 25 18:32 UTC │ 09 Oct 25 18:33 UTC │
│ start │ -p NoKubernetes-711902 --no-kubernetes --memory=3072 --alsologtostderr -v=5 --driver=docker --container-runtime=docker │ NoKubernetes-711902 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ 09 Oct 25 18:33 UTC │
│ delete │ -p NoKubernetes-711902 │ NoKubernetes-711902 │ jenkins │ v1.37.0 │ 09 Oct 25 18:33 UTC │ 09 Oct 25 18:33 UTC │
│ start │ -p NoKubernetes-711902 --no-kubernetes --memory=3072 --alsologtostderr -v=5 --driver=docker --container-runtime=docker │ NoKubernetes-711902 │ jenkins │ v1.37.0 │ 09 Oct 25 18:33 UTC │ 09 Oct 25 18:33 UTC │
│ stop │ stopped-upgrade-731100 stop │ stopped-upgrade-731100 │ jenkins │ v1.32.0 │ 09 Oct 25 18:33 UTC │ │
└────────────┴─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┴─────────────────────────────┴──────────┴─────────┴─────────────────────┴─────────────────────┘
==> Last Start <==
Log file created at: 2025/10/09 18:33:09
Running on machine: ubuntu-20-agent
Binary: Built with gc go1.24.6 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I1009 18:33:09.726820 281272 out.go:360] Setting OutFile to fd 1 ...
I1009 18:33:09.727079 281272 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1009 18:33:09.727087 281272 out.go:374] Setting ErrFile to fd 2...
I1009 18:33:09.727091 281272 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1009 18:33:09.727313 281272 root.go:338] Updating PATH: /home/jenkins/minikube-integration/21139-11202/.minikube/bin
I1009 18:33:09.727812 281272 out.go:368] Setting JSON to false
I1009 18:33:09.729162 281272 start.go:131] hostinfo: {"hostname":"ubuntu-20-agent","uptime":4531,"bootTime":1760030259,"procs":346,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"22.04","kernelVersion":"6.8.0-1041-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I1009 18:33:09.729247 281272 start.go:141] virtualization: kvm guest
I1009 18:33:09.731346 281272 out.go:179] * [NoKubernetes-711902] minikube v1.37.0 on Ubuntu 22.04 (kvm/amd64)
I1009 18:33:09.732706 281272 notify.go:220] Checking for updates...
I1009 18:33:09.732755 281272 out.go:179] - MINIKUBE_LOCATION=21139
I1009 18:33:09.734187 281272 out.go:179] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I1009 18:33:09.735610 281272 out.go:179] - KUBECONFIG=/home/jenkins/minikube-integration/21139-11202/kubeconfig
I1009 18:33:09.736931 281272 out.go:179] - MINIKUBE_HOME=/home/jenkins/minikube-integration/21139-11202/.minikube
I1009 18:33:09.738081 281272 out.go:179] - MINIKUBE_BIN=out/minikube-linux-amd64
I1009 18:33:09.739247 281272 out.go:179] - MINIKUBE_FORCE_SYSTEMD=
I1009 18:33:09.741194 281272 config.go:182] Loaded profile config "offline-docker-695286": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.34.1
I1009 18:33:09.741378 281272 config.go:182] Loaded profile config "pause-261319": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.34.1
I1009 18:33:09.741509 281272 config.go:182] Loaded profile config "stopped-upgrade-731100": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.28.3
I1009 18:33:09.741540 281272 start.go:1899] No Kubernetes flag is set, setting Kubernetes version to v0.0.0
I1009 18:33:09.741667 281272 driver.go:421] Setting default libvirt URI to qemu:///system
I1009 18:33:09.769435 281272 docker.go:123] docker version: linux-28.5.1:Docker Engine - Community
I1009 18:33:09.769662 281272 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1009 18:33:09.828796 281272 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:3 ContainersRunning:3 ContainersPaused:0 ContainersStopped:0 Images:4 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:false CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:64 OomKillDisable:false NGoroutines:76 SystemTime:2025-10-09 18:33:09.818416947 +0000 UTC LoggingDriver:json-file CgroupDriver:systemd NEventsListener:0 KernelVersion:6.8.0-1041-gcp OperatingSystem:Ubuntu 22.04.5 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33652174848 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent Labels:[] ExperimentalBuild:false ServerVersion:28.5.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:b98a3aace656320842a23f4a392a33f46af97866 Expected:} RuncCommit:{ID:v1.3.0-0-g4ca628d1 Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin name=cgroupns] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map
[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.29.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.40.0] map[Name:model Path:/usr/libexec/docker/cli-plugins/docker-model SchemaVersion:0.1.0 ShortDescription:Docker Model Runner Vendor:Docker Inc. Version:v0.1.42] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I1009 18:33:09.828892 281272 docker.go:318] overlay module found
I1009 18:33:09.830759 281272 out.go:179] * Using the docker driver based on user configuration
I1009 18:33:09.832210 281272 start.go:305] selected driver: docker
I1009 18:33:09.832225 281272 start.go:925] validating driver "docker" against <nil>
I1009 18:33:09.832236 281272 start.go:936] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I1009 18:33:09.832801 281272 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1009 18:33:09.893498 281272 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:3 ContainersRunning:3 ContainersPaused:0 ContainersStopped:0 Images:4 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:false CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:64 OomKillDisable:false NGoroutines:76 SystemTime:2025-10-09 18:33:09.883713298 +0000 UTC LoggingDriver:json-file CgroupDriver:systemd NEventsListener:0 KernelVersion:6.8.0-1041-gcp OperatingSystem:Ubuntu 22.04.5 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33652174848 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent Labels:[] ExperimentalBuild:false ServerVersion:28.5.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:b98a3aace656320842a23f4a392a33f46af97866 Expected:} RuncCommit:{ID:v1.3.0-0-g4ca628d1 Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin name=cgroupns] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map
[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.29.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.40.0] map[Name:model Path:/usr/libexec/docker/cli-plugins/docker-model SchemaVersion:0.1.0 ShortDescription:Docker Model Runner Vendor:Docker Inc. Version:v0.1.42] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I1009 18:33:09.893624 281272 start.go:1899] No Kubernetes flag is set, setting Kubernetes version to v0.0.0
I1009 18:33:09.893695 281272 start_flags.go:327] no existing cluster config was found, will generate one from the flags
I1009 18:33:09.893889 281272 start_flags.go:974] Wait components to verify : map[apiserver:true system_pods:true]
I1009 18:33:09.895859 281272 out.go:179] * Using Docker driver with root privileges
I1009 18:33:09.897317 281272 cni.go:84] Creating CNI manager for ""
I1009 18:33:09.897387 281272 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I1009 18:33:09.897398 281272 start_flags.go:336] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I1009 18:33:09.897423 281272 start.go:1899] No Kubernetes flag is set, setting Kubernetes version to v0.0.0
I1009 18:33:09.897472 281272 start.go:349] cluster config:
{Name:NoKubernetes-711902 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v0.0.0 ClusterName:NoKubernetes-711902 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local Conta
inerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v0.0.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I1009 18:33:09.898767 281272 out.go:179] * Starting minikube without Kubernetes in cluster NoKubernetes-711902
I1009 18:33:09.899911 281272 cache.go:133] Beginning downloading kic base image for docker with docker
I1009 18:33:09.901159 281272 out.go:179] * Pulling base image v0.0.48-1759745255-21703 ...
I1009 18:33:09.902378 281272 cache.go:58] Skipping Kubernetes image caching due to --no-kubernetes flag
I1009 18:33:09.902483 281272 image.go:81] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92 in local docker daemon
I1009 18:33:09.902514 281272 profile.go:143] Saving config to /home/jenkins/minikube-integration/21139-11202/.minikube/profiles/NoKubernetes-711902/config.json ...
I1009 18:33:09.902552 281272 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21139-11202/.minikube/profiles/NoKubernetes-711902/config.json: {Name:mk7ed4d136dae909a52d53e49ef555ca81eefd7b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1009 18:33:09.924695 281272 image.go:100] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92 in local docker daemon, skipping pull
I1009 18:33:09.924718 281272 cache.go:157] gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92 exists in daemon, skipping load
I1009 18:33:09.924738 281272 cache.go:242] Successfully downloaded all kic artifacts
I1009 18:33:09.924765 281272 start.go:360] acquireMachinesLock for NoKubernetes-711902: {Name:mk77177cbada15c1cc71ca39efd4970ee070e38a Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I1009 18:33:09.924827 281272 start.go:364] duration metric: took 46.818µs to acquireMachinesLock for "NoKubernetes-711902"
I1009 18:33:09.924851 281272 start.go:93] Provisioning new machine with config: &{Name:NoKubernetes-711902 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v0.0.0 ClusterName:NoKubernetes-711902 Namespace:default APIServerHAVIP: APISer
verName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v0.0.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSoc
k: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v0.0.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I1009 18:33:09.925024 281272 start.go:125] createHost starting for "" (driver="docker")
I1009 18:33:12.931630 265579 kubeadm.go:322] [apiclient] All control plane components are healthy after 5.502455 seconds
I1009 18:33:12.931770 265579 kubeadm.go:322] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I1009 18:33:12.945750 265579 kubeadm.go:322] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I1009 18:33:13.465875 265579 kubeadm.go:322] [upload-certs] Skipping phase. Please see --upload-certs
I1009 18:33:13.466133 265579 kubeadm.go:322] [mark-control-plane] Marking the node stopped-upgrade-731100 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I1009 18:33:13.975633 265579 kubeadm.go:322] [bootstrap-token] Using token: 2jh3nb.jrabffh7nvf4snwz
I1009 18:33:13.977481 265579 out.go:204] - Configuring RBAC rules ...
I1009 18:33:13.977640 265579 kubeadm.go:322] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I1009 18:33:13.981530 265579 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I1009 18:33:13.989295 265579 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I1009 18:33:13.992957 265579 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I1009 18:33:13.995784 265579 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I1009 18:33:13.998990 265579 kubeadm.go:322] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I1009 18:33:14.011058 265579 kubeadm.go:322] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I1009 18:33:14.185848 265579 kubeadm.go:322] [addons] Applied essential addon: CoreDNS
I1009 18:33:14.386651 265579 kubeadm.go:322] [addons] Applied essential addon: kube-proxy
I1009 18:33:14.387580 265579 kubeadm.go:322]
I1009 18:33:14.387688 265579 kubeadm.go:322] Your Kubernetes control-plane has initialized successfully!
I1009 18:33:14.387693 265579 kubeadm.go:322]
I1009 18:33:14.387858 265579 kubeadm.go:322] To start using your cluster, you need to run the following as a regular user:
I1009 18:33:14.387865 265579 kubeadm.go:322]
I1009 18:33:14.387898 265579 kubeadm.go:322] mkdir -p $HOME/.kube
I1009 18:33:14.387962 265579 kubeadm.go:322] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I1009 18:33:14.388049 265579 kubeadm.go:322] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I1009 18:33:14.388061 265579 kubeadm.go:322]
I1009 18:33:14.388132 265579 kubeadm.go:322] Alternatively, if you are the root user, you can run:
I1009 18:33:14.388137 265579 kubeadm.go:322]
I1009 18:33:14.388199 265579 kubeadm.go:322] export KUBECONFIG=/etc/kubernetes/admin.conf
I1009 18:33:14.388205 265579 kubeadm.go:322]
I1009 18:33:14.388304 265579 kubeadm.go:322] You should now deploy a pod network to the cluster.
I1009 18:33:14.388408 265579 kubeadm.go:322] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I1009 18:33:14.388497 265579 kubeadm.go:322] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I1009 18:33:14.388503 265579 kubeadm.go:322]
I1009 18:33:14.388647 265579 kubeadm.go:322] You can now join any number of control-plane nodes by copying certificate authorities
I1009 18:33:14.388750 265579 kubeadm.go:322] and service account keys on each node and then running the following as root:
I1009 18:33:14.388758 265579 kubeadm.go:322]
I1009 18:33:14.388898 265579 kubeadm.go:322] kubeadm join control-plane.minikube.internal:8443 --token 2jh3nb.jrabffh7nvf4snwz \
I1009 18:33:14.389022 265579 kubeadm.go:322] --discovery-token-ca-cert-hash sha256:50b267e6e7636e43468b46774c69e9781b208d8d6164a67ebb20e03119db0f84 \
I1009 18:33:14.389060 265579 kubeadm.go:322] --control-plane
I1009 18:33:14.389070 265579 kubeadm.go:322]
I1009 18:33:14.389201 265579 kubeadm.go:322] Then you can join any number of worker nodes by running the following on each as root:
I1009 18:33:14.389213 265579 kubeadm.go:322]
I1009 18:33:14.389319 265579 kubeadm.go:322] kubeadm join control-plane.minikube.internal:8443 --token 2jh3nb.jrabffh7nvf4snwz \
I1009 18:33:14.389471 265579 kubeadm.go:322] --discovery-token-ca-cert-hash sha256:50b267e6e7636e43468b46774c69e9781b208d8d6164a67ebb20e03119db0f84
I1009 18:33:14.391453 265579 kubeadm.go:322] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/6.8.0-1041-gcp\n", err: exit status 1
I1009 18:33:14.391574 265579 kubeadm.go:322] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I1009 18:33:14.391614 265579 cni.go:84] Creating CNI manager for ""
I1009 18:33:14.391631 265579 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I1009 18:33:14.393306 265579 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I1009 18:33:14.394495 265579 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I1009 18:33:14.405223 265579 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (457 bytes)
I1009 18:33:14.425689 265579 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I1009 18:33:14.425804 265579 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.3/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I1009 18:33:14.425816 265579 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.3/kubectl label nodes minikube.k8s.io/version=v1.32.0 minikube.k8s.io/commit=8220a6eb95f0a4d75f7f2d7b14cef975f050512d minikube.k8s.io/name=stopped-upgrade-731100 minikube.k8s.io/updated_at=2025_10_09T18_33_14_0700 minikube.k8s.io/primary=true --all --overwrite --kubeconfig=/var/lib/minikube/kubeconfig
I1009 18:33:14.434356 265579 ops.go:34] apiserver oom_adj: -16
I1009 18:33:14.518361 265579 kubeadm.go:1081] duration metric: took 92.625873ms to wait for elevateKubeSystemPrivileges.
I1009 18:33:14.518397 265579 kubeadm.go:406] StartCluster complete in 10.424255465s
I1009 18:33:14.518419 265579 settings.go:142] acquiring lock: {Name:mk068a134d3e4d40102c45be73d7211efe819775 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1009 18:33:14.518544 265579 settings.go:150] Updating kubeconfig: /tmp/legacy_kubeconfig880889346
I1009 18:33:14.519175 265579 lock.go:35] WriteFile acquiring /tmp/legacy_kubeconfig880889346: {Name:mk211495ca79b7e4f04aeeeac75527902ecb4d94 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1009 18:33:14.519497 265579 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I1009 18:33:14.519501 265579 addons.go:499] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false helm-tiller:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:false volumesnapshots:false]
I1009 18:33:14.519578 265579 addons.go:69] Setting storage-provisioner=true in profile "stopped-upgrade-731100"
I1009 18:33:14.519580 265579 addons.go:69] Setting default-storageclass=true in profile "stopped-upgrade-731100"
I1009 18:33:14.519622 265579 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "stopped-upgrade-731100"
I1009 18:33:14.519640 265579 addons.go:231] Setting addon storage-provisioner=true in "stopped-upgrade-731100"
I1009 18:33:14.519706 265579 host.go:66] Checking if "stopped-upgrade-731100" exists ...
I1009 18:33:14.519799 265579 config.go:182] Loaded profile config "stopped-upgrade-731100": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.28.3
I1009 18:33:14.520071 265579 cli_runner.go:164] Run: docker container inspect stopped-upgrade-731100 --format={{.State.Status}}
I1009 18:33:14.520240 265579 cli_runner.go:164] Run: docker container inspect stopped-upgrade-731100 --format={{.State.Status}}
I1009 18:33:14.541424 265579 kapi.go:248] "coredns" deployment in "kube-system" namespace and "stopped-upgrade-731100" context rescaled to 1 replicas
I1009 18:33:14.541458 265579 start.go:223] Will wait 6m0s for node &{Name: IP:192.168.103.2 Port:8443 KubernetesVersion:v1.28.3 ContainerRuntime:docker ControlPlane:true Worker:true}
I1009 18:33:14.542920 265579 out.go:177] * Verifying Kubernetes components...
I1009 18:33:14.544123 265579 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I1009 18:33:14.545665 265579 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I1009 18:33:09.814481 264343 pod_ready.go:99] pod "coredns-66bc5c9577-xfh69" in "kube-system" namespace is gone: getting pod "coredns-66bc5c9577-xfh69" in "kube-system" namespace (will retry): pods "coredns-66bc5c9577-xfh69" not found
I1009 18:33:09.814496 264343 pod_ready.go:86] duration metric: took 11.002870301s for pod "coredns-66bc5c9577-xfh69" in "kube-system" namespace to be "Ready" or be gone ...
I1009 18:33:09.814506 264343 pod_ready.go:83] waiting for pod "coredns-66bc5c9577-xqpvs" in "kube-system" namespace to be "Ready" or be gone ...
W1009 18:33:11.819352 264343 pod_ready.go:104] pod "coredns-66bc5c9577-xqpvs" is not "Ready", error: <nil>
W1009 18:33:13.821852 264343 pod_ready.go:104] pod "coredns-66bc5c9577-xqpvs" is not "Ready", error: <nil>
W1009 18:33:12.156435 264341 pod_ready.go:104] pod "coredns-66bc5c9577-qzfd8" is not "Ready", error: <nil>
W1009 18:33:14.157176 264341 pod_ready.go:104] pod "coredns-66bc5c9577-qzfd8" is not "Ready", error: <nil>
I1009 18:33:09.926837 281272 out.go:252] * Creating docker container (CPUs=2, Memory=3072MB) ...
I1009 18:33:09.927080 281272 start.go:159] libmachine.API.Create for "NoKubernetes-711902" (driver="docker")
I1009 18:33:09.927121 281272 client.go:168] LocalClient.Create starting
I1009 18:33:09.927172 281272 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem
I1009 18:33:09.927204 281272 main.go:141] libmachine: Decoding PEM data...
I1009 18:33:09.927219 281272 main.go:141] libmachine: Parsing certificate...
I1009 18:33:09.927274 281272 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21139-11202/.minikube/certs/cert.pem
I1009 18:33:09.927294 281272 main.go:141] libmachine: Decoding PEM data...
I1009 18:33:09.927310 281272 main.go:141] libmachine: Parsing certificate...
I1009 18:33:09.927670 281272 cli_runner.go:164] Run: docker network inspect NoKubernetes-711902 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W1009 18:33:09.945553 281272 cli_runner.go:211] docker network inspect NoKubernetes-711902 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I1009 18:33:09.945659 281272 network_create.go:284] running [docker network inspect NoKubernetes-711902] to gather additional debugging logs...
I1009 18:33:09.945687 281272 cli_runner.go:164] Run: docker network inspect NoKubernetes-711902
W1009 18:33:09.963224 281272 cli_runner.go:211] docker network inspect NoKubernetes-711902 returned with exit code 1
I1009 18:33:09.963257 281272 network_create.go:287] error running [docker network inspect NoKubernetes-711902]: docker network inspect NoKubernetes-711902: exit status 1
stdout:
[]
stderr:
Error response from daemon: network NoKubernetes-711902 not found
I1009 18:33:09.963289 281272 network_create.go:289] output of [docker network inspect NoKubernetes-711902]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network NoKubernetes-711902 not found
** /stderr **
I1009 18:33:09.963420 281272 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I1009 18:33:09.982077 281272 network.go:211] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-36d16f1d4123 IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:0a:9a:9a:77:cb:7c} reservation:<nil>}
I1009 18:33:09.982636 281272 network.go:211] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-8adbd295428d IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:22:1a:8a:aa:f1:33} reservation:<nil>}
I1009 18:33:09.983150 281272 network.go:211] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-9d29dc29aaeb IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:a2:d7:08:db:e1:fb} reservation:<nil>}
I1009 18:33:09.983716 281272 network.go:211] skipping subnet 192.168.76.0/24 that is taken: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName:br-9c468df904f3 IfaceIPv4:192.168.76.1 IfaceMTU:1500 IfaceMAC:0a:e0:89:50:00:c6} reservation:<nil>}
I1009 18:33:09.984371 281272 network.go:211] skipping subnet 192.168.85.0/24 that is taken: &{IP:192.168.85.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.85.0/24 Gateway:192.168.85.1 ClientMin:192.168.85.2 ClientMax:192.168.85.254 Broadcast:192.168.85.255 IsPrivate:true Interface:{IfaceName:br-b0a9a97c4749 IfaceIPv4:192.168.85.1 IfaceMTU:1500 IfaceMAC:4e:1e:0c:e6:fc:72} reservation:<nil>}
I1009 18:33:09.985210 281272 network.go:206] using free private subnet 192.168.94.0/24: &{IP:192.168.94.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.94.0/24 Gateway:192.168.94.1 ClientMin:192.168.94.2 ClientMax:192.168.94.254 Broadcast:192.168.94.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc00205ad10}
I1009 18:33:09.985234 281272 network_create.go:124] attempt to create docker network NoKubernetes-711902 192.168.94.0/24 with gateway 192.168.94.1 and MTU of 1500 ...
I1009 18:33:09.985291 281272 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.94.0/24 --gateway=192.168.94.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=NoKubernetes-711902 NoKubernetes-711902
I1009 18:33:10.046655 281272 network_create.go:108] docker network NoKubernetes-711902 192.168.94.0/24 created
I1009 18:33:10.046710 281272 kic.go:121] calculated static IP "192.168.94.2" for the "NoKubernetes-711902" container
I1009 18:33:10.046804 281272 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I1009 18:33:10.065918 281272 cli_runner.go:164] Run: docker volume create NoKubernetes-711902 --label name.minikube.sigs.k8s.io=NoKubernetes-711902 --label created_by.minikube.sigs.k8s.io=true
I1009 18:33:10.086849 281272 oci.go:103] Successfully created a docker volume NoKubernetes-711902
I1009 18:33:10.086930 281272 cli_runner.go:164] Run: docker run --rm --name NoKubernetes-711902-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=NoKubernetes-711902 --entrypoint /usr/bin/test -v NoKubernetes-711902:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92 -d /var/lib
I1009 18:33:10.518077 281272 oci.go:107] Successfully prepared a docker volume NoKubernetes-711902
I1009 18:33:10.518163 281272 preload.go:178] Skipping preload logic due to --no-kubernetes flag
W1009 18:33:10.518272 281272 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
W1009 18:33:10.518314 281272 oci.go:252] Your kernel does not support CPU cfs period/quota or the cgroup is not mounted.
I1009 18:33:10.518359 281272 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I1009 18:33:10.582290 281272 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname NoKubernetes-711902 --name NoKubernetes-711902 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=NoKubernetes-711902 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=NoKubernetes-711902 --network NoKubernetes-711902 --ip 192.168.94.2 --volume NoKubernetes-711902:/var --security-opt apparmor=unconfined --memory=3072mb -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92
I1009 18:33:10.896323 281272 cli_runner.go:164] Run: docker container inspect NoKubernetes-711902 --format={{.State.Running}}
I1009 18:33:10.918963 281272 cli_runner.go:164] Run: docker container inspect NoKubernetes-711902 --format={{.State.Status}}
I1009 18:33:10.938118 281272 cli_runner.go:164] Run: docker exec NoKubernetes-711902 stat /var/lib/dpkg/alternatives/iptables
I1009 18:33:10.992503 281272 oci.go:144] the created container "NoKubernetes-711902" has a running status.
I1009 18:33:10.992533 281272 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa...
I1009 18:33:11.562080 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa.pub -> /home/docker/.ssh/authorized_keys
I1009 18:33:11.562135 281272 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I1009 18:33:11.588890 281272 cli_runner.go:164] Run: docker container inspect NoKubernetes-711902 --format={{.State.Status}}
I1009 18:33:11.607636 281272 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I1009 18:33:11.607661 281272 kic_runner.go:114] Args: [docker exec --privileged NoKubernetes-711902 chown docker:docker /home/docker/.ssh/authorized_keys]
I1009 18:33:11.652898 281272 cli_runner.go:164] Run: docker container inspect NoKubernetes-711902 --format={{.State.Status}}
I1009 18:33:11.672707 281272 machine.go:93] provisionDockerMachine start ...
I1009 18:33:11.672817 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:11.691520 281272 main.go:141] libmachine: Using SSH client type: native
I1009 18:33:11.691795 281272 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x840040] 0x842d40 <nil> [] 0s} 127.0.0.1 33003 <nil> <nil>}
I1009 18:33:11.691817 281272 main.go:141] libmachine: About to run SSH command:
hostname
I1009 18:33:11.837755 281272 main.go:141] libmachine: SSH cmd err, output: <nil>: NoKubernetes-711902
I1009 18:33:11.837785 281272 ubuntu.go:182] provisioning hostname "NoKubernetes-711902"
I1009 18:33:11.837858 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:11.858762 281272 main.go:141] libmachine: Using SSH client type: native
I1009 18:33:11.859058 281272 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x840040] 0x842d40 <nil> [] 0s} 127.0.0.1 33003 <nil> <nil>}
I1009 18:33:11.859077 281272 main.go:141] libmachine: About to run SSH command:
sudo hostname NoKubernetes-711902 && echo "NoKubernetes-711902" | sudo tee /etc/hostname
I1009 18:33:12.017403 281272 main.go:141] libmachine: SSH cmd err, output: <nil>: NoKubernetes-711902
I1009 18:33:12.017483 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:12.034992 281272 main.go:141] libmachine: Using SSH client type: native
I1009 18:33:12.035252 281272 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x840040] 0x842d40 <nil> [] 0s} 127.0.0.1 33003 <nil> <nil>}
I1009 18:33:12.035272 281272 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sNoKubernetes-711902' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 NoKubernetes-711902/g' /etc/hosts;
else
echo '127.0.1.1 NoKubernetes-711902' | sudo tee -a /etc/hosts;
fi
fi
I1009 18:33:12.184962 281272 main.go:141] libmachine: SSH cmd err, output: <nil>:
I1009 18:33:12.185000 281272 ubuntu.go:188] set auth options {CertDir:/home/jenkins/minikube-integration/21139-11202/.minikube CaCertPath:/home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/21139-11202/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/21139-11202/.minikube}
I1009 18:33:12.185034 281272 ubuntu.go:190] setting up certificates
I1009 18:33:12.185047 281272 provision.go:84] configureAuth start
I1009 18:33:12.185114 281272 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" NoKubernetes-711902
I1009 18:33:12.205934 281272 provision.go:143] copyHostCerts
I1009 18:33:12.205971 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem -> /home/jenkins/minikube-integration/21139-11202/.minikube/ca.pem
I1009 18:33:12.206006 281272 exec_runner.go:144] found /home/jenkins/minikube-integration/21139-11202/.minikube/ca.pem, removing ...
I1009 18:33:12.206019 281272 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21139-11202/.minikube/ca.pem
I1009 18:33:12.206113 281272 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/21139-11202/.minikube/ca.pem (1078 bytes)
I1009 18:33:12.206228 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/cert.pem -> /home/jenkins/minikube-integration/21139-11202/.minikube/cert.pem
I1009 18:33:12.206256 281272 exec_runner.go:144] found /home/jenkins/minikube-integration/21139-11202/.minikube/cert.pem, removing ...
I1009 18:33:12.206263 281272 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21139-11202/.minikube/cert.pem
I1009 18:33:12.206311 281272 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/21139-11202/.minikube/cert.pem (1123 bytes)
I1009 18:33:12.206381 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/key.pem -> /home/jenkins/minikube-integration/21139-11202/.minikube/key.pem
I1009 18:33:12.206406 281272 exec_runner.go:144] found /home/jenkins/minikube-integration/21139-11202/.minikube/key.pem, removing ...
I1009 18:33:12.206412 281272 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21139-11202/.minikube/key.pem
I1009 18:33:12.206448 281272 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/21139-11202/.minikube/key.pem (1675 bytes)
I1009 18:33:12.206526 281272 provision.go:117] generating server cert: /home/jenkins/minikube-integration/21139-11202/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca-key.pem org=jenkins.NoKubernetes-711902 san=[127.0.0.1 192.168.94.2 NoKubernetes-711902 localhost minikube]
I1009 18:33:12.882770 281272 provision.go:177] copyRemoteCerts
I1009 18:33:12.882824 281272 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I1009 18:33:12.882859 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:12.901749 281272 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33003 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa Username:docker}
I1009 18:33:13.007807 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem -> /etc/docker/ca.pem
I1009 18:33:13.007877 281272 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I1009 18:33:13.028112 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/machines/server.pem -> /etc/docker/server.pem
I1009 18:33:13.028171 281272 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21139-11202/.minikube/machines/server.pem --> /etc/docker/server.pem (1224 bytes)
I1009 18:33:13.046902 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/machines/server-key.pem -> /etc/docker/server-key.pem
I1009 18:33:13.046959 281272 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21139-11202/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I1009 18:33:13.065410 281272 provision.go:87] duration metric: took 880.347046ms to configureAuth
I1009 18:33:13.065435 281272 ubuntu.go:206] setting minikube options for container-runtime
I1009 18:33:13.065668 281272 config.go:182] Loaded profile config "NoKubernetes-711902": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v0.0.0
I1009 18:33:13.065740 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:13.084447 281272 main.go:141] libmachine: Using SSH client type: native
I1009 18:33:13.084715 281272 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x840040] 0x842d40 <nil> [] 0s} 127.0.0.1 33003 <nil> <nil>}
I1009 18:33:13.084730 281272 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I1009 18:33:13.232775 281272 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I1009 18:33:13.232799 281272 ubuntu.go:71] root file system type: overlay
I1009 18:33:13.232968 281272 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I1009 18:33:13.233039 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:13.254715 281272 main.go:141] libmachine: Using SSH client type: native
I1009 18:33:13.254967 281272 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x840040] 0x842d40 <nil> [] 0s} 127.0.0.1 33003 <nil> <nil>}
I1009 18:33:13.255043 281272 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network-online.target nss-lookup.target docker.socket firewalld.service containerd.service time-set.target
Wants=network-online.target containerd.service
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=always
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 \
-H fd:// --containerd=/run/containerd/containerd.sock \
-H unix:///var/run/docker.sock \
--default-ulimit=nofile=1048576:1048576 \
--tlsverify \
--tlscacert /etc/docker/ca.pem \
--tlscert /etc/docker/server.pem \
--tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I1009 18:33:13.424326 281272 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network-online.target nss-lookup.target docker.socket firewalld.service containerd.service time-set.target
Wants=network-online.target containerd.service
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=always
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H fd:// --containerd=/run/containerd/containerd.sock -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
I1009 18:33:13.424393 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:13.444310 281272 main.go:141] libmachine: Using SSH client type: native
I1009 18:33:13.444506 281272 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x840040] 0x842d40 <nil> [] 0s} 127.0.0.1 33003 <nil> <nil>}
I1009 18:33:13.444527 281272 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I1009 18:33:14.654138 281272 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2025-10-02 14:52:52.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2025-10-09 18:33:13.422146133 +0000
@@ -9,23 +9,34 @@
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
Restart=always
+
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H fd:// --containerd=/run/containerd/containerd.sock -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
+
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I1009 18:33:14.654174 281272 machine.go:96] duration metric: took 2.981445688s to provisionDockerMachine
I1009 18:33:14.654187 281272 client.go:171] duration metric: took 4.727059425s to LocalClient.Create
I1009 18:33:14.654212 281272 start.go:167] duration metric: took 4.727134072s to libmachine.API.Create "NoKubernetes-711902"
I1009 18:33:14.654220 281272 start.go:293] postStartSetup for "NoKubernetes-711902" (driver="docker")
I1009 18:33:14.654232 281272 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I1009 18:33:14.654295 281272 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I1009 18:33:14.654345 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:14.680894 281272 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33003 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa Username:docker}
I1009 18:33:14.544979 265579 addons.go:231] Setting addon default-storageclass=true in "stopped-upgrade-731100"
I1009 18:33:14.547744 265579 addons.go:423] installing /etc/kubernetes/addons/storage-provisioner.yaml
I1009 18:33:14.547759 265579 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I1009 18:33:14.547758 265579 host.go:66] Checking if "stopped-upgrade-731100" exists ...
I1009 18:33:14.547816 265579 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" stopped-upgrade-731100
I1009 18:33:14.548246 265579 cli_runner.go:164] Run: docker container inspect stopped-upgrade-731100 --format={{.State.Status}}
I1009 18:33:14.569783 265579 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32998 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/stopped-upgrade-731100/id_rsa Username:docker}
I1009 18:33:14.570364 265579 addons.go:423] installing /etc/kubernetes/addons/storageclass.yaml
I1009 18:33:14.570376 265579 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I1009 18:33:14.570431 265579 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" stopped-upgrade-731100
I1009 18:33:14.595217 265579 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32998 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/stopped-upgrade-731100/id_rsa Username:docker}
I1009 18:33:14.604074 265579 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.103.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.28.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I1009 18:33:14.604933 265579 api_server.go:52] waiting for apiserver process to appear ...
I1009 18:33:14.604971 265579 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I1009 18:33:14.680524 265579 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.3/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I1009 18:33:14.703459 265579 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.3/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I1009 18:33:15.256158 265579 start.go:926] {"host.minikube.internal": 192.168.103.1} host record injected into CoreDNS's ConfigMap
I1009 18:33:15.256230 265579 api_server.go:72] duration metric: took 714.740057ms to wait for apiserver process to appear ...
I1009 18:33:15.256247 265579 api_server.go:88] waiting for apiserver healthz status ...
I1009 18:33:15.256267 265579 api_server.go:253] Checking apiserver healthz at https://192.168.103.2:8443/healthz ...
I1009 18:33:15.262308 265579 api_server.go:279] https://192.168.103.2:8443/healthz returned 200:
ok
I1009 18:33:15.263824 265579 api_server.go:141] control plane version: v1.28.3
I1009 18:33:15.263841 265579 api_server.go:131] duration metric: took 7.588233ms to wait for apiserver health ...
I1009 18:33:15.263848 265579 system_pods.go:43] waiting for kube-system pods to appear ...
I1009 18:33:15.271088 265579 system_pods.go:59] 4 kube-system pods found
I1009 18:33:15.271113 265579 system_pods.go:61] "etcd-stopped-upgrade-731100" [054d6170-1a85-4625-8037-37ebefcc4574] Pending
I1009 18:33:15.271124 265579 system_pods.go:61] "kube-apiserver-stopped-upgrade-731100" [899ed27c-d0e7-4d8e-a4ee-15298ddee726] Pending
I1009 18:33:15.271130 265579 system_pods.go:61] "kube-controller-manager-stopped-upgrade-731100" [a7f2a9ed-be6b-4e22-b4da-1305a3e56228] Pending
I1009 18:33:15.271136 265579 system_pods.go:61] "kube-scheduler-stopped-upgrade-731100" [8ec66241-649a-46c8-9e47-bf861cb836f9] Pending
I1009 18:33:15.271143 265579 system_pods.go:74] duration metric: took 7.288345ms to wait for pod list to return data ...
I1009 18:33:15.271152 265579 kubeadm.go:581] duration metric: took 729.665151ms to wait for : map[apiserver:true system_pods:true] ...
I1009 18:33:15.271165 265579 node_conditions.go:102] verifying NodePressure condition ...
I1009 18:33:15.275042 265579 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I1009 18:33:15.275060 265579 node_conditions.go:123] node cpu capacity is 8
I1009 18:33:15.275075 265579 node_conditions.go:105] duration metric: took 3.904965ms to run NodePressure ...
I1009 18:33:15.275088 265579 start.go:228] waiting for startup goroutines ...
I1009 18:33:15.427838 265579 out.go:177] * Enabled addons: storage-provisioner, default-storageclass
I1009 18:33:15.429174 265579 addons.go:502] enable addons completed in 909.675466ms: enabled=[storage-provisioner default-storageclass]
I1009 18:33:15.429202 265579 start.go:233] waiting for cluster config update ...
I1009 18:33:15.429212 265579 start.go:242] writing updated cluster config ...
I1009 18:33:15.429444 265579 ssh_runner.go:195] Run: rm -f paused
I1009 18:33:15.479963 265579 start.go:600] kubectl: 1.34.1, cluster: 1.28.3 (minor skew: 6)
I1009 18:33:15.481658 265579 out.go:177]
W1009 18:33:15.483575 265579 out.go:239] ! /usr/local/bin/kubectl is version 1.34.1, which may have incompatibilities with Kubernetes 1.28.3.
I1009 18:33:15.484917 265579 out.go:177] - Want kubectl v1.28.3? Try 'minikube kubectl -- get pods -A'
I1009 18:33:15.486762 265579 out.go:177] * Done! kubectl is now configured to use "stopped-upgrade-731100" cluster and "default" namespace by default
I1009 18:33:14.793997 281272 ssh_runner.go:195] Run: cat /etc/os-release
I1009 18:33:14.797769 281272 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I1009 18:33:14.797803 281272 info.go:137] Remote host: Debian GNU/Linux 12 (bookworm)
I1009 18:33:14.797827 281272 filesync.go:126] Scanning /home/jenkins/minikube-integration/21139-11202/.minikube/addons for local assets ...
I1009 18:33:14.797883 281272 filesync.go:126] Scanning /home/jenkins/minikube-integration/21139-11202/.minikube/files for local assets ...
I1009 18:33:14.797983 281272 filesync.go:149] local asset: /home/jenkins/minikube-integration/21139-11202/.minikube/files/etc/ssl/certs/147382.pem -> 147382.pem in /etc/ssl/certs
I1009 18:33:14.797996 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/files/etc/ssl/certs/147382.pem -> /etc/ssl/certs/147382.pem
I1009 18:33:14.798080 281272 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I1009 18:33:14.806796 281272 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21139-11202/.minikube/files/etc/ssl/certs/147382.pem --> /etc/ssl/certs/147382.pem (1708 bytes)
I1009 18:33:14.829305 281272 start.go:296] duration metric: took 175.07094ms for postStartSetup
I1009 18:33:14.829741 281272 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" NoKubernetes-711902
I1009 18:33:14.849149 281272 profile.go:143] Saving config to /home/jenkins/minikube-integration/21139-11202/.minikube/profiles/NoKubernetes-711902/config.json ...
I1009 18:33:14.849424 281272 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I1009 18:33:14.849473 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:14.868321 281272 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33003 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa Username:docker}
I1009 18:33:14.970174 281272 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I1009 18:33:14.974960 281272 start.go:128] duration metric: took 5.049916115s to createHost
I1009 18:33:14.974987 281272 start.go:83] releasing machines lock for "NoKubernetes-711902", held for 5.050146708s
I1009 18:33:14.975061 281272 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" NoKubernetes-711902
I1009 18:33:15.002433 281272 ssh_runner.go:195] Run: cat /version.json
I1009 18:33:15.002450 281272 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I1009 18:33:15.002489 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:15.002501 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:15.021468 281272 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33003 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa Username:docker}
I1009 18:33:15.022782 281272 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33003 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa Username:docker}
I1009 18:33:15.126408 281272 ssh_runner.go:195] Run: systemctl --version
I1009 18:33:15.204672 281272 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
W1009 18:33:15.210075 281272 cni.go:209] loopback cni configuration skipped: "/etc/cni/net.d/*loopback.conf*" not found
I1009 18:33:15.210143 281272 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I1009 18:33:15.240346 281272 cni.go:262] disabled [/etc/cni/net.d/10-crio-bridge.conflist.disabled, /etc/cni/net.d/87-podman-bridge.conflist] bridge cni config(s)
I1009 18:33:15.240375 281272 start.go:495] detecting cgroup driver to use...
I1009 18:33:15.240410 281272 detect.go:190] detected "systemd" cgroup driver on host os
I1009 18:33:15.240539 281272 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I1009 18:33:15.259251 281272 binary.go:59] Skipping Kubernetes binary download due to --no-kubernetes flag
I1009 18:33:15.259380 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.9"|' /etc/containerd/config.toml"
I1009 18:33:15.271964 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I1009 18:33:15.282979 281272 containerd.go:146] configuring containerd to use "systemd" as cgroup driver...
I1009 18:33:15.283042 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = true|g' /etc/containerd/config.toml"
I1009 18:33:15.293870 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I1009 18:33:15.304902 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I1009 18:33:15.318043 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I1009 18:33:15.330292 281272 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I1009 18:33:15.340561 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I1009 18:33:15.352781 281272 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I1009 18:33:15.363348 281272 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I1009 18:33:15.373310 281272 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1009 18:33:15.469841 281272 ssh_runner.go:195] Run: sudo systemctl restart containerd
I1009 18:33:15.558603 281272 start.go:495] detecting cgroup driver to use...
I1009 18:33:15.558650 281272 detect.go:190] detected "systemd" cgroup driver on host os
I1009 18:33:15.558699 281272 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I1009 18:33:15.575205 281272 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I1009 18:33:15.589442 281272 ssh_runner.go:195] Run: sudo systemctl stop -f containerd
I1009 18:33:15.609869 281272 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I1009 18:33:15.624274 281272 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I1009 18:33:15.640218 281272 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/dockershim.sock
" | sudo tee /etc/crictl.yaml"
I1009 18:33:15.658180 281272 binary.go:59] Skipping Kubernetes binary download due to --no-kubernetes flag
I1009 18:33:15.658267 281272 ssh_runner.go:195] Run: which cri-dockerd
I1009 18:33:15.662366 281272 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I1009 18:33:15.673211 281272 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (189 bytes)
I1009 18:33:15.687836 281272 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I1009 18:33:15.811338 281272 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I1009 18:33:15.895027 281272 docker.go:575] configuring docker to use "systemd" as cgroup driver...
I1009 18:33:15.895147 281272 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (129 bytes)
I1009 18:33:15.908669 281272 ssh_runner.go:195] Run: sudo systemctl reset-failed docker
I1009 18:33:15.921005 281272 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1009 18:33:16.001315 281272 ssh_runner.go:195] Run: sudo systemctl restart docker
I1009 18:33:16.810802 281272 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
I1009 18:33:16.825293 281272 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I1009 18:33:16.852612 281272 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I1009 18:33:16.884684 281272 out.go:252] * Preparing Docker 28.5.0 ...
I1009 18:33:16.885018 281272 ssh_runner.go:195] Run: rm -f paused
I1009 18:33:16.892737 281272 out.go:179] * Done! minikube is ready without Kubernetes!
I1009 18:33:16.896066 281272 out.go:203] ╭───────────────────────────────────────────────────────────────────────────────────────╮
│ │
│ * Things to try without Kubernetes ... │
│ │
│ - "minikube ssh" to SSH into minikube's node. │
│ - "minikube docker-env" to point your docker-cli to the docker inside minikube. │
│ - "minikube image" to build images without docker. │
│ │
╰───────────────────────────────────────────────────────────────────────────────────────╯
==> Docker <==
Oct 09 18:33:15 NoKubernetes-711902 dockerd[597]: time="2025-10-09T18:33:15.602382711Z" level=info msg="Waiting for containerd to be ready to restart event processing" module=libcontainerd namespace=plugins.moby
Oct 09 18:33:16 NoKubernetes-711902 systemd[1]: Stopping docker.service - Docker Application Container Engine...
Oct 09 18:33:16 NoKubernetes-711902 dockerd[597]: time="2025-10-09T18:33:16.013280603Z" level=info msg="Processing signal 'terminated'"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[597]: time="2025-10-09T18:33:16.014470939Z" level=warning msg="Error while testing if containerd API is ready" error="Canceled: grpc: the client connection is closing"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[597]: time="2025-10-09T18:33:16.014839474Z" level=info msg="Daemon shutdown complete"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[597]: time="2025-10-09T18:33:16.014900748Z" level=warning msg="Error while testing if containerd API is ready" error="Canceled: context canceled while waiting for connections to become ready"
Oct 09 18:33:16 NoKubernetes-711902 systemd[1]: docker.service: Deactivated successfully.
Oct 09 18:33:16 NoKubernetes-711902 systemd[1]: Stopped docker.service - Docker Application Container Engine.
Oct 09 18:33:16 NoKubernetes-711902 systemd[1]: Starting docker.service - Docker Application Container Engine...
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.124911078Z" level=info msg="Starting up"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.125794051Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.125894724Z" level=info msg="CDI directory does not exist, skipping: failed to monitor for changes: no such file or directory" dir=/var/run/cdi
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.125913849Z" level=info msg="CDI directory does not exist, skipping: failed to monitor for changes: no such file or directory" dir=/etc/cdi
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.135738123Z" level=info msg="Creating a containerd client" address=/run/containerd/containerd.sock timeout=1m0s
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.138799900Z" level=info msg="[graphdriver] trying configured driver: overlay2"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.146138652Z" level=info msg="Loading containers: start."
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.767942334Z" level=info msg="Loading containers: done."
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.781483515Z" level=info msg="Docker daemon" commit=cd04830 containerd-snapshotter=false storage-driver=overlay2 version=28.5.0
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.781550857Z" level=info msg="Initializing buildkit"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.801568203Z" level=info msg="Completed buildkit initialization"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.808683377Z" level=info msg="Daemon has completed initialization"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.808817533Z" level=info msg="API listen on /run/docker.sock"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.808897097Z" level=info msg="API listen on [::]:2376"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.808824253Z" level=info msg="API listen on /var/run/docker.sock"
Oct 09 18:33:16 NoKubernetes-711902 systemd[1]: Started docker.service - Docker Application Container Engine.
==> container status <==
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
time="2025-10-09T18:33:17Z" level=fatal msg="validate service connection: validate CRI v1 runtime API for endpoint \"unix:///var/run/dockershim.sock\": rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial unix /var/run/dockershim.sock: connect: no such file or directory\""
==> describe nodes <==
command /bin/bash -c "sudo /var/lib/minikube/binaries/v0.0.0/kubectl describe nodes --kubeconfig=/var/lib/minikube/kubeconfig" failed with error: /bin/bash -c "sudo /var/lib/minikube/binaries/v0.0.0/kubectl describe nodes --kubeconfig=/var/lib/minikube/kubeconfig": Process exited with status 1
stdout:
stderr:
sudo: /var/lib/minikube/binaries/v0.0.0/kubectl: command not found
==> dmesg <==
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 0e b0 08 c2 8f ad 08 06
[Oct 9 18:05] IPv4: martian source 10.244.0.1 from 10.244.0.15, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 66 86 b6 64 71 c5 08 06
[Oct 9 18:17] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 12 a7 d3 2c b6 cd 08 06
[Oct 9 18:20] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff be a6 34 11 a0 3e 08 06
[Oct 9 18:27] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 86 7b 16 d4 1f 2a 08 06
[ +0.175207] IPv4: martian source 10.244.0.1 from 10.244.0.3, on dev eth0
[ +0.000009] ll header: 00000000: ff ff ff ff ff ff 72 f4 98 6f 82 5e 08 06
[Oct 9 18:28] IPv4: martian source 10.244.0.1 from 10.244.0.4, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 6a 05 82 5f ef 5c 08 06
[Oct 9 18:29] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 2e de 88 79 fd 2b 08 06
[Oct 9 18:31] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000009] ll header: 00000000: ff ff ff ff ff ff 12 56 f0 8a a2 50 08 06
[Oct 9 18:32] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 66 3d 78 60 14 43 08 06
[ +0.047279] IPv4: martian source 10.244.0.1 from 10.244.0.3, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 82 eb 5a ac 74 61 08 06
[ +0.016595] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff b6 84 b7 e8 55 53 08 06
[ +0.001635] IPv4: martian source 10.244.0.1 from 10.244.0.3, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 7a f1 ee 0c 72 39 08 06
==> kernel <==
18:33:17 up 1:15, 0 user, load average: 7.49, 3.01, 2.01
Linux NoKubernetes-711902 6.8.0-1041-gcp #43~22.04.1-Ubuntu SMP Wed Sep 24 23:11:19 UTC 2025 x86_64 GNU/Linux
PRETTY_NAME="Debian GNU/Linux 12 (bookworm)"
==> kubelet <==
-- No entries --
-- /stdout --
helpers_test.go:262: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p NoKubernetes-711902 -n NoKubernetes-711902
helpers_test.go:262: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.APIServer}} -p NoKubernetes-711902 -n NoKubernetes-711902: exit status 6 (297.903921ms)
-- stdout --
Stopped
WARNING: Your kubectl is pointing to stale minikube-vm.
To fix the kubectl context, run `minikube update-context`
-- /stdout --
** stderr **
E1009 18:33:18.090416 284606 status.go:458] kubeconfig endpoint: get endpoint: "NoKubernetes-711902" does not appear in /home/jenkins/minikube-integration/21139-11202/kubeconfig
** /stderr **
helpers_test.go:262: status error: exit status 6 (may be ok)
helpers_test.go:264: "NoKubernetes-711902" apiserver is not running, skipping kubectl commands (state="Stopped")
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:223: ======> post-mortem[TestNoKubernetes/serial/VerifyNok8sNoK8sDownloads]: network settings <======
helpers_test.go:230: HOST ENV snapshots: PROXY env: HTTP_PROXY="<empty>" HTTPS_PROXY="<empty>" NO_PROXY="<empty>"
helpers_test.go:238: ======> post-mortem[TestNoKubernetes/serial/VerifyNok8sNoK8sDownloads]: docker inspect <======
helpers_test.go:239: (dbg) Run: docker inspect NoKubernetes-711902
helpers_test.go:243: (dbg) docker inspect NoKubernetes-711902:
-- stdout --
[
{
"Id": "fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab",
"Created": "2025-10-09T18:33:10.603053192Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 281743,
"ExitCode": 0,
"Error": "",
"StartedAt": "2025-10-09T18:33:10.645086458Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:c6fde2176fdc734a0b1cf5396bccb3dc7d4299b26808035c9aa3b16b26946dbd",
"ResolvConfPath": "/var/lib/docker/containers/fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab/hostname",
"HostsPath": "/var/lib/docker/containers/fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab/hosts",
"LogPath": "/var/lib/docker/containers/fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab/fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab-json.log",
"Name": "/NoKubernetes-711902",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"NoKubernetes-711902:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "NoKubernetes-711902",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "private",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 3221225472,
"NanoCpus": 0,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 6442450944,
"MemorySwappiness": null,
"OomKillDisable": null,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"ID": "fdb0ba874a7d24b7acf6f88412b6ec1668c7614f1538d0e4654f749985f862ab",
"LowerDir": "/var/lib/docker/overlay2/dd5a88ca62f066a91467d9e879b36568f1b8b0752c4e706dff8c6e0c841d1ab8-init/diff:/var/lib/docker/overlay2/ef1c51768be2d023b7ddbd111b70e98dd2519a049dcc4d492c339f9603c55e42/diff",
"MergedDir": "/var/lib/docker/overlay2/dd5a88ca62f066a91467d9e879b36568f1b8b0752c4e706dff8c6e0c841d1ab8/merged",
"UpperDir": "/var/lib/docker/overlay2/dd5a88ca62f066a91467d9e879b36568f1b8b0752c4e706dff8c6e0c841d1ab8/diff",
"WorkDir": "/var/lib/docker/overlay2/dd5a88ca62f066a91467d9e879b36568f1b8b0752c4e706dff8c6e0c841d1ab8/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "NoKubernetes-711902",
"Source": "/var/lib/docker/volumes/NoKubernetes-711902/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "NoKubernetes-711902",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "NoKubernetes-711902",
"name.minikube.sigs.k8s.io": "NoKubernetes-711902",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "61282bb5119f90b697bdf1852d26e94baa57332f950b1822e4b6a969012bcd83",
"SandboxKey": "/var/run/docker/netns/61282bb5119f",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33003"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33004"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33007"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33005"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33006"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"NoKubernetes-711902": {
"IPAMConfig": {
"IPv4Address": "192.168.94.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "5a:5b:d0:94:11:ee",
"DriverOpts": null,
"GwPriority": 0,
"NetworkID": "d94f619ba0be9fc76aa2ff0b29d465f2b02b4a1aab78b268d3bf882cd9d17b28",
"EndpointID": "f089f78d38806d2c8fec53b7b6b12215669293e7f04d09926489230674a9aea7",
"Gateway": "192.168.94.1",
"IPAddress": "192.168.94.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"NoKubernetes-711902",
"fdb0ba874a7d"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p NoKubernetes-711902 -n NoKubernetes-711902
helpers_test.go:247: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.Host}} -p NoKubernetes-711902 -n NoKubernetes-711902: exit status 6 (305.073131ms)
-- stdout --
Running
WARNING: Your kubectl is pointing to stale minikube-vm.
To fix the kubectl context, run `minikube update-context`
-- /stdout --
** stderr **
E1009 18:33:18.415856 284734 status.go:458] kubeconfig endpoint: get endpoint: "NoKubernetes-711902" does not appear in /home/jenkins/minikube-integration/21139-11202/kubeconfig
** /stderr **
helpers_test.go:247: status error: exit status 6 (may be ok)
helpers_test.go:252: <<< TestNoKubernetes/serial/VerifyNok8sNoK8sDownloads FAILED: start of post-mortem logs <<<
helpers_test.go:253: ======> post-mortem[TestNoKubernetes/serial/VerifyNok8sNoK8sDownloads]: minikube logs <======
helpers_test.go:255: (dbg) Run: out/minikube-linux-amd64 -p NoKubernetes-711902 logs -n 25
helpers_test.go:260: TestNoKubernetes/serial/VerifyNok8sNoK8sDownloads logs:
-- stdout --
==> Audit <==
┌────────────┬─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┬─────────────────────────────┬──────────┬─────────┬─────────────────────┬─────────────────────┐
│ COMMAND │ ARGS │ PROFILE │ USER │ VERSION │ START TIME │ END TIME │
├────────────┼─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┼─────────────────────────────┼──────────┼─────────┼─────────────────────┼─────────────────────┤
│ stop │ -p scheduled-stop-803375 --schedule 5m │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 5m │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 5m │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 15s │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 15s │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 15s │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --cancel-scheduled │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ 09 Oct 25 18:29 UTC │
│ stop │ -p scheduled-stop-803375 --schedule 15s │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 15s │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ │
│ stop │ -p scheduled-stop-803375 --schedule 15s │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:29 UTC │ 09 Oct 25 18:30 UTC │
│ delete │ -p scheduled-stop-803375 │ scheduled-stop-803375 │ jenkins │ v1.37.0 │ 09 Oct 25 18:30 UTC │ 09 Oct 25 18:30 UTC │
│ start │ -p skaffold-381243 --memory=3072 --driver=docker --container-runtime=docker │ skaffold-381243 │ jenkins │ v1.37.0 │ 09 Oct 25 18:30 UTC │ 09 Oct 25 18:31 UTC │
│ docker-env │ --shell none -p skaffold-381243 --user=skaffold │ skaffold-381243 │ skaffold │ v1.37.0 │ 09 Oct 25 18:31 UTC │ 09 Oct 25 18:31 UTC │
│ delete │ -p skaffold-381243 │ skaffold-381243 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ 09 Oct 25 18:32 UTC │
│ start │ -p insufficient-storage-290618 --memory=3072 --output=json --wait=true --driver=docker --container-runtime=docker │ insufficient-storage-290618 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ │
│ delete │ -p insufficient-storage-290618 │ insufficient-storage-290618 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ 09 Oct 25 18:32 UTC │
│ start │ -p pause-261319 --memory=3072 --install-addons=false --wait=all --driver=docker --container-runtime=docker │ pause-261319 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ │
│ start │ -p NoKubernetes-711902 --no-kubernetes --kubernetes-version=v1.28.0 --driver=docker --container-runtime=docker │ NoKubernetes-711902 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ │
│ start │ -p offline-docker-695286 --alsologtostderr -v=1 --memory=3072 --wait=true --driver=docker --container-runtime=docker │ offline-docker-695286 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ │
│ start │ -p NoKubernetes-711902 --memory=3072 --alsologtostderr -v=5 --driver=docker --container-runtime=docker │ NoKubernetes-711902 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ 09 Oct 25 18:32 UTC │
│ start │ -p stopped-upgrade-731100 --memory=3072 --vm-driver=docker --container-runtime=docker │ stopped-upgrade-731100 │ jenkins │ v1.32.0 │ 09 Oct 25 18:32 UTC │ 09 Oct 25 18:33 UTC │
│ start │ -p NoKubernetes-711902 --no-kubernetes --memory=3072 --alsologtostderr -v=5 --driver=docker --container-runtime=docker │ NoKubernetes-711902 │ jenkins │ v1.37.0 │ 09 Oct 25 18:32 UTC │ 09 Oct 25 18:33 UTC │
│ delete │ -p NoKubernetes-711902 │ NoKubernetes-711902 │ jenkins │ v1.37.0 │ 09 Oct 25 18:33 UTC │ 09 Oct 25 18:33 UTC │
│ start │ -p NoKubernetes-711902 --no-kubernetes --memory=3072 --alsologtostderr -v=5 --driver=docker --container-runtime=docker │ NoKubernetes-711902 │ jenkins │ v1.37.0 │ 09 Oct 25 18:33 UTC │ 09 Oct 25 18:33 UTC │
│ stop │ stopped-upgrade-731100 stop │ stopped-upgrade-731100 │ jenkins │ v1.32.0 │ 09 Oct 25 18:33 UTC │ │
└────────────┴─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┴─────────────────────────────┴──────────┴─────────┴─────────────────────┴─────────────────────┘
==> Last Start <==
Log file created at: 2025/10/09 18:33:09
Running on machine: ubuntu-20-agent
Binary: Built with gc go1.24.6 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I1009 18:33:09.726820 281272 out.go:360] Setting OutFile to fd 1 ...
I1009 18:33:09.727079 281272 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1009 18:33:09.727087 281272 out.go:374] Setting ErrFile to fd 2...
I1009 18:33:09.727091 281272 out.go:408] TERM=,COLORTERM=, which probably does not support color
I1009 18:33:09.727313 281272 root.go:338] Updating PATH: /home/jenkins/minikube-integration/21139-11202/.minikube/bin
I1009 18:33:09.727812 281272 out.go:368] Setting JSON to false
I1009 18:33:09.729162 281272 start.go:131] hostinfo: {"hostname":"ubuntu-20-agent","uptime":4531,"bootTime":1760030259,"procs":346,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"22.04","kernelVersion":"6.8.0-1041-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I1009 18:33:09.729247 281272 start.go:141] virtualization: kvm guest
I1009 18:33:09.731346 281272 out.go:179] * [NoKubernetes-711902] minikube v1.37.0 on Ubuntu 22.04 (kvm/amd64)
I1009 18:33:09.732706 281272 notify.go:220] Checking for updates...
I1009 18:33:09.732755 281272 out.go:179] - MINIKUBE_LOCATION=21139
I1009 18:33:09.734187 281272 out.go:179] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I1009 18:33:09.735610 281272 out.go:179] - KUBECONFIG=/home/jenkins/minikube-integration/21139-11202/kubeconfig
I1009 18:33:09.736931 281272 out.go:179] - MINIKUBE_HOME=/home/jenkins/minikube-integration/21139-11202/.minikube
I1009 18:33:09.738081 281272 out.go:179] - MINIKUBE_BIN=out/minikube-linux-amd64
I1009 18:33:09.739247 281272 out.go:179] - MINIKUBE_FORCE_SYSTEMD=
I1009 18:33:09.741194 281272 config.go:182] Loaded profile config "offline-docker-695286": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.34.1
I1009 18:33:09.741378 281272 config.go:182] Loaded profile config "pause-261319": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.34.1
I1009 18:33:09.741509 281272 config.go:182] Loaded profile config "stopped-upgrade-731100": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.28.3
I1009 18:33:09.741540 281272 start.go:1899] No Kubernetes flag is set, setting Kubernetes version to v0.0.0
I1009 18:33:09.741667 281272 driver.go:421] Setting default libvirt URI to qemu:///system
I1009 18:33:09.769435 281272 docker.go:123] docker version: linux-28.5.1:Docker Engine - Community
I1009 18:33:09.769662 281272 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1009 18:33:09.828796 281272 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:3 ContainersRunning:3 ContainersPaused:0 ContainersStopped:0 Images:4 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:false CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:64 OomKillDisable:false NGoroutines:76 SystemTime:2025-10-09 18:33:09.818416947 +0000 UTC LoggingDriver:json-file CgroupDriver:systemd NEventsListener:0 KernelVersion:6.8.0-1041-gcp OperatingSystem:Ubuntu 22.04.5 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33652174848 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent Labels:[] ExperimentalBuild:false ServerVersion:28.5.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:b98a3aace656320842a23f4a392a33f46af97866 Expected:} RuncCommit:{ID:v1.3.0-0-g4ca628d1 Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin name=cgroupns] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map
[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.29.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.40.0] map[Name:model Path:/usr/libexec/docker/cli-plugins/docker-model SchemaVersion:0.1.0 ShortDescription:Docker Model Runner Vendor:Docker Inc. Version:v0.1.42] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I1009 18:33:09.828892 281272 docker.go:318] overlay module found
I1009 18:33:09.830759 281272 out.go:179] * Using the docker driver based on user configuration
I1009 18:33:09.832210 281272 start.go:305] selected driver: docker
I1009 18:33:09.832225 281272 start.go:925] validating driver "docker" against <nil>
I1009 18:33:09.832236 281272 start.go:936] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I1009 18:33:09.832801 281272 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1009 18:33:09.893498 281272 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:3 ContainersRunning:3 ContainersPaused:0 ContainersStopped:0 Images:4 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:false CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:64 OomKillDisable:false NGoroutines:76 SystemTime:2025-10-09 18:33:09.883713298 +0000 UTC LoggingDriver:json-file CgroupDriver:systemd NEventsListener:0 KernelVersion:6.8.0-1041-gcp OperatingSystem:Ubuntu 22.04.5 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[::1/128 127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33652174848 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent Labels:[] ExperimentalBuild:false ServerVersion:28.5.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:b98a3aace656320842a23f4a392a33f46af97866 Expected:} RuncCommit:{ID:v1.3.0-0-g4ca628d1 Expected:} InitCommit:{ID:de40ad0 Expected:} SecurityOptions:[name=apparmor name=seccomp,profile=builtin name=cgroupns] ProductLicense: Warnings:<nil> ServerErrors:[] ClientInfo:{Debug:false Plugins:[map
[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.29.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.40.0] map[Name:model Path:/usr/libexec/docker/cli-plugins/docker-model SchemaVersion:0.1.0 ShortDescription:Docker Model Runner Vendor:Docker Inc. Version:v0.1.42] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I1009 18:33:09.893624 281272 start.go:1899] No Kubernetes flag is set, setting Kubernetes version to v0.0.0
I1009 18:33:09.893695 281272 start_flags.go:327] no existing cluster config was found, will generate one from the flags
I1009 18:33:09.893889 281272 start_flags.go:974] Wait components to verify : map[apiserver:true system_pods:true]
I1009 18:33:09.895859 281272 out.go:179] * Using Docker driver with root privileges
I1009 18:33:09.897317 281272 cni.go:84] Creating CNI manager for ""
I1009 18:33:09.897387 281272 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I1009 18:33:09.897398 281272 start_flags.go:336] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I1009 18:33:09.897423 281272 start.go:1899] No Kubernetes flag is set, setting Kubernetes version to v0.0.0
I1009 18:33:09.897472 281272 start.go:349] cluster config:
{Name:NoKubernetes-711902 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v0.0.0 ClusterName:NoKubernetes-711902 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local Conta
inerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v0.0.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I1009 18:33:09.898767 281272 out.go:179] * Starting minikube without Kubernetes in cluster NoKubernetes-711902
I1009 18:33:09.899911 281272 cache.go:133] Beginning downloading kic base image for docker with docker
I1009 18:33:09.901159 281272 out.go:179] * Pulling base image v0.0.48-1759745255-21703 ...
I1009 18:33:09.902378 281272 cache.go:58] Skipping Kubernetes image caching due to --no-kubernetes flag
I1009 18:33:09.902483 281272 image.go:81] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92 in local docker daemon
I1009 18:33:09.902514 281272 profile.go:143] Saving config to /home/jenkins/minikube-integration/21139-11202/.minikube/profiles/NoKubernetes-711902/config.json ...
I1009 18:33:09.902552 281272 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/21139-11202/.minikube/profiles/NoKubernetes-711902/config.json: {Name:mk7ed4d136dae909a52d53e49ef555ca81eefd7b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1009 18:33:09.924695 281272 image.go:100] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92 in local docker daemon, skipping pull
I1009 18:33:09.924718 281272 cache.go:157] gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92 exists in daemon, skipping load
I1009 18:33:09.924738 281272 cache.go:242] Successfully downloaded all kic artifacts
I1009 18:33:09.924765 281272 start.go:360] acquireMachinesLock for NoKubernetes-711902: {Name:mk77177cbada15c1cc71ca39efd4970ee070e38a Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I1009 18:33:09.924827 281272 start.go:364] duration metric: took 46.818µs to acquireMachinesLock for "NoKubernetes-711902"
I1009 18:33:09.924851 281272 start.go:93] Provisioning new machine with config: &{Name:NoKubernetes-711902 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92 Memory:3072 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v0.0.0 ClusterName:NoKubernetes-711902 Namespace:default APIServerHAVIP: APISer
verName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v0.0.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s MountString: Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false DisableCoreDNSLog:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSoc
k: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v0.0.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I1009 18:33:09.925024 281272 start.go:125] createHost starting for "" (driver="docker")
I1009 18:33:12.931630 265579 kubeadm.go:322] [apiclient] All control plane components are healthy after 5.502455 seconds
I1009 18:33:12.931770 265579 kubeadm.go:322] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I1009 18:33:12.945750 265579 kubeadm.go:322] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I1009 18:33:13.465875 265579 kubeadm.go:322] [upload-certs] Skipping phase. Please see --upload-certs
I1009 18:33:13.466133 265579 kubeadm.go:322] [mark-control-plane] Marking the node stopped-upgrade-731100 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I1009 18:33:13.975633 265579 kubeadm.go:322] [bootstrap-token] Using token: 2jh3nb.jrabffh7nvf4snwz
I1009 18:33:13.977481 265579 out.go:204] - Configuring RBAC rules ...
I1009 18:33:13.977640 265579 kubeadm.go:322] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I1009 18:33:13.981530 265579 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I1009 18:33:13.989295 265579 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I1009 18:33:13.992957 265579 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I1009 18:33:13.995784 265579 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I1009 18:33:13.998990 265579 kubeadm.go:322] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I1009 18:33:14.011058 265579 kubeadm.go:322] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I1009 18:33:14.185848 265579 kubeadm.go:322] [addons] Applied essential addon: CoreDNS
I1009 18:33:14.386651 265579 kubeadm.go:322] [addons] Applied essential addon: kube-proxy
I1009 18:33:14.387580 265579 kubeadm.go:322]
I1009 18:33:14.387688 265579 kubeadm.go:322] Your Kubernetes control-plane has initialized successfully!
I1009 18:33:14.387693 265579 kubeadm.go:322]
I1009 18:33:14.387858 265579 kubeadm.go:322] To start using your cluster, you need to run the following as a regular user:
I1009 18:33:14.387865 265579 kubeadm.go:322]
I1009 18:33:14.387898 265579 kubeadm.go:322] mkdir -p $HOME/.kube
I1009 18:33:14.387962 265579 kubeadm.go:322] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I1009 18:33:14.388049 265579 kubeadm.go:322] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I1009 18:33:14.388061 265579 kubeadm.go:322]
I1009 18:33:14.388132 265579 kubeadm.go:322] Alternatively, if you are the root user, you can run:
I1009 18:33:14.388137 265579 kubeadm.go:322]
I1009 18:33:14.388199 265579 kubeadm.go:322] export KUBECONFIG=/etc/kubernetes/admin.conf
I1009 18:33:14.388205 265579 kubeadm.go:322]
I1009 18:33:14.388304 265579 kubeadm.go:322] You should now deploy a pod network to the cluster.
I1009 18:33:14.388408 265579 kubeadm.go:322] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I1009 18:33:14.388497 265579 kubeadm.go:322] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I1009 18:33:14.388503 265579 kubeadm.go:322]
I1009 18:33:14.388647 265579 kubeadm.go:322] You can now join any number of control-plane nodes by copying certificate authorities
I1009 18:33:14.388750 265579 kubeadm.go:322] and service account keys on each node and then running the following as root:
I1009 18:33:14.388758 265579 kubeadm.go:322]
I1009 18:33:14.388898 265579 kubeadm.go:322] kubeadm join control-plane.minikube.internal:8443 --token 2jh3nb.jrabffh7nvf4snwz \
I1009 18:33:14.389022 265579 kubeadm.go:322] --discovery-token-ca-cert-hash sha256:50b267e6e7636e43468b46774c69e9781b208d8d6164a67ebb20e03119db0f84 \
I1009 18:33:14.389060 265579 kubeadm.go:322] --control-plane
I1009 18:33:14.389070 265579 kubeadm.go:322]
I1009 18:33:14.389201 265579 kubeadm.go:322] Then you can join any number of worker nodes by running the following on each as root:
I1009 18:33:14.389213 265579 kubeadm.go:322]
I1009 18:33:14.389319 265579 kubeadm.go:322] kubeadm join control-plane.minikube.internal:8443 --token 2jh3nb.jrabffh7nvf4snwz \
I1009 18:33:14.389471 265579 kubeadm.go:322] --discovery-token-ca-cert-hash sha256:50b267e6e7636e43468b46774c69e9781b208d8d6164a67ebb20e03119db0f84
I1009 18:33:14.391453 265579 kubeadm.go:322] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/6.8.0-1041-gcp\n", err: exit status 1
I1009 18:33:14.391574 265579 kubeadm.go:322] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I1009 18:33:14.391614 265579 cni.go:84] Creating CNI manager for ""
I1009 18:33:14.391631 265579 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I1009 18:33:14.393306 265579 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I1009 18:33:14.394495 265579 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I1009 18:33:14.405223 265579 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (457 bytes)
I1009 18:33:14.425689 265579 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I1009 18:33:14.425804 265579 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.3/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I1009 18:33:14.425816 265579 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.3/kubectl label nodes minikube.k8s.io/version=v1.32.0 minikube.k8s.io/commit=8220a6eb95f0a4d75f7f2d7b14cef975f050512d minikube.k8s.io/name=stopped-upgrade-731100 minikube.k8s.io/updated_at=2025_10_09T18_33_14_0700 minikube.k8s.io/primary=true --all --overwrite --kubeconfig=/var/lib/minikube/kubeconfig
I1009 18:33:14.434356 265579 ops.go:34] apiserver oom_adj: -16
I1009 18:33:14.518361 265579 kubeadm.go:1081] duration metric: took 92.625873ms to wait for elevateKubeSystemPrivileges.
I1009 18:33:14.518397 265579 kubeadm.go:406] StartCluster complete in 10.424255465s
I1009 18:33:14.518419 265579 settings.go:142] acquiring lock: {Name:mk068a134d3e4d40102c45be73d7211efe819775 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1009 18:33:14.518544 265579 settings.go:150] Updating kubeconfig: /tmp/legacy_kubeconfig880889346
I1009 18:33:14.519175 265579 lock.go:35] WriteFile acquiring /tmp/legacy_kubeconfig880889346: {Name:mk211495ca79b7e4f04aeeeac75527902ecb4d94 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1009 18:33:14.519497 265579 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I1009 18:33:14.519501 265579 addons.go:499] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:false csi-hostpath-driver:false dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:false gvisor:false headlamp:false helm-tiller:false inaccel:false ingress:false ingress-dns:false inspektor-gadget:false istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:false nvidia-device-plugin:false nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:false registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:false volumesnapshots:false]
I1009 18:33:14.519578 265579 addons.go:69] Setting storage-provisioner=true in profile "stopped-upgrade-731100"
I1009 18:33:14.519580 265579 addons.go:69] Setting default-storageclass=true in profile "stopped-upgrade-731100"
I1009 18:33:14.519622 265579 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "stopped-upgrade-731100"
I1009 18:33:14.519640 265579 addons.go:231] Setting addon storage-provisioner=true in "stopped-upgrade-731100"
I1009 18:33:14.519706 265579 host.go:66] Checking if "stopped-upgrade-731100" exists ...
I1009 18:33:14.519799 265579 config.go:182] Loaded profile config "stopped-upgrade-731100": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.28.3
I1009 18:33:14.520071 265579 cli_runner.go:164] Run: docker container inspect stopped-upgrade-731100 --format={{.State.Status}}
I1009 18:33:14.520240 265579 cli_runner.go:164] Run: docker container inspect stopped-upgrade-731100 --format={{.State.Status}}
I1009 18:33:14.541424 265579 kapi.go:248] "coredns" deployment in "kube-system" namespace and "stopped-upgrade-731100" context rescaled to 1 replicas
I1009 18:33:14.541458 265579 start.go:223] Will wait 6m0s for node &{Name: IP:192.168.103.2 Port:8443 KubernetesVersion:v1.28.3 ContainerRuntime:docker ControlPlane:true Worker:true}
I1009 18:33:14.542920 265579 out.go:177] * Verifying Kubernetes components...
I1009 18:33:14.544123 265579 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I1009 18:33:14.545665 265579 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I1009 18:33:09.814481 264343 pod_ready.go:99] pod "coredns-66bc5c9577-xfh69" in "kube-system" namespace is gone: getting pod "coredns-66bc5c9577-xfh69" in "kube-system" namespace (will retry): pods "coredns-66bc5c9577-xfh69" not found
I1009 18:33:09.814496 264343 pod_ready.go:86] duration metric: took 11.002870301s for pod "coredns-66bc5c9577-xfh69" in "kube-system" namespace to be "Ready" or be gone ...
I1009 18:33:09.814506 264343 pod_ready.go:83] waiting for pod "coredns-66bc5c9577-xqpvs" in "kube-system" namespace to be "Ready" or be gone ...
W1009 18:33:11.819352 264343 pod_ready.go:104] pod "coredns-66bc5c9577-xqpvs" is not "Ready", error: <nil>
W1009 18:33:13.821852 264343 pod_ready.go:104] pod "coredns-66bc5c9577-xqpvs" is not "Ready", error: <nil>
W1009 18:33:12.156435 264341 pod_ready.go:104] pod "coredns-66bc5c9577-qzfd8" is not "Ready", error: <nil>
W1009 18:33:14.157176 264341 pod_ready.go:104] pod "coredns-66bc5c9577-qzfd8" is not "Ready", error: <nil>
I1009 18:33:09.926837 281272 out.go:252] * Creating docker container (CPUs=2, Memory=3072MB) ...
I1009 18:33:09.927080 281272 start.go:159] libmachine.API.Create for "NoKubernetes-711902" (driver="docker")
I1009 18:33:09.927121 281272 client.go:168] LocalClient.Create starting
I1009 18:33:09.927172 281272 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem
I1009 18:33:09.927204 281272 main.go:141] libmachine: Decoding PEM data...
I1009 18:33:09.927219 281272 main.go:141] libmachine: Parsing certificate...
I1009 18:33:09.927274 281272 main.go:141] libmachine: Reading certificate data from /home/jenkins/minikube-integration/21139-11202/.minikube/certs/cert.pem
I1009 18:33:09.927294 281272 main.go:141] libmachine: Decoding PEM data...
I1009 18:33:09.927310 281272 main.go:141] libmachine: Parsing certificate...
I1009 18:33:09.927670 281272 cli_runner.go:164] Run: docker network inspect NoKubernetes-711902 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W1009 18:33:09.945553 281272 cli_runner.go:211] docker network inspect NoKubernetes-711902 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I1009 18:33:09.945659 281272 network_create.go:284] running [docker network inspect NoKubernetes-711902] to gather additional debugging logs...
I1009 18:33:09.945687 281272 cli_runner.go:164] Run: docker network inspect NoKubernetes-711902
W1009 18:33:09.963224 281272 cli_runner.go:211] docker network inspect NoKubernetes-711902 returned with exit code 1
I1009 18:33:09.963257 281272 network_create.go:287] error running [docker network inspect NoKubernetes-711902]: docker network inspect NoKubernetes-711902: exit status 1
stdout:
[]
stderr:
Error response from daemon: network NoKubernetes-711902 not found
I1009 18:33:09.963289 281272 network_create.go:289] output of [docker network inspect NoKubernetes-711902]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network NoKubernetes-711902 not found
** /stderr **
I1009 18:33:09.963420 281272 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I1009 18:33:09.982077 281272 network.go:211] skipping subnet 192.168.49.0/24 that is taken: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName:br-36d16f1d4123 IfaceIPv4:192.168.49.1 IfaceMTU:1500 IfaceMAC:0a:9a:9a:77:cb:7c} reservation:<nil>}
I1009 18:33:09.982636 281272 network.go:211] skipping subnet 192.168.58.0/24 that is taken: &{IP:192.168.58.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.58.0/24 Gateway:192.168.58.1 ClientMin:192.168.58.2 ClientMax:192.168.58.254 Broadcast:192.168.58.255 IsPrivate:true Interface:{IfaceName:br-8adbd295428d IfaceIPv4:192.168.58.1 IfaceMTU:1500 IfaceMAC:22:1a:8a:aa:f1:33} reservation:<nil>}
I1009 18:33:09.983150 281272 network.go:211] skipping subnet 192.168.67.0/24 that is taken: &{IP:192.168.67.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.67.0/24 Gateway:192.168.67.1 ClientMin:192.168.67.2 ClientMax:192.168.67.254 Broadcast:192.168.67.255 IsPrivate:true Interface:{IfaceName:br-9d29dc29aaeb IfaceIPv4:192.168.67.1 IfaceMTU:1500 IfaceMAC:a2:d7:08:db:e1:fb} reservation:<nil>}
I1009 18:33:09.983716 281272 network.go:211] skipping subnet 192.168.76.0/24 that is taken: &{IP:192.168.76.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.76.0/24 Gateway:192.168.76.1 ClientMin:192.168.76.2 ClientMax:192.168.76.254 Broadcast:192.168.76.255 IsPrivate:true Interface:{IfaceName:br-9c468df904f3 IfaceIPv4:192.168.76.1 IfaceMTU:1500 IfaceMAC:0a:e0:89:50:00:c6} reservation:<nil>}
I1009 18:33:09.984371 281272 network.go:211] skipping subnet 192.168.85.0/24 that is taken: &{IP:192.168.85.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.85.0/24 Gateway:192.168.85.1 ClientMin:192.168.85.2 ClientMax:192.168.85.254 Broadcast:192.168.85.255 IsPrivate:true Interface:{IfaceName:br-b0a9a97c4749 IfaceIPv4:192.168.85.1 IfaceMTU:1500 IfaceMAC:4e:1e:0c:e6:fc:72} reservation:<nil>}
I1009 18:33:09.985210 281272 network.go:206] using free private subnet 192.168.94.0/24: &{IP:192.168.94.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.94.0/24 Gateway:192.168.94.1 ClientMin:192.168.94.2 ClientMax:192.168.94.254 Broadcast:192.168.94.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc00205ad10}
I1009 18:33:09.985234 281272 network_create.go:124] attempt to create docker network NoKubernetes-711902 192.168.94.0/24 with gateway 192.168.94.1 and MTU of 1500 ...
I1009 18:33:09.985291 281272 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.94.0/24 --gateway=192.168.94.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=NoKubernetes-711902 NoKubernetes-711902
I1009 18:33:10.046655 281272 network_create.go:108] docker network NoKubernetes-711902 192.168.94.0/24 created
I1009 18:33:10.046710 281272 kic.go:121] calculated static IP "192.168.94.2" for the "NoKubernetes-711902" container
I1009 18:33:10.046804 281272 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I1009 18:33:10.065918 281272 cli_runner.go:164] Run: docker volume create NoKubernetes-711902 --label name.minikube.sigs.k8s.io=NoKubernetes-711902 --label created_by.minikube.sigs.k8s.io=true
I1009 18:33:10.086849 281272 oci.go:103] Successfully created a docker volume NoKubernetes-711902
I1009 18:33:10.086930 281272 cli_runner.go:164] Run: docker run --rm --name NoKubernetes-711902-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=NoKubernetes-711902 --entrypoint /usr/bin/test -v NoKubernetes-711902:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92 -d /var/lib
I1009 18:33:10.518077 281272 oci.go:107] Successfully prepared a docker volume NoKubernetes-711902
I1009 18:33:10.518163 281272 preload.go:178] Skipping preload logic due to --no-kubernetes flag
W1009 18:33:10.518272 281272 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
W1009 18:33:10.518314 281272 oci.go:252] Your kernel does not support CPU cfs period/quota or the cgroup is not mounted.
I1009 18:33:10.518359 281272 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I1009 18:33:10.582290 281272 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname NoKubernetes-711902 --name NoKubernetes-711902 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=NoKubernetes-711902 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=NoKubernetes-711902 --network NoKubernetes-711902 --ip 192.168.94.2 --volume NoKubernetes-711902:/var --security-opt apparmor=unconfined --memory=3072mb -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.48-1759745255-21703@sha256:cb5cd2ea26aaf2d64a5ec385670af2f770e759461e4b662fd7a8fae305b74c92
I1009 18:33:10.896323 281272 cli_runner.go:164] Run: docker container inspect NoKubernetes-711902 --format={{.State.Running}}
I1009 18:33:10.918963 281272 cli_runner.go:164] Run: docker container inspect NoKubernetes-711902 --format={{.State.Status}}
I1009 18:33:10.938118 281272 cli_runner.go:164] Run: docker exec NoKubernetes-711902 stat /var/lib/dpkg/alternatives/iptables
I1009 18:33:10.992503 281272 oci.go:144] the created container "NoKubernetes-711902" has a running status.
I1009 18:33:10.992533 281272 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa...
I1009 18:33:11.562080 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa.pub -> /home/docker/.ssh/authorized_keys
I1009 18:33:11.562135 281272 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I1009 18:33:11.588890 281272 cli_runner.go:164] Run: docker container inspect NoKubernetes-711902 --format={{.State.Status}}
I1009 18:33:11.607636 281272 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I1009 18:33:11.607661 281272 kic_runner.go:114] Args: [docker exec --privileged NoKubernetes-711902 chown docker:docker /home/docker/.ssh/authorized_keys]
I1009 18:33:11.652898 281272 cli_runner.go:164] Run: docker container inspect NoKubernetes-711902 --format={{.State.Status}}
I1009 18:33:11.672707 281272 machine.go:93] provisionDockerMachine start ...
I1009 18:33:11.672817 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:11.691520 281272 main.go:141] libmachine: Using SSH client type: native
I1009 18:33:11.691795 281272 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x840040] 0x842d40 <nil> [] 0s} 127.0.0.1 33003 <nil> <nil>}
I1009 18:33:11.691817 281272 main.go:141] libmachine: About to run SSH command:
hostname
I1009 18:33:11.837755 281272 main.go:141] libmachine: SSH cmd err, output: <nil>: NoKubernetes-711902
I1009 18:33:11.837785 281272 ubuntu.go:182] provisioning hostname "NoKubernetes-711902"
I1009 18:33:11.837858 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:11.858762 281272 main.go:141] libmachine: Using SSH client type: native
I1009 18:33:11.859058 281272 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x840040] 0x842d40 <nil> [] 0s} 127.0.0.1 33003 <nil> <nil>}
I1009 18:33:11.859077 281272 main.go:141] libmachine: About to run SSH command:
sudo hostname NoKubernetes-711902 && echo "NoKubernetes-711902" | sudo tee /etc/hostname
I1009 18:33:12.017403 281272 main.go:141] libmachine: SSH cmd err, output: <nil>: NoKubernetes-711902
I1009 18:33:12.017483 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:12.034992 281272 main.go:141] libmachine: Using SSH client type: native
I1009 18:33:12.035252 281272 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x840040] 0x842d40 <nil> [] 0s} 127.0.0.1 33003 <nil> <nil>}
I1009 18:33:12.035272 281272 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\sNoKubernetes-711902' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 NoKubernetes-711902/g' /etc/hosts;
else
echo '127.0.1.1 NoKubernetes-711902' | sudo tee -a /etc/hosts;
fi
fi
I1009 18:33:12.184962 281272 main.go:141] libmachine: SSH cmd err, output: <nil>:
I1009 18:33:12.185000 281272 ubuntu.go:188] set auth options {CertDir:/home/jenkins/minikube-integration/21139-11202/.minikube CaCertPath:/home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/21139-11202/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/21139-11202/.minikube}
I1009 18:33:12.185034 281272 ubuntu.go:190] setting up certificates
I1009 18:33:12.185047 281272 provision.go:84] configureAuth start
I1009 18:33:12.185114 281272 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" NoKubernetes-711902
I1009 18:33:12.205934 281272 provision.go:143] copyHostCerts
I1009 18:33:12.205971 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem -> /home/jenkins/minikube-integration/21139-11202/.minikube/ca.pem
I1009 18:33:12.206006 281272 exec_runner.go:144] found /home/jenkins/minikube-integration/21139-11202/.minikube/ca.pem, removing ...
I1009 18:33:12.206019 281272 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21139-11202/.minikube/ca.pem
I1009 18:33:12.206113 281272 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/21139-11202/.minikube/ca.pem (1078 bytes)
I1009 18:33:12.206228 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/cert.pem -> /home/jenkins/minikube-integration/21139-11202/.minikube/cert.pem
I1009 18:33:12.206256 281272 exec_runner.go:144] found /home/jenkins/minikube-integration/21139-11202/.minikube/cert.pem, removing ...
I1009 18:33:12.206263 281272 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21139-11202/.minikube/cert.pem
I1009 18:33:12.206311 281272 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/21139-11202/.minikube/cert.pem (1123 bytes)
I1009 18:33:12.206381 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/key.pem -> /home/jenkins/minikube-integration/21139-11202/.minikube/key.pem
I1009 18:33:12.206406 281272 exec_runner.go:144] found /home/jenkins/minikube-integration/21139-11202/.minikube/key.pem, removing ...
I1009 18:33:12.206412 281272 exec_runner.go:203] rm: /home/jenkins/minikube-integration/21139-11202/.minikube/key.pem
I1009 18:33:12.206448 281272 exec_runner.go:151] cp: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/21139-11202/.minikube/key.pem (1675 bytes)
I1009 18:33:12.206526 281272 provision.go:117] generating server cert: /home/jenkins/minikube-integration/21139-11202/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca-key.pem org=jenkins.NoKubernetes-711902 san=[127.0.0.1 192.168.94.2 NoKubernetes-711902 localhost minikube]
I1009 18:33:12.882770 281272 provision.go:177] copyRemoteCerts
I1009 18:33:12.882824 281272 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I1009 18:33:12.882859 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:12.901749 281272 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33003 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa Username:docker}
I1009 18:33:13.007807 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem -> /etc/docker/ca.pem
I1009 18:33:13.007877 281272 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21139-11202/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I1009 18:33:13.028112 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/machines/server.pem -> /etc/docker/server.pem
I1009 18:33:13.028171 281272 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21139-11202/.minikube/machines/server.pem --> /etc/docker/server.pem (1224 bytes)
I1009 18:33:13.046902 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/machines/server-key.pem -> /etc/docker/server-key.pem
I1009 18:33:13.046959 281272 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21139-11202/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I1009 18:33:13.065410 281272 provision.go:87] duration metric: took 880.347046ms to configureAuth
I1009 18:33:13.065435 281272 ubuntu.go:206] setting minikube options for container-runtime
I1009 18:33:13.065668 281272 config.go:182] Loaded profile config "NoKubernetes-711902": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v0.0.0
I1009 18:33:13.065740 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:13.084447 281272 main.go:141] libmachine: Using SSH client type: native
I1009 18:33:13.084715 281272 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x840040] 0x842d40 <nil> [] 0s} 127.0.0.1 33003 <nil> <nil>}
I1009 18:33:13.084730 281272 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I1009 18:33:13.232775 281272 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I1009 18:33:13.232799 281272 ubuntu.go:71] root file system type: overlay
I1009 18:33:13.232968 281272 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I1009 18:33:13.233039 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:13.254715 281272 main.go:141] libmachine: Using SSH client type: native
I1009 18:33:13.254967 281272 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x840040] 0x842d40 <nil> [] 0s} 127.0.0.1 33003 <nil> <nil>}
I1009 18:33:13.255043 281272 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network-online.target nss-lookup.target docker.socket firewalld.service containerd.service time-set.target
Wants=network-online.target containerd.service
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=always
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 \
-H fd:// --containerd=/run/containerd/containerd.sock \
-H unix:///var/run/docker.sock \
--default-ulimit=nofile=1048576:1048576 \
--tlsverify \
--tlscacert /etc/docker/ca.pem \
--tlscert /etc/docker/server.pem \
--tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I1009 18:33:13.424326 281272 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network-online.target nss-lookup.target docker.socket firewalld.service containerd.service time-set.target
Wants=network-online.target containerd.service
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=always
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H fd:// --containerd=/run/containerd/containerd.sock -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
I1009 18:33:13.424393 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:13.444310 281272 main.go:141] libmachine: Using SSH client type: native
I1009 18:33:13.444506 281272 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x840040] 0x842d40 <nil> [] 0s} 127.0.0.1 33003 <nil> <nil>}
I1009 18:33:13.444527 281272 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I1009 18:33:14.654138 281272 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2025-10-02 14:52:52.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2025-10-09 18:33:13.422146133 +0000
@@ -9,23 +9,34 @@
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
Restart=always
+
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H fd:// --containerd=/run/containerd/containerd.sock -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
+
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I1009 18:33:14.654174 281272 machine.go:96] duration metric: took 2.981445688s to provisionDockerMachine
I1009 18:33:14.654187 281272 client.go:171] duration metric: took 4.727059425s to LocalClient.Create
I1009 18:33:14.654212 281272 start.go:167] duration metric: took 4.727134072s to libmachine.API.Create "NoKubernetes-711902"
I1009 18:33:14.654220 281272 start.go:293] postStartSetup for "NoKubernetes-711902" (driver="docker")
I1009 18:33:14.654232 281272 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I1009 18:33:14.654295 281272 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I1009 18:33:14.654345 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:14.680894 281272 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33003 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa Username:docker}
I1009 18:33:14.544979 265579 addons.go:231] Setting addon default-storageclass=true in "stopped-upgrade-731100"
I1009 18:33:14.547744 265579 addons.go:423] installing /etc/kubernetes/addons/storage-provisioner.yaml
I1009 18:33:14.547759 265579 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I1009 18:33:14.547758 265579 host.go:66] Checking if "stopped-upgrade-731100" exists ...
I1009 18:33:14.547816 265579 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" stopped-upgrade-731100
I1009 18:33:14.548246 265579 cli_runner.go:164] Run: docker container inspect stopped-upgrade-731100 --format={{.State.Status}}
I1009 18:33:14.569783 265579 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32998 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/stopped-upgrade-731100/id_rsa Username:docker}
I1009 18:33:14.570364 265579 addons.go:423] installing /etc/kubernetes/addons/storageclass.yaml
I1009 18:33:14.570376 265579 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I1009 18:33:14.570431 265579 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" stopped-upgrade-731100
I1009 18:33:14.595217 265579 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32998 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/stopped-upgrade-731100/id_rsa Username:docker}
I1009 18:33:14.604074 265579 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.103.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.28.3/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I1009 18:33:14.604933 265579 api_server.go:52] waiting for apiserver process to appear ...
I1009 18:33:14.604971 265579 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I1009 18:33:14.680524 265579 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.3/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I1009 18:33:14.703459 265579 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.3/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I1009 18:33:15.256158 265579 start.go:926] {"host.minikube.internal": 192.168.103.1} host record injected into CoreDNS's ConfigMap
I1009 18:33:15.256230 265579 api_server.go:72] duration metric: took 714.740057ms to wait for apiserver process to appear ...
I1009 18:33:15.256247 265579 api_server.go:88] waiting for apiserver healthz status ...
I1009 18:33:15.256267 265579 api_server.go:253] Checking apiserver healthz at https://192.168.103.2:8443/healthz ...
I1009 18:33:15.262308 265579 api_server.go:279] https://192.168.103.2:8443/healthz returned 200:
ok
I1009 18:33:15.263824 265579 api_server.go:141] control plane version: v1.28.3
I1009 18:33:15.263841 265579 api_server.go:131] duration metric: took 7.588233ms to wait for apiserver health ...
I1009 18:33:15.263848 265579 system_pods.go:43] waiting for kube-system pods to appear ...
I1009 18:33:15.271088 265579 system_pods.go:59] 4 kube-system pods found
I1009 18:33:15.271113 265579 system_pods.go:61] "etcd-stopped-upgrade-731100" [054d6170-1a85-4625-8037-37ebefcc4574] Pending
I1009 18:33:15.271124 265579 system_pods.go:61] "kube-apiserver-stopped-upgrade-731100" [899ed27c-d0e7-4d8e-a4ee-15298ddee726] Pending
I1009 18:33:15.271130 265579 system_pods.go:61] "kube-controller-manager-stopped-upgrade-731100" [a7f2a9ed-be6b-4e22-b4da-1305a3e56228] Pending
I1009 18:33:15.271136 265579 system_pods.go:61] "kube-scheduler-stopped-upgrade-731100" [8ec66241-649a-46c8-9e47-bf861cb836f9] Pending
I1009 18:33:15.271143 265579 system_pods.go:74] duration metric: took 7.288345ms to wait for pod list to return data ...
I1009 18:33:15.271152 265579 kubeadm.go:581] duration metric: took 729.665151ms to wait for : map[apiserver:true system_pods:true] ...
I1009 18:33:15.271165 265579 node_conditions.go:102] verifying NodePressure condition ...
I1009 18:33:15.275042 265579 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I1009 18:33:15.275060 265579 node_conditions.go:123] node cpu capacity is 8
I1009 18:33:15.275075 265579 node_conditions.go:105] duration metric: took 3.904965ms to run NodePressure ...
I1009 18:33:15.275088 265579 start.go:228] waiting for startup goroutines ...
I1009 18:33:15.427838 265579 out.go:177] * Enabled addons: storage-provisioner, default-storageclass
I1009 18:33:15.429174 265579 addons.go:502] enable addons completed in 909.675466ms: enabled=[storage-provisioner default-storageclass]
I1009 18:33:15.429202 265579 start.go:233] waiting for cluster config update ...
I1009 18:33:15.429212 265579 start.go:242] writing updated cluster config ...
I1009 18:33:15.429444 265579 ssh_runner.go:195] Run: rm -f paused
I1009 18:33:15.479963 265579 start.go:600] kubectl: 1.34.1, cluster: 1.28.3 (minor skew: 6)
I1009 18:33:15.481658 265579 out.go:177]
W1009 18:33:15.483575 265579 out.go:239] ! /usr/local/bin/kubectl is version 1.34.1, which may have incompatibilities with Kubernetes 1.28.3.
I1009 18:33:15.484917 265579 out.go:177] - Want kubectl v1.28.3? Try 'minikube kubectl -- get pods -A'
I1009 18:33:15.486762 265579 out.go:177] * Done! kubectl is now configured to use "stopped-upgrade-731100" cluster and "default" namespace by default
I1009 18:33:14.793997 281272 ssh_runner.go:195] Run: cat /etc/os-release
I1009 18:33:14.797769 281272 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I1009 18:33:14.797803 281272 info.go:137] Remote host: Debian GNU/Linux 12 (bookworm)
I1009 18:33:14.797827 281272 filesync.go:126] Scanning /home/jenkins/minikube-integration/21139-11202/.minikube/addons for local assets ...
I1009 18:33:14.797883 281272 filesync.go:126] Scanning /home/jenkins/minikube-integration/21139-11202/.minikube/files for local assets ...
I1009 18:33:14.797983 281272 filesync.go:149] local asset: /home/jenkins/minikube-integration/21139-11202/.minikube/files/etc/ssl/certs/147382.pem -> 147382.pem in /etc/ssl/certs
I1009 18:33:14.797996 281272 vm_assets.go:164] NewFileAsset: /home/jenkins/minikube-integration/21139-11202/.minikube/files/etc/ssl/certs/147382.pem -> /etc/ssl/certs/147382.pem
I1009 18:33:14.798080 281272 ssh_runner.go:195] Run: sudo mkdir -p /etc/ssl/certs
I1009 18:33:14.806796 281272 ssh_runner.go:362] scp /home/jenkins/minikube-integration/21139-11202/.minikube/files/etc/ssl/certs/147382.pem --> /etc/ssl/certs/147382.pem (1708 bytes)
I1009 18:33:14.829305 281272 start.go:296] duration metric: took 175.07094ms for postStartSetup
I1009 18:33:14.829741 281272 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" NoKubernetes-711902
I1009 18:33:14.849149 281272 profile.go:143] Saving config to /home/jenkins/minikube-integration/21139-11202/.minikube/profiles/NoKubernetes-711902/config.json ...
I1009 18:33:14.849424 281272 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I1009 18:33:14.849473 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:14.868321 281272 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33003 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa Username:docker}
I1009 18:33:14.970174 281272 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I1009 18:33:14.974960 281272 start.go:128] duration metric: took 5.049916115s to createHost
I1009 18:33:14.974987 281272 start.go:83] releasing machines lock for "NoKubernetes-711902", held for 5.050146708s
I1009 18:33:14.975061 281272 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" NoKubernetes-711902
I1009 18:33:15.002433 281272 ssh_runner.go:195] Run: cat /version.json
I1009 18:33:15.002450 281272 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I1009 18:33:15.002489 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:15.002501 281272 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" NoKubernetes-711902
I1009 18:33:15.021468 281272 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33003 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa Username:docker}
I1009 18:33:15.022782 281272 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33003 SSHKeyPath:/home/jenkins/minikube-integration/21139-11202/.minikube/machines/NoKubernetes-711902/id_rsa Username:docker}
I1009 18:33:15.126408 281272 ssh_runner.go:195] Run: systemctl --version
I1009 18:33:15.204672 281272 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
W1009 18:33:15.210075 281272 cni.go:209] loopback cni configuration skipped: "/etc/cni/net.d/*loopback.conf*" not found
I1009 18:33:15.210143 281272 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I1009 18:33:15.240346 281272 cni.go:262] disabled [/etc/cni/net.d/10-crio-bridge.conflist.disabled, /etc/cni/net.d/87-podman-bridge.conflist] bridge cni config(s)
I1009 18:33:15.240375 281272 start.go:495] detecting cgroup driver to use...
I1009 18:33:15.240410 281272 detect.go:190] detected "systemd" cgroup driver on host os
I1009 18:33:15.240539 281272 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I1009 18:33:15.259251 281272 binary.go:59] Skipping Kubernetes binary download due to --no-kubernetes flag
I1009 18:33:15.259380 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.9"|' /etc/containerd/config.toml"
I1009 18:33:15.271964 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I1009 18:33:15.282979 281272 containerd.go:146] configuring containerd to use "systemd" as cgroup driver...
I1009 18:33:15.283042 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = true|g' /etc/containerd/config.toml"
I1009 18:33:15.293870 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I1009 18:33:15.304902 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I1009 18:33:15.318043 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I1009 18:33:15.330292 281272 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I1009 18:33:15.340561 281272 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I1009 18:33:15.352781 281272 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I1009 18:33:15.363348 281272 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I1009 18:33:15.373310 281272 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1009 18:33:15.469841 281272 ssh_runner.go:195] Run: sudo systemctl restart containerd
I1009 18:33:15.558603 281272 start.go:495] detecting cgroup driver to use...
I1009 18:33:15.558650 281272 detect.go:190] detected "systemd" cgroup driver on host os
I1009 18:33:15.558699 281272 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I1009 18:33:15.575205 281272 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I1009 18:33:15.589442 281272 ssh_runner.go:195] Run: sudo systemctl stop -f containerd
I1009 18:33:15.609869 281272 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I1009 18:33:15.624274 281272 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I1009 18:33:15.640218 281272 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/dockershim.sock
" | sudo tee /etc/crictl.yaml"
I1009 18:33:15.658180 281272 binary.go:59] Skipping Kubernetes binary download due to --no-kubernetes flag
I1009 18:33:15.658267 281272 ssh_runner.go:195] Run: which cri-dockerd
I1009 18:33:15.662366 281272 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I1009 18:33:15.673211 281272 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (189 bytes)
I1009 18:33:15.687836 281272 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I1009 18:33:15.811338 281272 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I1009 18:33:15.895027 281272 docker.go:575] configuring docker to use "systemd" as cgroup driver...
I1009 18:33:15.895147 281272 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (129 bytes)
I1009 18:33:15.908669 281272 ssh_runner.go:195] Run: sudo systemctl reset-failed docker
I1009 18:33:15.921005 281272 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1009 18:33:16.001315 281272 ssh_runner.go:195] Run: sudo systemctl restart docker
I1009 18:33:16.810802 281272 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
I1009 18:33:16.825293 281272 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I1009 18:33:16.852612 281272 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I1009 18:33:16.884684 281272 out.go:252] * Preparing Docker 28.5.0 ...
I1009 18:33:16.885018 281272 ssh_runner.go:195] Run: rm -f paused
I1009 18:33:16.892737 281272 out.go:179] * Done! minikube is ready without Kubernetes!
I1009 18:33:16.896066 281272 out.go:203] ╭───────────────────────────────────────────────────────────────────────────────────────╮
│ │
│ * Things to try without Kubernetes ... │
│ │
│ - "minikube ssh" to SSH into minikube's node. │
│ - "minikube docker-env" to point your docker-cli to the docker inside minikube. │
│ - "minikube image" to build images without docker. │
│ │
╰───────────────────────────────────────────────────────────────────────────────────────╯
==> Docker <==
Oct 09 18:33:15 NoKubernetes-711902 dockerd[597]: time="2025-10-09T18:33:15.602382711Z" level=info msg="Waiting for containerd to be ready to restart event processing" module=libcontainerd namespace=plugins.moby
Oct 09 18:33:16 NoKubernetes-711902 systemd[1]: Stopping docker.service - Docker Application Container Engine...
Oct 09 18:33:16 NoKubernetes-711902 dockerd[597]: time="2025-10-09T18:33:16.013280603Z" level=info msg="Processing signal 'terminated'"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[597]: time="2025-10-09T18:33:16.014470939Z" level=warning msg="Error while testing if containerd API is ready" error="Canceled: grpc: the client connection is closing"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[597]: time="2025-10-09T18:33:16.014839474Z" level=info msg="Daemon shutdown complete"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[597]: time="2025-10-09T18:33:16.014900748Z" level=warning msg="Error while testing if containerd API is ready" error="Canceled: context canceled while waiting for connections to become ready"
Oct 09 18:33:16 NoKubernetes-711902 systemd[1]: docker.service: Deactivated successfully.
Oct 09 18:33:16 NoKubernetes-711902 systemd[1]: Stopped docker.service - Docker Application Container Engine.
Oct 09 18:33:16 NoKubernetes-711902 systemd[1]: Starting docker.service - Docker Application Container Engine...
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.124911078Z" level=info msg="Starting up"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.125794051Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.125894724Z" level=info msg="CDI directory does not exist, skipping: failed to monitor for changes: no such file or directory" dir=/var/run/cdi
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.125913849Z" level=info msg="CDI directory does not exist, skipping: failed to monitor for changes: no such file or directory" dir=/etc/cdi
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.135738123Z" level=info msg="Creating a containerd client" address=/run/containerd/containerd.sock timeout=1m0s
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.138799900Z" level=info msg="[graphdriver] trying configured driver: overlay2"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.146138652Z" level=info msg="Loading containers: start."
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.767942334Z" level=info msg="Loading containers: done."
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.781483515Z" level=info msg="Docker daemon" commit=cd04830 containerd-snapshotter=false storage-driver=overlay2 version=28.5.0
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.781550857Z" level=info msg="Initializing buildkit"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.801568203Z" level=info msg="Completed buildkit initialization"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.808683377Z" level=info msg="Daemon has completed initialization"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.808817533Z" level=info msg="API listen on /run/docker.sock"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.808897097Z" level=info msg="API listen on [::]:2376"
Oct 09 18:33:16 NoKubernetes-711902 dockerd[1049]: time="2025-10-09T18:33:16.808824253Z" level=info msg="API listen on /var/run/docker.sock"
Oct 09 18:33:16 NoKubernetes-711902 systemd[1]: Started docker.service - Docker Application Container Engine.
==> container status <==
time="2025-10-09T18:33:18Z" level=fatal msg="validate service connection: validate CRI v1 runtime API for endpoint \"unix:///var/run/dockershim.sock\": rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial unix /var/run/dockershim.sock: connect: no such file or directory\""
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
==> describe nodes <==
command /bin/bash -c "sudo /var/lib/minikube/binaries/v0.0.0/kubectl describe nodes --kubeconfig=/var/lib/minikube/kubeconfig" failed with error: /bin/bash -c "sudo /var/lib/minikube/binaries/v0.0.0/kubectl describe nodes --kubeconfig=/var/lib/minikube/kubeconfig": Process exited with status 1
stdout:
stderr:
sudo: /var/lib/minikube/binaries/v0.0.0/kubectl: command not found
==> dmesg <==
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 0e b0 08 c2 8f ad 08 06
[Oct 9 18:05] IPv4: martian source 10.244.0.1 from 10.244.0.15, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 66 86 b6 64 71 c5 08 06
[Oct 9 18:17] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 12 a7 d3 2c b6 cd 08 06
[Oct 9 18:20] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff be a6 34 11 a0 3e 08 06
[Oct 9 18:27] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 86 7b 16 d4 1f 2a 08 06
[ +0.175207] IPv4: martian source 10.244.0.1 from 10.244.0.3, on dev eth0
[ +0.000009] ll header: 00000000: ff ff ff ff ff ff 72 f4 98 6f 82 5e 08 06
[Oct 9 18:28] IPv4: martian source 10.244.0.1 from 10.244.0.4, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 6a 05 82 5f ef 5c 08 06
[Oct 9 18:29] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 2e de 88 79 fd 2b 08 06
[Oct 9 18:31] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000009] ll header: 00000000: ff ff ff ff ff ff 12 56 f0 8a a2 50 08 06
[Oct 9 18:32] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 66 3d 78 60 14 43 08 06
[ +0.047279] IPv4: martian source 10.244.0.1 from 10.244.0.3, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 82 eb 5a ac 74 61 08 06
[ +0.016595] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff b6 84 b7 e8 55 53 08 06
[ +0.001635] IPv4: martian source 10.244.0.1 from 10.244.0.3, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 7a f1 ee 0c 72 39 08 06
==> kernel <==
18:33:18 up 1:15, 0 user, load average: 7.49, 3.01, 2.01
Linux NoKubernetes-711902 6.8.0-1041-gcp #43~22.04.1-Ubuntu SMP Wed Sep 24 23:11:19 UTC 2025 x86_64 GNU/Linux
PRETTY_NAME="Debian GNU/Linux 12 (bookworm)"
==> kubelet <==
-- No entries --
-- /stdout --
helpers_test.go:262: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p NoKubernetes-711902 -n NoKubernetes-711902
helpers_test.go:262: (dbg) Non-zero exit: out/minikube-linux-amd64 status --format={{.APIServer}} -p NoKubernetes-711902 -n NoKubernetes-711902: exit status 6 (298.907407ms)
-- stdout --
Stopped
WARNING: Your kubectl is pointing to stale minikube-vm.
To fix the kubectl context, run `minikube update-context`
-- /stdout --
** stderr **
E1009 18:33:19.260803 285043 status.go:458] kubeconfig endpoint: get endpoint: "NoKubernetes-711902" does not appear in /home/jenkins/minikube-integration/21139-11202/kubeconfig
** /stderr **
helpers_test.go:262: status error: exit status 6 (may be ok)
helpers_test.go:264: "NoKubernetes-711902" apiserver is not running, skipping kubectl commands (state="Stopped")
--- FAIL: TestNoKubernetes/serial/VerifyNok8sNoK8sDownloads (2.37s)