
-------------
1) K3s logs
-------------

Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.361060    5772 serving.go:355] Generated self-signed cert in-memory
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.363907    5772 event.go:294] "Event occurred" object="kube-system/kuberouter" fieldPath="" kind="Addon" apiVersion="k3s.cattle.io/v1" type="Normal" reason="ApplyingManifest" message="Applying manifest at \"/mnt/Metzwerkspeicher/ix-applications/k3s/server/manifests/kuberouter.yaml\""
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.372003    5772 event.go:294] "Event occurred" object="kube-system/kuberouter" fieldPath="" kind="Addon" apiVersion="k3s.cattle.io/v1" type="Normal" reason="AppliedManifest" message="Applied manifest at \"/mnt/Metzwerkspeicher/ix-applications/k3s/server/manifests/kuberouter.yaml\""
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.376010    5772 event.go:294] "Event occurred" object="kube-system/multus-daemonset" fieldPath="" kind="Addon" apiVersion="k3s.cattle.io/v1" type="Normal" reason="ApplyingManifest" message="Applying manifest at \"/mnt/Metzwerkspeicher/ix-applications/k3s/server/manifests/multus-daemonset.yaml\""
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.393441    5772 event.go:294] "Event occurred" object="kube-system/multus-daemonset" fieldPath="" kind="Addon" apiVersion="k3s.cattle.io/v1" type="Normal" reason="AppliedManifest" message="Applied manifest at \"/mnt/Metzwerkspeicher/ix-applications/k3s/server/manifests/multus-daemonset.yaml\""
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.402708    5772 event.go:294] "Event occurred" object="kube-system/nfs-csi" fieldPath="" kind="Addon" apiVersion="k3s.cattle.io/v1" type="Normal" reason="ApplyingManifest" message="Applying manifest at \"/mnt/Metzwerkspeicher/ix-applications/k3s/server/manifests/nfs-csi.yaml\""
Dec 24 07:36:40 Server k3s[5772]: time="2023-12-24T07:36:40+01:00" level=info msg="Starting discovery.k8s.io/v1, Kind=EndpointSlice controller"
Dec 24 07:36:40 Server k3s[5772]: W1224 07:36:40.405544    5772 controllermanager.go:288] "cloud-node" is disabled
Dec 24 07:36:40 Server k3s[5772]: W1224 07:36:40.405557    5772 controllermanager.go:288] "cloud-node-lifecycle" is disabled
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.406058    5772 controllermanager.go:311] Started "service"
Dec 24 07:36:40 Server k3s[5772]: W1224 07:36:40.406071    5772 controllermanager.go:288] "route" is disabled
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.406769    5772 controller.go:227] Starting service controller
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.406790    5772 shared_informer.go:270] Waiting for caches to sync for service
Dec 24 07:36:40 Server k3s[5772]: time="2023-12-24T07:36:40+01:00" level=info msg="Starting helm.cattle.io/v1, Kind=HelmChart controller"
Dec 24 07:36:40 Server k3s[5772]: time="2023-12-24T07:36:40+01:00" level=info msg="Starting helm.cattle.io/v1, Kind=HelmChartConfig controller"
Dec 24 07:36:40 Server k3s[5772]: time="2023-12-24T07:36:40+01:00" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=ClusterRoleBinding controller"
Dec 24 07:36:40 Server k3s[5772]: time="2023-12-24T07:36:40+01:00" level=info msg="Starting batch/v1, Kind=Job controller"
Dec 24 07:36:40 Server k3s[5772]: time="2023-12-24T07:36:40+01:00" level=info msg="Starting /v1, Kind=Secret controller"
Dec 24 07:36:40 Server k3s[5772]: time="2023-12-24T07:36:40+01:00" level=info msg="Starting /v1, Kind=ConfigMap controller"
Dec 24 07:36:40 Server k3s[5772]: time="2023-12-24T07:36:40+01:00" level=info msg="Starting /v1, Kind=ServiceAccount controller"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.506928    5772 shared_informer.go:277] Caches are synced for service
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.594288    5772 event.go:294] "Event occurred" object="kube-system/nfs-csi" fieldPath="" kind="Addon" apiVersion="k3s.cattle.io/v1" type="Normal" reason="AppliedManifest" message="Applied manifest at \"/mnt/Metzwerkspeicher/ix-applications/k3s/server/manifests/nfs-csi.yaml\""
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.598383    5772 event.go:294] "Event occurred" object="kube-system/rolebindings" fieldPath="" kind="Addon" apiVersion="k3s.cattle.io/v1" type="Normal" reason="ApplyingManifest" message="Applying manifest at \"/mnt/Metzwerkspeicher/ix-applications/k3s/server/manifests/rolebindings.yaml\""
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.606575    5772 event.go:294] "Event occurred" object="kube-system/rolebindings" fieldPath="" kind="Addon" apiVersion="k3s.cattle.io/v1" type="Normal" reason="AppliedManifest" message="Applied manifest at \"/mnt/Metzwerkspeicher/ix-applications/k3s/server/manifests/rolebindings.yaml\""
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.610338    5772 event.go:294] "Event occurred" object="kube-system/smb-csi" fieldPath="" kind="Addon" apiVersion="k3s.cattle.io/v1" type="Normal" reason="ApplyingManifest" message="Applying manifest at \"/mnt/Metzwerkspeicher/ix-applications/k3s/server/manifests/smb-csi.yaml\""
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.634606    5772 event.go:294] "Event occurred" object="kube-system/smb-csi" fieldPath="" kind="Addon" apiVersion="k3s.cattle.io/v1" type="Normal" reason="AppliedManifest" message="Applied manifest at \"/mnt/Metzwerkspeicher/ix-applications/k3s/server/manifests/smb-csi.yaml\""
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.648464    5772 event.go:294] "Event occurred" object="kube-system/zfs-operator" fieldPath="" kind="Addon" apiVersion="k3s.cattle.io/v1" type="Normal" reason="ApplyingManifest" message="Applying manifest at \"/mnt/Metzwerkspeicher/ix-applications/k3s/server/manifests/zfs-operator.yaml\""
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.736701    5772 controllermanager.go:182] Version: v1.26.6+k3s-e18037a7-dirty
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.736735    5772 controllermanager.go:184] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.739050    5772 configmap_cafile_content.go:202] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.739076    5772 shared_informer.go:270] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.739138    5772 configmap_cafile_content.go:202] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.739146    5772 shared_informer.go:270] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.739259    5772 requestheader_controller.go:169] Starting RequestHeaderAuthRequestController
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.739271    5772 shared_informer.go:270] Waiting for caches to sync for RequestHeaderAuthRequestController
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.739342    5772 secure_serving.go:210] Serving securely on 127.0.0.1:10257
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.740033    5772 tlsconfig.go:240] "Starting DynamicServingCertificateController"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.794729    5772 shared_informer.go:270] Waiting for caches to sync for tokens
Dec 24 07:36:40 Server k3s[5772]: time="2023-12-24T07:36:40+01:00" level=info msg="Updating TLS secret for kube-system/k3s-serving (count: 11): map[listener.cattle.io/cn-0.0.0.0:0.0.0.0 listener.cattle.io/cn-127.0.0.1:127.0.0.1 listener.cattle.io/cn-172.17.0.1:172.17.0.1 listener.cattle.io/cn-192.168.3.5:192.168.3.5 listener.cattle.io/cn-__1-f16284:::1 listener.cattle.io/cn-ix-truenas:ix-truenas listener.cattle.io/cn-kubernetes:kubernetes listener.cattle.io/cn-kubernetes.default:kubernetes.default listener.cattle.io/cn-kubernetes.default.svc:kubernetes.default.svc listener.cattle.io/cn-kubernetes.default.svc.cluster.local:kubernetes.default.svc.cluster.local listener.cattle.io/cn-localhost:localhost listener.cattle.io/fingerprint:SHA1=79AAA80589B7B37A923BBA58B2EA722CDA7626AA]"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.797563    5772 controllermanager.go:622] Started "statefulset"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.798707    5772 stateful_set.go:152] Starting stateful set controller
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.798724    5772 shared_informer.go:270] Waiting for caches to sync for stateful set
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.810799    5772 node_lifecycle_controller.go:492] Controller will reconcile labels.
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.810865    5772 controllermanager.go:622] Started "nodelifecycle"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.810980    5772 node_lifecycle_controller.go:527] Sending events to api server.
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.811002    5772 node_lifecycle_controller.go:538] Starting node controller
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.811009    5772 shared_informer.go:270] Waiting for caches to sync for taint
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.813074    5772 controllermanager.go:622] Started "ttl-after-finished"
Dec 24 07:36:40 Server k3s[5772]: E1224 07:36:40.814814    5772 core.go:207] failed to start cloud node lifecycle controller: no cloud provider provided
Dec 24 07:36:40 Server k3s[5772]: W1224 07:36:40.814836    5772 controllermanager.go:600] Skipping "cloud-node-lifecycle"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.822586    5772 ttlafterfinished_controller.go:104] Starting TTL after finished controller
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.822611    5772 shared_informer.go:270] Waiting for caches to sync for TTL after finished
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.823052    5772 controllermanager.go:622] Started "endpointslice"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.824903    5772 controllermanager.go:622] Started "endpointslicemirroring"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.826799    5772 controllermanager.go:622] Started "podgc"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.826912    5772 gc_controller.go:102] Starting GC controller
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.826928    5772 shared_informer.go:270] Waiting for caches to sync for GC
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.828856    5772 controllermanager.go:622] Started "serviceaccount"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.829408    5772 serviceaccounts_controller.go:111] Starting service account controller
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.829418    5772 shared_informer.go:270] Waiting for caches to sync for service account
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.833914    5772 endpointslicemirroring_controller.go:211] Starting EndpointSliceMirroring controller
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.833921    5772 endpointslice_controller.go:257] Starting endpoint slice controller
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.833944    5772 shared_informer.go:270] Waiting for caches to sync for endpoint_slice_mirroring
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.833954    5772 shared_informer.go:270] Waiting for caches to sync for endpoint_slice
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.834252    5772 controllermanager.go:622] Started "garbagecollector"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.834299    5772 garbagecollector.go:154] Starting garbage collector controller
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.834313    5772 shared_informer.go:270] Waiting for caches to sync for garbage collector
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.834372    5772 graph_builder.go:291] GraphBuilder running
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.836813    5772 controllermanager.go:622] Started "replicaset"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.836930    5772 replica_set.go:201] Starting replicaset controller
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.836942    5772 shared_informer.go:270] Waiting for caches to sync for ReplicaSet
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.839416    5772 shared_informer.go:277] Caches are synced for RequestHeaderAuthRequestController
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.839423    5772 shared_informer.go:277] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.839479    5772 shared_informer.go:277] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.894912    5772 shared_informer.go:277] Caches are synced for tokens
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.913296    5772 server.go:152] "Starting Kubernetes Scheduler" version="v1.26.6+k3s-e18037a7-dirty"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.913325    5772 server.go:154] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.915747    5772 requestheader_controller.go:169] Starting RequestHeaderAuthRequestController
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.915760    5772 shared_informer.go:270] Waiting for caches to sync for RequestHeaderAuthRequestController
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.915777    5772 configmap_cafile_content.go:202] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.915783    5772 shared_informer.go:270] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.915805    5772 configmap_cafile_content.go:202] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file"
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.915815    5772 shared_informer.go:270] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.915864    5772 secure_serving.go:210] Serving securely on 127.0.0.1:10259
Dec 24 07:36:40 Server k3s[5772]: I1224 07:36:40.916050    5772 tlsconfig.go:240] "Starting DynamicServingCertificateController"
Dec 24 07:36:41 Server k3s[5772]: I1224 07:36:41.016848    5772 shared_informer.go:277] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file
Dec 24 07:36:41 Server k3s[5772]: I1224 07:36:41.016955    5772 shared_informer.go:277] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
Dec 24 07:36:41 Server k3s[5772]: I1224 07:36:41.016977    5772 shared_informer.go:277] Caches are synced for RequestHeaderAuthRequestController
Dec 24 07:36:41 Server k3s[5772]: time="2023-12-24T07:36:41+01:00" level=info msg="Waiting for containerd startup: rpc error: code = Unknown desc = server is not initialized yet"
Dec 24 07:36:42 Server k3s[5772]: time="2023-12-24T07:36:42+01:00" level=info msg="containerd is now running"
Dec 24 07:36:42 Server k3s[5772]: time="2023-12-24T07:36:42+01:00" level=info msg="Connecting to proxy" url="wss://127.0.0.1:6443/v1-k3s/connect"
Dec 24 07:36:42 Server k3s[5772]: time="2023-12-24T07:36:42+01:00" level=info msg="Running kubelet --address=0.0.0.0 --allowed-unsafe-sysctls=net.ipv4.ip_forward,net.ipv6.conf.all.forwarding --anonymous-auth=false --authentication-token-webhook=true --authorization-mode=Webhook --cgroup-driver=systemd --client-ca-file=/mnt/Metzwerkspeicher/ix-applications/k3s/agent/client-ca.crt --cluster-dns=172.17.0.10 --cluster-domain=cluster.local --config=/etc/rancher/k3s/kubelet_config.yaml --container-runtime-endpoint=unix:///run/k3s/containerd/containerd.sock --containerd=/run/k3s/containerd/containerd.sock --eviction-hard=imagefs.available<5%,nodefs.available<5% --eviction-minimum-reclaim=imagefs.available=10%,nodefs.available=10% --fail-swap-on=false --healthz-bind-address=127.0.0.1 --hostname-override=ix-truenas --kubeconfig=/mnt/Metzwerkspeicher/ix-applications/k3s/agent/kubelet.kubeconfig --node-ip=0.0.0.0 --node-labels= --pod-infra-container-image=rancher/mirrored-pause:3.6 --pod-manifest-path=/mnt/Metzwerkspeicher/ix-applications/k3s/agent/pod-manifests --protect-kernel-defaults=true --read-only-port=0 --resolv-conf=/etc/resolv.conf --serialize-image-pulls=false --tls-cert-file=/mnt/Metzwerkspeicher/ix-applications/k3s/agent/serving-kubelet.crt --tls-private-key-file=/mnt/Metzwerkspeicher/ix-applications/k3s/agent/serving-kubelet.key"
Dec 24 07:36:42 Server k3s[5772]: time="2023-12-24T07:36:42+01:00" level=info msg="Handling backend connection request [ix-truenas]"
Dec 24 07:36:42 Server k3s[5772]: Flag --containerd has been deprecated, This is a cadvisor flag that was mistakenly registered with the Kubelet. Due to legacy concerns, it will follow the standard CLI deprecation timeline before being removed.
Dec 24 07:36:42 Server k3s[5772]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI.
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.152210    5772 server.go:197] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime"
Dec 24 07:36:42 Server k3s[5772]: time="2023-12-24T07:36:42+01:00" level=info msg="Annotations and labels have already set on node: ix-truenas"
Dec 24 07:36:42 Server k3s[5772]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI.
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.209611    5772 server.go:407] "Kubelet version" kubeletVersion="v1.26.6+k3s-e18037a7-dirty"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.209628    5772 server.go:409] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.213619    5772 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/mnt/Metzwerkspeicher/ix-applications/k3s/agent/client-ca.crt"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.248985    5772 server.go:654] "--cgroups-per-qos enabled, but --cgroup-root was not specified.  defaulting to /"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.249823    5772 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[]
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.250051    5772 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={RuntimeCgroupsName: SystemCgroupsName: KubeletCgroupsName: KubeletOOMScoreAdj:-999 ContainerRuntime: CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:systemd KubeletRootDir:/var/lib/kubelet ProtectKernelDefaults:true NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: ReservedSystemCPUs: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[] SystemReserved:map[] HardEvictionThresholds:[{Signal:nodefs.available Operator:LessThan Value:{Quantity:<nil> Percentage:0.05} GracePeriod:0s MinReclaim:<nil>} {Signal:imagefs.available Operator:LessThan Value:{Quantity:<nil> Percentage:0.05} GracePeriod:0s MinReclaim:<nil>}]} QOSReserved:map[] CPUManagerPolicy:none CPUManagerPolicyOptions:map[] ExperimentalTopologyManagerScope:container CPUManagerReconcilePeriod:10s ExperimentalMemoryManagerPolicy:None ExperimentalMemoryManagerReservedMemory:[] ExperimentalPodPidsLimit:-1 EnforceCPULimits:true CPUCFSQuotaPeriod:100ms ExperimentalTopologyManagerPolicy:none ExperimentalTopologyManagerPolicyOptions:map[]}
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.250996    5772 topology_manager.go:134] "Creating topology manager with policy per scope" topologyPolicyName="none" topologyScopeName="container"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.251234    5772 container_manager_linux.go:308] "Creating device plugin manager"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.264579    5772 state_mem.go:36] "Initialized new in-memory state store"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.268225    5772 kubelet.go:398] "Attempting to sync node with API server"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.268286    5772 kubelet.go:286] "Adding static pod path" path="/mnt/Metzwerkspeicher/ix-applications/k3s/agent/pod-manifests"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.268621    5772 kubelet.go:297] "Adding apiserver pod source"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.268678    5772 apiserver.go:42] "Waiting for node sync before watching apiserver pods"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.272373    5772 kuberuntime_manager.go:244] "Container runtime initialized" containerRuntime="containerd" version="truenas/v1.7.1-k3s1" apiVersion="v1"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.276178    5772 nodeshutdown_manager_linux.go:137] "Creating node shutdown manager" shutdownGracePeriodRequested="15s" shutdownGracePeriodCriticalPods="10s" shutdownGracePeriodByPodPriority=[{Priority:0 ShutdownGracePeriodSeconds:5} {Priority:2000000000 ShutdownGracePeriodSeconds:10}]
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.276729    5772 server.go:1181] "Started kubelet"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.277036    5772 server.go:161] "Starting to listen" address="0.0.0.0" port=10250
Dec 24 07:36:42 Server k3s[5772]: E1224 07:36:42.277681    5772 cri_stats_provider.go:455] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/mnt/Metzwerkspeicher/ix-applications/k3s/agent/containerd/io.containerd.snapshotter.v1.overlayfs"
Dec 24 07:36:42 Server k3s[5772]: E1224 07:36:42.277712    5772 kubelet.go:1386] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.278155    5772 server.go:451] "Adding debug handlers to kubelet server"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.285408    5772 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.285949    5772 volume_manager.go:293] "Starting Kubelet Volume Manager"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.286030    5772 desired_state_of_world_populator.go:151] "Desired state populator starts to run"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.287483    5772 scope.go:115] "RemoveContainer" containerID="e5905aa085384807599db9007cdd3e8c951db9350ea082b7fe73800f1aa8af73"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.312041    5772 cpu_manager.go:214] "Starting CPU manager" policy="none"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.312056    5772 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.312080    5772 state_mem.go:36] "Initialized new in-memory state store"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.312752    5772 state_mem.go:88] "Updated default CPUSet" cpuSet=""
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.312782    5772 state_mem.go:96] "Updated CPUSet assignments" assignments=map[]
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.312830    5772 policy_none.go:49] "None policy: Start"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.314000    5772 memory_manager.go:169] "Starting memorymanager" policy="None"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.314054    5772 state_mem.go:35] "Initializing new in-memory state store"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.314387    5772 state_mem.go:75] "Updated machine memory state"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.378722    5772 scope.go:115] "RemoveContainer" containerID="7ba0157b0f55f70d8fee715901d77640fb365737bf1cbd78c2bb1cad768ccaec"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.391697    5772 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv4
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.407695    5772 plugin_manager.go:118] "Starting Kubelet Plugin Manager"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.409591    5772 kuberuntime_manager.go:1114] "Updating runtime config through cri with podcidr" CIDR="172.16.0.0/16"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.409985    5772 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="172.16.0.0/16"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.418574    5772 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv6
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.418764    5772 status_manager.go:176] "Starting to sync pod status with apiserver"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.418841    5772 kubelet.go:2113] "Starting kubelet main sync loop"
Dec 24 07:36:42 Server k3s[5772]: E1224 07:36:42.419091    5772 kubelet.go:2137] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.434636    5772 scope.go:115] "RemoveContainer" containerID="40bfa901f0a5a6ac35b6b8bb4b45d8da0d828264bb91cc12a22cb6b8588ad658"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.513302    5772 scope.go:115] "RemoveContainer" containerID="d6403e0a773151c9a90c0623048a29af564a0d1d5e5400fcd1ee37412238f3ff"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519379    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f6de42afefc2ec70574d8b0f96996b7f793aefbcbdc0ae368ef7c2007c6a48b"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519396    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="816c17b59a30c268333f742692e95a5b572f81e3e60673767c1f6a224c4885a3"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519414    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e7f3b1a7d2ec384669397fbe1f7ad77f59a9a63310b9fc33c507fc0cbef7503"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519421    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="789dabe34ac867bd5d2d9a70ea0353a55b47a217020c6bcdce365b24d8ebdc3c"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519430    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44eed0940b04904e6b845cb433236c62f564e0aa4de5db2db5b19bb5da2b7bb0"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519441    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbc89515365b8c4da61600445d886b56fb080ee5f1b5460d27cc057e5bffc696"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519449    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98258011043e1fb762b0713705ebad2a57869896607c048ade96438809c929eb"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519457    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30d5456a4a289972825c47e94d73006605204272418e3dd7cd01e70e04dee4d2"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519464    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc9617256dadba1ff8558b0f42ae75d0c6e1bde5f7a46929f46081b937a027d5"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519474    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac610a2dcbc362396aa871541fda3cb6f8b5b9257f9b12eabff833500d6db758"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519483    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8368e3650403fedcbbdfba96260a46d1a7177973f2991a628e040c31dd05404a"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519496    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ba0157b0f55f70d8fee715901d77640fb365737bf1cbd78c2bb1cad768ccaec"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519504    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c513fd7fcec6ed9f9c4909205e29230d2724d75c3562c5fcf944727e1138559"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519510    5772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae8434ee15df9dfeb47011fa3e1eb18361861f2ed71ca60703dc52c6e72a2316"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.519523    5772 scope.go:115] "RemoveContainer" containerID="c4e05d7f8b87d2b031bce89f5689fc35f87bce9846332572650156d3799a8a08"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.533679    5772 kubelet_node_status.go:70] "Attempting to register node" node="ix-truenas"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.539704    5772 kubelet_node_status.go:108] "Node was previously registered" node="ix-truenas"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.539789    5772 kubelet_node_status.go:73] "Successfully registered node" node="ix-truenas"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.581863    5772 scope.go:115] "RemoveContainer" containerID="c4e05d7f8b87d2b031bce89f5689fc35f87bce9846332572650156d3799a8a08"
Dec 24 07:36:42 Server k3s[5772]: E1224 07:36:42.582390    5772 remote_runtime.go:415] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c4e05d7f8b87d2b031bce89f5689fc35f87bce9846332572650156d3799a8a08\": not found" containerID="c4e05d7f8b87d2b031bce89f5689fc35f87bce9846332572650156d3799a8a08"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.582417    5772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={Type:containerd ID:c4e05d7f8b87d2b031bce89f5689fc35f87bce9846332572650156d3799a8a08} err="failed to get container status \"c4e05d7f8b87d2b031bce89f5689fc35f87bce9846332572650156d3799a8a08\": rpc error: code = NotFound desc = an error occurred when try to find container \"c4e05d7f8b87d2b031bce89f5689fc35f87bce9846332572650156d3799a8a08\": not found"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.582426    5772 scope.go:115] "RemoveContainer" containerID="f4ff6e96db3e5dd784e5ec163d8d2c9eb1d961eda5d8a9ca50d8a2d169ca8584"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.605239    5772 scope.go:115] "RemoveContainer" containerID="f4ff6e96db3e5dd784e5ec163d8d2c9eb1d961eda5d8a9ca50d8a2d169ca8584"
Dec 24 07:36:42 Server k3s[5772]: E1224 07:36:42.606729    5772 remote_runtime.go:368] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to set removing state for container \"f4ff6e96db3e5dd784e5ec163d8d2c9eb1d961eda5d8a9ca50d8a2d169ca8584\": container is already in removing state" containerID="f4ff6e96db3e5dd784e5ec163d8d2c9eb1d961eda5d8a9ca50d8a2d169ca8584"
Dec 24 07:36:42 Server k3s[5772]: E1224 07:36:42.606781    5772 kuberuntime_gc.go:147] "Failed to remove container" err="rpc error: code = Unknown desc = failed to set removing state for container \"f4ff6e96db3e5dd784e5ec163d8d2c9eb1d961eda5d8a9ca50d8a2d169ca8584\": container is already in removing state" containerID="f4ff6e96db3e5dd784e5ec163d8d2c9eb1d961eda5d8a9ca50d8a2d169ca8584"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.606794    5772 scope.go:115] "RemoveContainer" containerID="c4e05d7f8b87d2b031bce89f5689fc35f87bce9846332572650156d3799a8a08"
Dec 24 07:36:42 Server k3s[5772]: E1224 07:36:42.607137    5772 kuberuntime_gc.go:147] "Failed to remove container" err="failed to get container status \"c4e05d7f8b87d2b031bce89f5689fc35f87bce9846332572650156d3799a8a08\": rpc error: code = NotFound desc = an error occurred when try to find container \"c4e05d7f8b87d2b031bce89f5689fc35f87bce9846332572650156d3799a8a08\": not found" containerID="c4e05d7f8b87d2b031bce89f5689fc35f87bce9846332572650156d3799a8a08"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.622830    5772 event.go:294] "Event occurred" object="kube-system/zfs-operator" fieldPath="" kind="Addon" apiVersion="k3s.cattle.io/v1" type="Normal" reason="AppliedManifest" message="Applied manifest at \"/mnt/Metzwerkspeicher/ix-applications/k3s/server/manifests/zfs-operator.yaml\""
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.650927    5772 scope.go:115] "RemoveContainer" containerID="f4ff6e96db3e5dd784e5ec163d8d2c9eb1d961eda5d8a9ca50d8a2d169ca8584"
Dec 24 07:36:42 Server k3s[5772]: E1224 07:36:42.651209    5772 remote_runtime.go:415] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"f4ff6e96db3e5dd784e5ec163d8d2c9eb1d961eda5d8a9ca50d8a2d169ca8584\": not found" containerID="f4ff6e96db3e5dd784e5ec163d8d2c9eb1d961eda5d8a9ca50d8a2d169ca8584"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.651238    5772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={Type:containerd ID:f4ff6e96db3e5dd784e5ec163d8d2c9eb1d961eda5d8a9ca50d8a2d169ca8584} err="failed to get container status \"f4ff6e96db3e5dd784e5ec163d8d2c9eb1d961eda5d8a9ca50d8a2d169ca8584\": rpc error: code = NotFound desc = an error occurred when try to find container \"f4ff6e96db3e5dd784e5ec163d8d2c9eb1d961eda5d8a9ca50d8a2d169ca8584\": not found"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.651249    5772 scope.go:115] "RemoveContainer" containerID="d6403e0a773151c9a90c0623048a29af564a0d1d5e5400fcd1ee37412238f3ff"
Dec 24 07:36:42 Server k3s[5772]: E1224 07:36:42.651448    5772 remote_runtime.go:415] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"d6403e0a773151c9a90c0623048a29af564a0d1d5e5400fcd1ee37412238f3ff\": not found" containerID="d6403e0a773151c9a90c0623048a29af564a0d1d5e5400fcd1ee37412238f3ff"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.651469    5772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={Type:containerd ID:d6403e0a773151c9a90c0623048a29af564a0d1d5e5400fcd1ee37412238f3ff} err="failed to get container status \"d6403e0a773151c9a90c0623048a29af564a0d1d5e5400fcd1ee37412238f3ff\": rpc error: code = NotFound desc = an error occurred when try to find container \"d6403e0a773151c9a90c0623048a29af564a0d1d5e5400fcd1ee37412238f3ff\": not found"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.651479    5772 scope.go:115] "RemoveContainer" containerID="40bfa901f0a5a6ac35b6b8bb4b45d8da0d828264bb91cc12a22cb6b8588ad658"
Dec 24 07:36:42 Server k3s[5772]: E1224 07:36:42.651679    5772 remote_runtime.go:415] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"40bfa901f0a5a6ac35b6b8bb4b45d8da0d828264bb91cc12a22cb6b8588ad658\": not found" containerID="40bfa901f0a5a6ac35b6b8bb4b45d8da0d828264bb91cc12a22cb6b8588ad658"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.651697    5772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={Type:containerd ID:40bfa901f0a5a6ac35b6b8bb4b45d8da0d828264bb91cc12a22cb6b8588ad658} err="failed to get container status \"40bfa901f0a5a6ac35b6b8bb4b45d8da0d828264bb91cc12a22cb6b8588ad658\": rpc error: code = NotFound desc = an error occurred when try to find container \"40bfa901f0a5a6ac35b6b8bb4b45d8da0d828264bb91cc12a22cb6b8588ad658\": not found"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.651704    5772 scope.go:115] "RemoveContainer" containerID="40bfa901f0a5a6ac35b6b8bb4b45d8da0d828264bb91cc12a22cb6b8588ad658"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.651892    5772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={Type:containerd ID:40bfa901f0a5a6ac35b6b8bb4b45d8da0d828264bb91cc12a22cb6b8588ad658} err="failed to get container status \"40bfa901f0a5a6ac35b6b8bb4b45d8da0d828264bb91cc12a22cb6b8588ad658\": rpc error: code = NotFound desc = an error occurred when try to find container \"40bfa901f0a5a6ac35b6b8bb4b45d8da0d828264bb91cc12a22cb6b8588ad658\": not found"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.651904    5772 scope.go:115] "RemoveContainer" containerID="d6403e0a773151c9a90c0623048a29af564a0d1d5e5400fcd1ee37412238f3ff"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.652177    5772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={Type:containerd ID:d6403e0a773151c9a90c0623048a29af564a0d1d5e5400fcd1ee37412238f3ff} err="failed to get container status \"d6403e0a773151c9a90c0623048a29af564a0d1d5e5400fcd1ee37412238f3ff\": rpc error: code = NotFound desc = an error occurred when try to find container \"d6403e0a773151c9a90c0623048a29af564a0d1d5e5400fcd1ee37412238f3ff\": not found"
Dec 24 07:36:42 Server k3s[5772]: I1224 07:36:42.698240    5772 event.go:294] "Event occurred" object="ix-truenas" fieldPath="" kind="Node" apiVersion="" type="Normal" reason="NodePasswordValidationComplete" message="Deferred node password secret validation complete"
Dec 24 07:36:43 Server k3s[5772]: time="2023-12-24T07:36:43+01:00" level=info msg="Stopped tunnel to 127.0.0.1:6443"
Dec 24 07:36:43 Server k3s[5772]: time="2023-12-24T07:36:43+01:00" level=info msg="Proxy done" err="context canceled" url="wss://127.0.0.1:6443/v1-k3s/connect"
Dec 24 07:36:43 Server k3s[5772]: time="2023-12-24T07:36:43+01:00" level=info msg="Connecting to proxy" url="wss://192.168.3.5:6443/v1-k3s/connect"
Dec 24 07:36:43 Server k3s[5772]: time="2023-12-24T07:36:43+01:00" level=info msg="error in remotedialer server [400]: websocket: close 1006 (abnormal closure): unexpected EOF"
Dec 24 07:36:43 Server k3s[5772]: time="2023-12-24T07:36:43+01:00" level=info msg="Handling backend connection request [ix-truenas]"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.269037    5772 apiserver.go:52] "Watching apiserver"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.272150    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.272506    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.272596    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.272896    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.273104    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.273175    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.273231    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.273271    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.273310    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.273345    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.273608    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.273733    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.273872    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.274134    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.274189    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.312430    5772 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.394432    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9464b12d-9714-4068-b415-258e022c53c3-socket-dir\") pod \"csi-nfs-node-8dc5q\" (UID: \"9464b12d-9714-4068-b415-258e022c53c3\") " pod="kube-system/csi-nfs-node-8dc5q"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.394459    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pods-mount-dir\" (UniqueName: \"kubernetes.io/host-path/9464b12d-9714-4068-b415-258e022c53c3-pods-mount-dir\") pod \"csi-nfs-node-8dc5q\" (UID: \"9464b12d-9714-4068-b415-258e022c53c3\") " pod="kube-system/csi-nfs-node-8dc5q"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.394904    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9f7k\" (UniqueName: \"kubernetes.io/projected/6e8514be-5f5a-4378-b550-3504fedd7642-kube-api-access-p9f7k\") pod \"csi-smb-node-ljmtn\" (UID: \"6e8514be-5f5a-4378-b550-3504fedd7642\") " pod="kube-system/csi-smb-node-ljmtn"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.394954    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/281b688e-7c75-47b8-aeb8-234d997059af-socket-dir\") pod \"csi-nfs-controller-7b74694749-vldc4\" (UID: \"281b688e-7c75-47b8-aeb8-234d997059af\") " pod="kube-system/csi-nfs-controller-7b74694749-vldc4"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395023    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26swq\" (UniqueName: \"kubernetes.io/projected/013e3d9d-e200-4b53-b56a-e762a9decad5-kube-api-access-26swq\") pod \"openebs-zfs-node-bx29q\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") " pod="kube-system/openebs-zfs-node-bx29q"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395064    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9464b12d-9714-4068-b415-258e022c53c3-registration-dir\") pod \"csi-nfs-node-8dc5q\" (UID: \"9464b12d-9714-4068-b415-258e022c53c3\") " pod="kube-system/csi-nfs-node-8dc5q"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395106    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dnsmasq\" (UniqueName: \"kubernetes.io/host-path/70cfd1eb-640c-41b3-a322-48303d01ca4d-dnsmasq\") pod \"pihole-ff558c9db-z9l5k\" (UID: \"70cfd1eb-640c-41b3-a322-48303d01ca4d\") " pod="ix-pihole/pihole-ff558c9db-z9l5k"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395156    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l46l\" (UniqueName: \"kubernetes.io/projected/70cfd1eb-640c-41b3-a322-48303d01ca4d-kube-api-access-8l46l\") pod \"pihole-ff558c9db-z9l5k\" (UID: \"70cfd1eb-640c-41b3-a322-48303d01ca4d\") " pod="ix-pihole/pihole-ff558c9db-z9l5k"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395191    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/8794e834-b432-4b66-872d-5406ffc0ea25-socket-dir\") pod \"csi-nfs-controller-7b74694749-h72f9\" (UID: \"8794e834-b432-4b66-872d-5406ffc0ea25\") " pod="kube-system/csi-nfs-controller-7b74694749-h72f9"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395229    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpd2f\" (UniqueName: \"kubernetes.io/projected/281b688e-7c75-47b8-aeb8-234d997059af-kube-api-access-gpd2f\") pod \"csi-nfs-controller-7b74694749-vldc4\" (UID: \"281b688e-7c75-47b8-aeb8-234d997059af\") " pod="kube-system/csi-nfs-controller-7b74694749-vldc4"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395266    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-root\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-host-root\") pod \"openebs-zfs-node-bx29q\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") " pod="kube-system/openebs-zfs-node-bx29q"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395307    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"chroot-zfs\" (UniqueName: \"kubernetes.io/configmap/013e3d9d-e200-4b53-b56a-e762a9decad5-chroot-zfs\") pod \"openebs-zfs-node-bx29q\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") " pod="kube-system/openebs-zfs-node-bx29q"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395343    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5lcb\" (UniqueName: \"kubernetes.io/projected/ec0104ce-ec27-4b74-87d8-78cec5066e59-kube-api-access-n5lcb\") pod \"openebs-zfs-controller-0\" (UID: \"ec0104ce-ec27-4b74-87d8-78cec5066e59\") " pod="kube-system/openebs-zfs-controller-0"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395373    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sndrs\" (UniqueName: \"kubernetes.io/projected/9464b12d-9714-4068-b415-258e022c53c3-kube-api-access-sndrs\") pod \"csi-nfs-node-8dc5q\" (UID: \"9464b12d-9714-4068-b415-258e022c53c3\") " pod="kube-system/csi-nfs-node-8dc5q"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395394    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84vq7\" (UniqueName: \"kubernetes.io/projected/8794e834-b432-4b66-872d-5406ffc0ea25-kube-api-access-84vq7\") pod \"csi-nfs-controller-7b74694749-h72f9\" (UID: \"8794e834-b432-4b66-872d-5406ffc0ea25\") " pod="kube-system/csi-nfs-controller-7b74694749-h72f9"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395424    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dnsmasq\" (UniqueName: \"kubernetes.io/host-path/d12c8291-0283-4e24-b00a-251ce079dfcc-dnsmasq\") pod \"pihole-ff558c9db-7wnhn\" (UID: \"d12c8291-0283-4e24-b00a-251ce079dfcc\") " pod="ix-pihole/pihole-ff558c9db-7wnhn"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395451    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encr-keys\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-encr-keys\") pod \"openebs-zfs-node-bx29q\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") " pod="kube-system/openebs-zfs-node-bx29q"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395471    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpsww\" (UniqueName: \"kubernetes.io/projected/b65ae8ba-644a-4b60-b64c-eb90ef7c09ad-kube-api-access-lpsww\") pod \"snapshot-controller-546868dfb4-vxcw2\" (UID: \"b65ae8ba-644a-4b60-b64c-eb90ef7c09ad\") " pod="kube-system/snapshot-controller-546868dfb4-vxcw2"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395508    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-462sr\" (UniqueName: \"kubernetes.io/projected/72100aad-f257-4df2-89aa-369c77e3439b-kube-api-access-462sr\") pod \"coredns-59b4f5bbd5-tbbkx\" (UID: \"72100aad-f257-4df2-89aa-369c77e3439b\") " pod="kube-system/coredns-59b4f5bbd5-tbbkx"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395531    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6e8514be-5f5a-4378-b550-3504fedd7642-registration-dir\") pod \"csi-smb-node-ljmtn\" (UID: \"6e8514be-5f5a-4378-b550-3504fedd7642\") " pod="kube-system/csi-smb-node-ljmtn"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395569    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpw4s\" (UniqueName: \"kubernetes.io/projected/14f02546-7f11-4aa1-aeda-d8e9f10ecbf2-kube-api-access-bpw4s\") pod \"csi-smb-controller-7fbbb8fb6f-rcntv\" (UID: \"14f02546-7f11-4aa1-aeda-d8e9f10ecbf2\") " pod="kube-system/csi-smb-controller-7fbbb8fb6f-rcntv"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395594    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/70cfd1eb-640c-41b3-a322-48303d01ca4d-config\") pod \"pihole-ff558c9db-z9l5k\" (UID: \"70cfd1eb-640c-41b3-a322-48303d01ca4d\") " pod="ix-pihole/pihole-ff558c9db-z9l5k"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395613    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdgpk\" (UniqueName: \"kubernetes.io/projected/d12c8291-0283-4e24-b00a-251ce079dfcc-kube-api-access-kdgpk\") pod \"pihole-ff558c9db-7wnhn\" (UID: \"d12c8291-0283-4e24-b00a-251ce079dfcc\") " pod="ix-pihole/pihole-ff558c9db-7wnhn"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395630    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/ec0104ce-ec27-4b74-87d8-78cec5066e59-socket-dir\") pod \"openebs-zfs-controller-0\" (UID: \"ec0104ce-ec27-4b74-87d8-78cec5066e59\") " pod="kube-system/openebs-zfs-controller-0"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395655    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-config-volume\") pod \"coredns-59b4f5bbd5-tbbkx\" (UID: \"72100aad-f257-4df2-89aa-369c77e3439b\") " pod="kube-system/coredns-59b4f5bbd5-tbbkx"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395684    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfdmq\" (UniqueName: \"kubernetes.io/projected/0cadcc5e-31e6-4b41-ab31-6f13326debe1-kube-api-access-kfdmq\") pod \"csi-smb-controller-7fbbb8fb6f-dwvwr\" (UID: \"0cadcc5e-31e6-4b41-ab31-6f13326debe1\") " pod="kube-system/csi-smb-controller-7fbbb8fb6f-dwvwr"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395735    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24wjw\" (UniqueName: \"kubernetes.io/projected/5e2592f9-ea55-4817-8872-c5d937a57ef6-kube-api-access-24wjw\") pod \"snapshot-controller-546868dfb4-2z962\" (UID: \"5e2592f9-ea55-4817-8872-c5d937a57ef6\") " pod="kube-system/snapshot-controller-546868dfb4-2z962"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395776    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw7wd\" (UniqueName: \"kubernetes.io/projected/f7daa893-7b06-453f-9300-2a183e5f6d33-kube-api-access-gw7wd\") pod \"snapshot-controller-546868dfb4-48kvb\" (UID: \"f7daa893-7b06-453f-9300-2a183e5f6d33\") " pod="kube-system/snapshot-controller-546868dfb4-48kvb"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395792    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/d12c8291-0283-4e24-b00a-251ce079dfcc-config\") pod \"pihole-ff558c9db-7wnhn\" (UID: \"d12c8291-0283-4e24-b00a-251ce079dfcc\") " pod="ix-pihole/pihole-ff558c9db-7wnhn"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395810    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/0cadcc5e-31e6-4b41-ab31-6f13326debe1-socket-dir\") pod \"csi-smb-controller-7fbbb8fb6f-dwvwr\" (UID: \"0cadcc5e-31e6-4b41-ab31-6f13326debe1\") " pod="kube-system/csi-smb-controller-7fbbb8fb6f-dwvwr"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395825    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6e8514be-5f5a-4378-b550-3504fedd7642-socket-dir\") pod \"csi-smb-node-ljmtn\" (UID: \"6e8514be-5f5a-4378-b550-3504fedd7642\") " pod="kube-system/csi-smb-node-ljmtn"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395863    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6e8514be-5f5a-4378-b550-3504fedd7642-mountpoint-dir\") pod \"csi-smb-node-ljmtn\" (UID: \"6e8514be-5f5a-4378-b550-3504fedd7642\") " pod="kube-system/csi-smb-node-ljmtn"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395884    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pods-mount-dir\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-pods-mount-dir\") pod \"openebs-zfs-node-bx29q\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") " pod="kube-system/openebs-zfs-node-bx29q"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395909    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/14f02546-7f11-4aa1-aeda-d8e9f10ecbf2-socket-dir\") pod \"csi-smb-controller-7fbbb8fb6f-rcntv\" (UID: \"14f02546-7f11-4aa1-aeda-d8e9f10ecbf2\") " pod="kube-system/csi-smb-controller-7fbbb8fb6f-rcntv"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395929    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-config-volume\" (UniqueName: \"kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-custom-config-volume\") pod \"coredns-59b4f5bbd5-tbbkx\" (UID: \"72100aad-f257-4df2-89aa-369c77e3439b\") " pod="kube-system/coredns-59b4f5bbd5-tbbkx"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395947    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-registration-dir\") pod \"openebs-zfs-node-bx29q\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") " pod="kube-system/openebs-zfs-node-bx29q"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395970    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-plugin-dir\") pod \"openebs-zfs-node-bx29q\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") " pod="kube-system/openebs-zfs-node-bx29q"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.395989    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pods-mount-dir\" (UniqueName: \"kubernetes.io/host-path/8794e834-b432-4b66-872d-5406ffc0ea25-pods-mount-dir\") pod \"csi-nfs-controller-7b74694749-h72f9\" (UID: \"8794e834-b432-4b66-872d-5406ffc0ea25\") " pod="kube-system/csi-nfs-controller-7b74694749-h72f9"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.396015    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw289\" (UniqueName: \"kubernetes.io/projected/d19f51ba-3f7e-45a0-ad35-f6abb0cf0b28-kube-api-access-tw289\") pod \"snapshot-controller-546868dfb4-7xxcd\" (UID: \"d19f51ba-3f7e-45a0-ad35-f6abb0cf0b28\") " pod="kube-system/snapshot-controller-546868dfb4-7xxcd"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.396038    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pods-mount-dir\" (UniqueName: \"kubernetes.io/host-path/281b688e-7c75-47b8-aeb8-234d997059af-pods-mount-dir\") pod \"csi-nfs-controller-7b74694749-vldc4\" (UID: \"281b688e-7c75-47b8-aeb8-234d997059af\") " pod="kube-system/csi-nfs-controller-7b74694749-vldc4"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.396058    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-device-dir\") pod \"openebs-zfs-node-bx29q\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") " pod="kube-system/openebs-zfs-node-bx29q"
Dec 24 07:36:43 Server k3s[5772]: I1224 07:36:43.396081    5772 reconciler.go:41] "Reconciler: start to sync state"
Dec 24 07:36:43 Server k3s[5772]: E1224 07:36:43.499958    5772 configmap.go:199] Couldn't get configMap kube-system/coredns: object "kube-system"/"coredns" not registered
Dec 24 07:36:43 Server k3s[5772]: E1224 07:36:43.500057    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-config-volume podName:72100aad-f257-4df2-89aa-369c77e3439b nodeName:}" failed. No retries permitted until 2023-12-24 07:36:43.999994362 +0100 CET m=+8.609949353 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-config-volume") pod "coredns-59b4f5bbd5-tbbkx" (UID: "72100aad-f257-4df2-89aa-369c77e3439b") : object "kube-system"/"coredns" not registered
Dec 24 07:36:43 Server k3s[5772]: E1224 07:36:43.500632    5772 configmap.go:199] Couldn't get configMap kube-system/openebs-zfspv-bin: object "kube-system"/"openebs-zfspv-bin" not registered
Dec 24 07:36:43 Server k3s[5772]: E1224 07:36:43.500665    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/013e3d9d-e200-4b53-b56a-e762a9decad5-chroot-zfs podName:013e3d9d-e200-4b53-b56a-e762a9decad5 nodeName:}" failed. No retries permitted until 2023-12-24 07:36:44.000654724 +0100 CET m=+8.610609711 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "chroot-zfs" (UniqueName: "kubernetes.io/configmap/013e3d9d-e200-4b53-b56a-e762a9decad5-chroot-zfs") pod "openebs-zfs-node-bx29q" (UID: "013e3d9d-e200-4b53-b56a-e762a9decad5") : object "kube-system"/"openebs-zfspv-bin" not registered
Dec 24 07:36:43 Server k3s[5772]: E1224 07:36:43.500674    5772 configmap.go:199] Couldn't get configMap kube-system/coredns-custom: object "kube-system"/"coredns-custom" not registered
Dec 24 07:36:43 Server k3s[5772]: E1224 07:36:43.500718    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-custom-config-volume podName:72100aad-f257-4df2-89aa-369c77e3439b nodeName:}" failed. No retries permitted until 2023-12-24 07:36:44.000704896 +0100 CET m=+8.610659883 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "custom-config-volume" (UniqueName: "kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-custom-config-volume") pod "coredns-59b4f5bbd5-tbbkx" (UID: "72100aad-f257-4df2-89aa-369c77e3439b") : object "kube-system"/"coredns-custom" not registered
Dec 24 07:36:44 Server k3s[5772]: E1224 07:36:44.002689    5772 configmap.go:199] Couldn't get configMap kube-system/openebs-zfspv-bin: object "kube-system"/"openebs-zfspv-bin" not registered
Dec 24 07:36:44 Server k3s[5772]: E1224 07:36:44.002704    5772 configmap.go:199] Couldn't get configMap kube-system/coredns-custom: object "kube-system"/"coredns-custom" not registered
Dec 24 07:36:44 Server k3s[5772]: E1224 07:36:44.002738    5772 configmap.go:199] Couldn't get configMap kube-system/coredns: object "kube-system"/"coredns" not registered
Dec 24 07:36:44 Server k3s[5772]: E1224 07:36:44.002751    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/013e3d9d-e200-4b53-b56a-e762a9decad5-chroot-zfs podName:013e3d9d-e200-4b53-b56a-e762a9decad5 nodeName:}" failed. No retries permitted until 2023-12-24 07:36:45.002740464 +0100 CET m=+9.612695462 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "chroot-zfs" (UniqueName: "kubernetes.io/configmap/013e3d9d-e200-4b53-b56a-e762a9decad5-chroot-zfs") pod "openebs-zfs-node-bx29q" (UID: "013e3d9d-e200-4b53-b56a-e762a9decad5") : object "kube-system"/"openebs-zfspv-bin" not registered
Dec 24 07:36:44 Server k3s[5772]: E1224 07:36:44.002768    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-config-volume podName:72100aad-f257-4df2-89aa-369c77e3439b nodeName:}" failed. No retries permitted until 2023-12-24 07:36:45.002758981 +0100 CET m=+9.612713974 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-config-volume") pod "coredns-59b4f5bbd5-tbbkx" (UID: "72100aad-f257-4df2-89aa-369c77e3439b") : object "kube-system"/"coredns" not registered
Dec 24 07:36:44 Server k3s[5772]: E1224 07:36:44.002783    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-custom-config-volume podName:72100aad-f257-4df2-89aa-369c77e3439b nodeName:}" failed. No retries permitted until 2023-12-24 07:36:45.002776137 +0100 CET m=+9.612731125 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "custom-config-volume" (UniqueName: "kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-custom-config-volume") pod "coredns-59b4f5bbd5-tbbkx" (UID: "72100aad-f257-4df2-89aa-369c77e3439b") : object "kube-system"/"coredns-custom" not registered
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.604988    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"dnsmasq\" (UniqueName: \"kubernetes.io/host-path/70cfd1eb-640c-41b3-a322-48303d01ca4d-dnsmasq\") pod \"70cfd1eb-640c-41b3-a322-48303d01ca4d\" (UID: \"70cfd1eb-640c-41b3-a322-48303d01ca4d\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605017    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-462sr\" (UniqueName: \"kubernetes.io/projected/72100aad-f257-4df2-89aa-369c77e3439b-kube-api-access-462sr\") pod \"72100aad-f257-4df2-89aa-369c77e3439b\" (UID: \"72100aad-f257-4df2-89aa-369c77e3439b\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605060    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/70cfd1eb-640c-41b3-a322-48303d01ca4d-config\") pod \"70cfd1eb-640c-41b3-a322-48303d01ca4d\" (UID: \"70cfd1eb-640c-41b3-a322-48303d01ca4d\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605077    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfdmq\" (UniqueName: \"kubernetes.io/projected/0cadcc5e-31e6-4b41-ab31-6f13326debe1-kube-api-access-kfdmq\") pod \"0cadcc5e-31e6-4b41-ab31-6f13326debe1\" (UID: \"0cadcc5e-31e6-4b41-ab31-6f13326debe1\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605096    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/14f02546-7f11-4aa1-aeda-d8e9f10ecbf2-socket-dir\") pod \"14f02546-7f11-4aa1-aeda-d8e9f10ecbf2\" (UID: \"14f02546-7f11-4aa1-aeda-d8e9f10ecbf2\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605114    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/281b688e-7c75-47b8-aeb8-234d997059af-socket-dir\") pod \"281b688e-7c75-47b8-aeb8-234d997059af\" (UID: \"281b688e-7c75-47b8-aeb8-234d997059af\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605134    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpw4s\" (UniqueName: \"kubernetes.io/projected/14f02546-7f11-4aa1-aeda-d8e9f10ecbf2-kube-api-access-bpw4s\") pod \"14f02546-7f11-4aa1-aeda-d8e9f10ecbf2\" (UID: \"14f02546-7f11-4aa1-aeda-d8e9f10ecbf2\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605150    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-plugin-dir\") pod \"013e3d9d-e200-4b53-b56a-e762a9decad5\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605165    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"encr-keys\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-encr-keys\") pod \"013e3d9d-e200-4b53-b56a-e762a9decad5\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605180    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/d12c8291-0283-4e24-b00a-251ce079dfcc-config\") pod \"d12c8291-0283-4e24-b00a-251ce079dfcc\" (UID: \"d12c8291-0283-4e24-b00a-251ce079dfcc\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605196    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-device-dir\") pod \"013e3d9d-e200-4b53-b56a-e762a9decad5\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605211    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-registration-dir\") pod \"013e3d9d-e200-4b53-b56a-e762a9decad5\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605227    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"pods-mount-dir\" (UniqueName: \"kubernetes.io/host-path/281b688e-7c75-47b8-aeb8-234d997059af-pods-mount-dir\") pod \"281b688e-7c75-47b8-aeb8-234d997059af\" (UID: \"281b688e-7c75-47b8-aeb8-234d997059af\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605241    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"host-root\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-host-root\") pod \"013e3d9d-e200-4b53-b56a-e762a9decad5\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605258    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/ec0104ce-ec27-4b74-87d8-78cec5066e59-socket-dir\") pod \"ec0104ce-ec27-4b74-87d8-78cec5066e59\" (UID: \"ec0104ce-ec27-4b74-87d8-78cec5066e59\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605274    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/8794e834-b432-4b66-872d-5406ffc0ea25-socket-dir\") pod \"8794e834-b432-4b66-872d-5406ffc0ea25\" (UID: \"8794e834-b432-4b66-872d-5406ffc0ea25\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605292    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpsww\" (UniqueName: \"kubernetes.io/projected/b65ae8ba-644a-4b60-b64c-eb90ef7c09ad-kube-api-access-lpsww\") pod \"b65ae8ba-644a-4b60-b64c-eb90ef7c09ad\" (UID: \"b65ae8ba-644a-4b60-b64c-eb90ef7c09ad\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605307    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"dnsmasq\" (UniqueName: \"kubernetes.io/host-path/d12c8291-0283-4e24-b00a-251ce079dfcc-dnsmasq\") pod \"d12c8291-0283-4e24-b00a-251ce079dfcc\" (UID: \"d12c8291-0283-4e24-b00a-251ce079dfcc\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605324    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/0cadcc5e-31e6-4b41-ab31-6f13326debe1-socket-dir\") pod \"0cadcc5e-31e6-4b41-ab31-6f13326debe1\" (UID: \"0cadcc5e-31e6-4b41-ab31-6f13326debe1\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605338    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"pods-mount-dir\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-pods-mount-dir\") pod \"013e3d9d-e200-4b53-b56a-e762a9decad5\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605355    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"pods-mount-dir\" (UniqueName: \"kubernetes.io/host-path/8794e834-b432-4b66-872d-5406ffc0ea25-pods-mount-dir\") pod \"8794e834-b432-4b66-872d-5406ffc0ea25\" (UID: \"8794e834-b432-4b66-872d-5406ffc0ea25\") "
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605401    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70cfd1eb-640c-41b3-a322-48303d01ca4d-dnsmasq" (OuterVolumeSpecName: "dnsmasq") pod "70cfd1eb-640c-41b3-a322-48303d01ca4d" (UID: "70cfd1eb-640c-41b3-a322-48303d01ca4d"). InnerVolumeSpecName "dnsmasq". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605402    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70cfd1eb-640c-41b3-a322-48303d01ca4d-config" (OuterVolumeSpecName: "config") pod "70cfd1eb-640c-41b3-a322-48303d01ca4d" (UID: "70cfd1eb-640c-41b3-a322-48303d01ca4d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605465    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8794e834-b432-4b66-872d-5406ffc0ea25-pods-mount-dir" (OuterVolumeSpecName: "pods-mount-dir") pod "8794e834-b432-4b66-872d-5406ffc0ea25" (UID: "8794e834-b432-4b66-872d-5406ffc0ea25"). InnerVolumeSpecName "pods-mount-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605488    5772 reconciler_common.go:295] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/host-path/70cfd1eb-640c-41b3-a322-48303d01ca4d-config\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605502    5772 reconciler_common.go:295] "Volume detached for volume \"dnsmasq\" (UniqueName: \"kubernetes.io/host-path/70cfd1eb-640c-41b3-a322-48303d01ca4d-dnsmasq\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605569    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d12c8291-0283-4e24-b00a-251ce079dfcc-config" (OuterVolumeSpecName: "config") pod "d12c8291-0283-4e24-b00a-251ce079dfcc" (UID: "d12c8291-0283-4e24-b00a-251ce079dfcc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605642    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d12c8291-0283-4e24-b00a-251ce079dfcc-dnsmasq" (OuterVolumeSpecName: "dnsmasq") pod "d12c8291-0283-4e24-b00a-251ce079dfcc" (UID: "d12c8291-0283-4e24-b00a-251ce079dfcc"). InnerVolumeSpecName "dnsmasq". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: W1224 07:36:44.605667    5772 empty_dir.go:525] Warning: Failed to clear quota on /var/lib/kubelet/pods/8794e834-b432-4b66-872d-5406ffc0ea25/volumes/kubernetes.io~empty-dir/socket-dir: clearQuota called, but quotas disabled
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605727    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8794e834-b432-4b66-872d-5406ffc0ea25-socket-dir" (OuterVolumeSpecName: "socket-dir") pod "8794e834-b432-4b66-872d-5406ffc0ea25" (UID: "8794e834-b432-4b66-872d-5406ffc0ea25"). InnerVolumeSpecName "socket-dir". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605767    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/281b688e-7c75-47b8-aeb8-234d997059af-pods-mount-dir" (OuterVolumeSpecName: "pods-mount-dir") pod "281b688e-7c75-47b8-aeb8-234d997059af" (UID: "281b688e-7c75-47b8-aeb8-234d997059af"). InnerVolumeSpecName "pods-mount-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: W1224 07:36:44.605832    5772 empty_dir.go:525] Warning: Failed to clear quota on /var/lib/kubelet/pods/0cadcc5e-31e6-4b41-ab31-6f13326debe1/volumes/kubernetes.io~empty-dir/socket-dir: clearQuota called, but quotas disabled
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605887    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cadcc5e-31e6-4b41-ab31-6f13326debe1-socket-dir" (OuterVolumeSpecName: "socket-dir") pod "0cadcc5e-31e6-4b41-ab31-6f13326debe1" (UID: "0cadcc5e-31e6-4b41-ab31-6f13326debe1"). InnerVolumeSpecName "socket-dir". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.605996    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-pods-mount-dir" (OuterVolumeSpecName: "pods-mount-dir") pod "013e3d9d-e200-4b53-b56a-e762a9decad5" (UID: "013e3d9d-e200-4b53-b56a-e762a9decad5"). InnerVolumeSpecName "pods-mount-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: W1224 07:36:44.606171    5772 empty_dir.go:525] Warning: Failed to clear quota on /var/lib/kubelet/pods/14f02546-7f11-4aa1-aeda-d8e9f10ecbf2/volumes/kubernetes.io~empty-dir/socket-dir: clearQuota called, but quotas disabled
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.606226    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14f02546-7f11-4aa1-aeda-d8e9f10ecbf2-socket-dir" (OuterVolumeSpecName: "socket-dir") pod "14f02546-7f11-4aa1-aeda-d8e9f10ecbf2" (UID: "14f02546-7f11-4aa1-aeda-d8e9f10ecbf2"). InnerVolumeSpecName "socket-dir". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.606257    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-host-root" (OuterVolumeSpecName: "host-root") pod "013e3d9d-e200-4b53-b56a-e762a9decad5" (UID: "013e3d9d-e200-4b53-b56a-e762a9decad5"). InnerVolumeSpecName "host-root". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.606275    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-plugin-dir" (OuterVolumeSpecName: "plugin-dir") pod "013e3d9d-e200-4b53-b56a-e762a9decad5" (UID: "013e3d9d-e200-4b53-b56a-e762a9decad5"). InnerVolumeSpecName "plugin-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: W1224 07:36:44.606370    5772 empty_dir.go:525] Warning: Failed to clear quota on /var/lib/kubelet/pods/281b688e-7c75-47b8-aeb8-234d997059af/volumes/kubernetes.io~empty-dir/socket-dir: clearQuota called, but quotas disabled
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.606425    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/281b688e-7c75-47b8-aeb8-234d997059af-socket-dir" (OuterVolumeSpecName: "socket-dir") pod "281b688e-7c75-47b8-aeb8-234d997059af" (UID: "281b688e-7c75-47b8-aeb8-234d997059af"). InnerVolumeSpecName "socket-dir". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.606441    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-device-dir" (OuterVolumeSpecName: "device-dir") pod "013e3d9d-e200-4b53-b56a-e762a9decad5" (UID: "013e3d9d-e200-4b53-b56a-e762a9decad5"). InnerVolumeSpecName "device-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.606452    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-registration-dir" (OuterVolumeSpecName: "registration-dir") pod "013e3d9d-e200-4b53-b56a-e762a9decad5" (UID: "013e3d9d-e200-4b53-b56a-e762a9decad5"). InnerVolumeSpecName "registration-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.606457    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-encr-keys" (OuterVolumeSpecName: "encr-keys") pod "013e3d9d-e200-4b53-b56a-e762a9decad5" (UID: "013e3d9d-e200-4b53-b56a-e762a9decad5"). InnerVolumeSpecName "encr-keys". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: W1224 07:36:44.606624    5772 empty_dir.go:525] Warning: Failed to clear quota on /var/lib/kubelet/pods/ec0104ce-ec27-4b74-87d8-78cec5066e59/volumes/kubernetes.io~empty-dir/socket-dir: clearQuota called, but quotas disabled
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.606936    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec0104ce-ec27-4b74-87d8-78cec5066e59-socket-dir" (OuterVolumeSpecName: "socket-dir") pod "ec0104ce-ec27-4b74-87d8-78cec5066e59" (UID: "ec0104ce-ec27-4b74-87d8-78cec5066e59"). InnerVolumeSpecName "socket-dir". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.606958    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72100aad-f257-4df2-89aa-369c77e3439b-kube-api-access-462sr" (OuterVolumeSpecName: "kube-api-access-462sr") pod "72100aad-f257-4df2-89aa-369c77e3439b" (UID: "72100aad-f257-4df2-89aa-369c77e3439b"). InnerVolumeSpecName "kube-api-access-462sr". PluginName "kubernetes.io/projected", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.607142    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b65ae8ba-644a-4b60-b64c-eb90ef7c09ad-kube-api-access-lpsww" (OuterVolumeSpecName: "kube-api-access-lpsww") pod "b65ae8ba-644a-4b60-b64c-eb90ef7c09ad" (UID: "b65ae8ba-644a-4b60-b64c-eb90ef7c09ad"). InnerVolumeSpecName "kube-api-access-lpsww". PluginName "kubernetes.io/projected", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.607169    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cadcc5e-31e6-4b41-ab31-6f13326debe1-kube-api-access-kfdmq" (OuterVolumeSpecName: "kube-api-access-kfdmq") pod "0cadcc5e-31e6-4b41-ab31-6f13326debe1" (UID: "0cadcc5e-31e6-4b41-ab31-6f13326debe1"). InnerVolumeSpecName "kube-api-access-kfdmq". PluginName "kubernetes.io/projected", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.607561    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14f02546-7f11-4aa1-aeda-d8e9f10ecbf2-kube-api-access-bpw4s" (OuterVolumeSpecName: "kube-api-access-bpw4s") pod "14f02546-7f11-4aa1-aeda-d8e9f10ecbf2" (UID: "14f02546-7f11-4aa1-aeda-d8e9f10ecbf2"). InnerVolumeSpecName "kube-api-access-bpw4s". PluginName "kubernetes.io/projected", VolumeGidValue ""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.669199    5772 request.go:690] Waited for 1.168616788s due to client-side throttling, not priority and fairness, request: POST:https://127.0.0.1:6443/api/v1/namespaces/ix-pihole/serviceaccounts/default/token
Dec 24 07:36:44 Server k3s[5772]: E1224 07:36:44.673595    5772 projected.go:292] Couldn't get configMap ix-pihole/kube-root-ca.crt: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:44 Server k3s[5772]: E1224 07:36:44.673610    5772 projected.go:198] Error preparing data for projected volume kube-api-access-8l46l for pod ix-pihole/pihole-ff558c9db-z9l5k: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:44 Server k3s[5772]: E1224 07:36:44.673650    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/70cfd1eb-640c-41b3-a322-48303d01ca4d-kube-api-access-8l46l podName:70cfd1eb-640c-41b3-a322-48303d01ca4d nodeName:}" failed. No retries permitted until 2023-12-24 07:36:45.173638593 +0100 CET m=+9.783593580 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-8l46l" (UniqueName: "kubernetes.io/projected/70cfd1eb-640c-41b3-a322-48303d01ca4d-kube-api-access-8l46l") pod "pihole-ff558c9db-z9l5k" (UID: "70cfd1eb-640c-41b3-a322-48303d01ca4d") : object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706297    5772 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-bpw4s\" (UniqueName: \"kubernetes.io/projected/14f02546-7f11-4aa1-aeda-d8e9f10ecbf2-kube-api-access-bpw4s\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706314    5772 reconciler_common.go:295] "Volume detached for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-plugin-dir\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706324    5772 reconciler_common.go:295] "Volume detached for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-device-dir\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706333    5772 reconciler_common.go:295] "Volume detached for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-registration-dir\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706342    5772 reconciler_common.go:295] "Volume detached for volume \"pods-mount-dir\" (UniqueName: \"kubernetes.io/host-path/281b688e-7c75-47b8-aeb8-234d997059af-pods-mount-dir\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706350    5772 reconciler_common.go:295] "Volume detached for volume \"encr-keys\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-encr-keys\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706358    5772 reconciler_common.go:295] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/host-path/d12c8291-0283-4e24-b00a-251ce079dfcc-config\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706367    5772 reconciler_common.go:295] "Volume detached for volume \"host-root\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-host-root\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706375    5772 reconciler_common.go:295] "Volume detached for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/ec0104ce-ec27-4b74-87d8-78cec5066e59-socket-dir\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706384    5772 reconciler_common.go:295] "Volume detached for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/8794e834-b432-4b66-872d-5406ffc0ea25-socket-dir\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706392    5772 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-lpsww\" (UniqueName: \"kubernetes.io/projected/b65ae8ba-644a-4b60-b64c-eb90ef7c09ad-kube-api-access-lpsww\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706401    5772 reconciler_common.go:295] "Volume detached for volume \"dnsmasq\" (UniqueName: \"kubernetes.io/host-path/d12c8291-0283-4e24-b00a-251ce079dfcc-dnsmasq\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706409    5772 reconciler_common.go:295] "Volume detached for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/0cadcc5e-31e6-4b41-ab31-6f13326debe1-socket-dir\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706417    5772 reconciler_common.go:295] "Volume detached for volume \"pods-mount-dir\" (UniqueName: \"kubernetes.io/host-path/013e3d9d-e200-4b53-b56a-e762a9decad5-pods-mount-dir\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706426    5772 reconciler_common.go:295] "Volume detached for volume \"pods-mount-dir\" (UniqueName: \"kubernetes.io/host-path/8794e834-b432-4b66-872d-5406ffc0ea25-pods-mount-dir\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706436    5772 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-462sr\" (UniqueName: \"kubernetes.io/projected/72100aad-f257-4df2-89aa-369c77e3439b-kube-api-access-462sr\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706444    5772 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-kfdmq\" (UniqueName: \"kubernetes.io/projected/0cadcc5e-31e6-4b41-ab31-6f13326debe1-kube-api-access-kfdmq\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706453    5772 reconciler_common.go:295] "Volume detached for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/14f02546-7f11-4aa1-aeda-d8e9f10ecbf2-socket-dir\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.706461    5772 reconciler_common.go:295] "Volume detached for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/281b688e-7c75-47b8-aeb8-234d997059af-socket-dir\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:44 Server k3s[5772]: I1224 07:36:44.899285    5772 kubelet_node_status.go:493] "Fast updating node status as it just became ready"
Dec 24 07:36:45 Server k3s[5772]: I1224 07:36:45.008723    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26swq\" (UniqueName: \"kubernetes.io/projected/013e3d9d-e200-4b53-b56a-e762a9decad5-kube-api-access-26swq\") pod \"013e3d9d-e200-4b53-b56a-e762a9decad5\" (UID: \"013e3d9d-e200-4b53-b56a-e762a9decad5\") "
Dec 24 07:36:45 Server k3s[5772]: E1224 07:36:45.008895    5772 configmap.go:199] Couldn't get configMap kube-system/coredns-custom: object "kube-system"/"coredns-custom" not registered
Dec 24 07:36:45 Server k3s[5772]: E1224 07:36:45.008933    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-custom-config-volume podName:72100aad-f257-4df2-89aa-369c77e3439b nodeName:}" failed. No retries permitted until 2023-12-24 07:36:47.008922311 +0100 CET m=+11.618877304 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "custom-config-volume" (UniqueName: "kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-custom-config-volume") pod "coredns-59b4f5bbd5-tbbkx" (UID: "72100aad-f257-4df2-89aa-369c77e3439b") : object "kube-system"/"coredns-custom" not registered
Dec 24 07:36:45 Server k3s[5772]: E1224 07:36:45.008971    5772 configmap.go:199] Couldn't get configMap kube-system/openebs-zfspv-bin: object "kube-system"/"openebs-zfspv-bin" not registered
Dec 24 07:36:45 Server k3s[5772]: E1224 07:36:45.008992    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/013e3d9d-e200-4b53-b56a-e762a9decad5-chroot-zfs podName:013e3d9d-e200-4b53-b56a-e762a9decad5 nodeName:}" failed. No retries permitted until 2023-12-24 07:36:47.008985205 +0100 CET m=+11.618940192 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "chroot-zfs" (UniqueName: "kubernetes.io/configmap/013e3d9d-e200-4b53-b56a-e762a9decad5-chroot-zfs") pod "openebs-zfs-node-bx29q" (UID: "013e3d9d-e200-4b53-b56a-e762a9decad5") : object "kube-system"/"openebs-zfspv-bin" not registered
Dec 24 07:36:45 Server k3s[5772]: E1224 07:36:45.009020    5772 configmap.go:199] Couldn't get configMap kube-system/coredns: object "kube-system"/"coredns" not registered
Dec 24 07:36:45 Server k3s[5772]: E1224 07:36:45.009039    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-config-volume podName:72100aad-f257-4df2-89aa-369c77e3439b nodeName:}" failed. No retries permitted until 2023-12-24 07:36:47.009032932 +0100 CET m=+11.618987920 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-config-volume") pod "coredns-59b4f5bbd5-tbbkx" (UID: "72100aad-f257-4df2-89aa-369c77e3439b") : object "kube-system"/"coredns" not registered
Dec 24 07:36:45 Server k3s[5772]: I1224 07:36:45.010064    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/013e3d9d-e200-4b53-b56a-e762a9decad5-kube-api-access-26swq" (OuterVolumeSpecName: "kube-api-access-26swq") pod "013e3d9d-e200-4b53-b56a-e762a9decad5" (UID: "013e3d9d-e200-4b53-b56a-e762a9decad5"). InnerVolumeSpecName "kube-api-access-26swq". PluginName "kubernetes.io/projected", VolumeGidValue ""
Dec 24 07:36:45 Server k3s[5772]: I1224 07:36:45.073959    5772 controller.go:615] quota admission added evaluator for: daemonsets.apps
Dec 24 07:36:45 Server k3s[5772]: I1224 07:36:45.109251    5772 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-26swq\" (UniqueName: \"kubernetes.io/projected/013e3d9d-e200-4b53-b56a-e762a9decad5-kube-api-access-26swq\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:45 Server k3s[5772]: time="2023-12-24T07:36:45+01:00" level=info msg="Tunnel authorizer set Kubelet Port 10250"
Dec 24 07:36:45 Server k3s[5772]: I1224 07:36:45.209566    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5lcb\" (UniqueName: \"kubernetes.io/projected/ec0104ce-ec27-4b74-87d8-78cec5066e59-kube-api-access-n5lcb\") pod \"ec0104ce-ec27-4b74-87d8-78cec5066e59\" (UID: \"ec0104ce-ec27-4b74-87d8-78cec5066e59\") "
Dec 24 07:36:45 Server k3s[5772]: E1224 07:36:45.209751    5772 projected.go:292] Couldn't get configMap ix-pihole/kube-root-ca.crt: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:45 Server k3s[5772]: E1224 07:36:45.209765    5772 projected.go:198] Error preparing data for projected volume kube-api-access-8l46l for pod ix-pihole/pihole-ff558c9db-z9l5k: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:45 Server k3s[5772]: E1224 07:36:45.209803    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/70cfd1eb-640c-41b3-a322-48303d01ca4d-kube-api-access-8l46l podName:70cfd1eb-640c-41b3-a322-48303d01ca4d nodeName:}" failed. No retries permitted until 2023-12-24 07:36:46.209790804 +0100 CET m=+10.819745792 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-8l46l" (UniqueName: "kubernetes.io/projected/70cfd1eb-640c-41b3-a322-48303d01ca4d-kube-api-access-8l46l") pod "pihole-ff558c9db-z9l5k" (UID: "70cfd1eb-640c-41b3-a322-48303d01ca4d") : object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:45 Server k3s[5772]: I1224 07:36:45.210667    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec0104ce-ec27-4b74-87d8-78cec5066e59-kube-api-access-n5lcb" (OuterVolumeSpecName: "kube-api-access-n5lcb") pod "ec0104ce-ec27-4b74-87d8-78cec5066e59" (UID: "ec0104ce-ec27-4b74-87d8-78cec5066e59"). InnerVolumeSpecName "kube-api-access-n5lcb". PluginName "kubernetes.io/projected", VolumeGidValue ""
Dec 24 07:36:45 Server k3s[5772]: I1224 07:36:45.310058    5772 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-n5lcb\" (UniqueName: \"kubernetes.io/projected/ec0104ce-ec27-4b74-87d8-78cec5066e59-kube-api-access-n5lcb\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:45 Server k3s[5772]: I1224 07:36:45.611177    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84vq7\" (UniqueName: \"kubernetes.io/projected/8794e834-b432-4b66-872d-5406ffc0ea25-kube-api-access-84vq7\") pod \"8794e834-b432-4b66-872d-5406ffc0ea25\" (UID: \"8794e834-b432-4b66-872d-5406ffc0ea25\") "
Dec 24 07:36:45 Server k3s[5772]: I1224 07:36:45.612349    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8794e834-b432-4b66-872d-5406ffc0ea25-kube-api-access-84vq7" (OuterVolumeSpecName: "kube-api-access-84vq7") pod "8794e834-b432-4b66-872d-5406ffc0ea25" (UID: "8794e834-b432-4b66-872d-5406ffc0ea25"). InnerVolumeSpecName "kube-api-access-84vq7". PluginName "kubernetes.io/projected", VolumeGidValue ""
Dec 24 07:36:45 Server k3s[5772]: I1224 07:36:45.711989    5772 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-84vq7\" (UniqueName: \"kubernetes.io/projected/8794e834-b432-4b66-872d-5406ffc0ea25-kube-api-access-84vq7\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:45 Server k3s[5772]: I1224 07:36:45.812986    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpd2f\" (UniqueName: \"kubernetes.io/projected/281b688e-7c75-47b8-aeb8-234d997059af-kube-api-access-gpd2f\") pod \"281b688e-7c75-47b8-aeb8-234d997059af\" (UID: \"281b688e-7c75-47b8-aeb8-234d997059af\") "
Dec 24 07:36:45 Server k3s[5772]: I1224 07:36:45.814280    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/281b688e-7c75-47b8-aeb8-234d997059af-kube-api-access-gpd2f" (OuterVolumeSpecName: "kube-api-access-gpd2f") pod "281b688e-7c75-47b8-aeb8-234d997059af" (UID: "281b688e-7c75-47b8-aeb8-234d997059af"). InnerVolumeSpecName "kube-api-access-gpd2f". PluginName "kubernetes.io/projected", VolumeGidValue ""
Dec 24 07:36:45 Server k3s[5772]: I1224 07:36:45.913444    5772 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-gpd2f\" (UniqueName: \"kubernetes.io/projected/281b688e-7c75-47b8-aeb8-234d997059af-kube-api-access-gpd2f\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:46 Server k3s[5772]: I1224 07:36:46.014445    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw7wd\" (UniqueName: \"kubernetes.io/projected/f7daa893-7b06-453f-9300-2a183e5f6d33-kube-api-access-gw7wd\") pod \"f7daa893-7b06-453f-9300-2a183e5f6d33\" (UID: \"f7daa893-7b06-453f-9300-2a183e5f6d33\") "
Dec 24 07:36:46 Server k3s[5772]: I1224 07:36:46.015603    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7daa893-7b06-453f-9300-2a183e5f6d33-kube-api-access-gw7wd" (OuterVolumeSpecName: "kube-api-access-gw7wd") pod "f7daa893-7b06-453f-9300-2a183e5f6d33" (UID: "f7daa893-7b06-453f-9300-2a183e5f6d33"). InnerVolumeSpecName "kube-api-access-gw7wd". PluginName "kubernetes.io/projected", VolumeGidValue ""
Dec 24 07:36:46 Server k3s[5772]: I1224 07:36:46.115211    5772 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-gw7wd\" (UniqueName: \"kubernetes.io/projected/f7daa893-7b06-453f-9300-2a183e5f6d33-kube-api-access-gw7wd\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:46 Server k3s[5772]: I1224 07:36:46.215718    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw289\" (UniqueName: \"kubernetes.io/projected/d19f51ba-3f7e-45a0-ad35-f6abb0cf0b28-kube-api-access-tw289\") pod \"d19f51ba-3f7e-45a0-ad35-f6abb0cf0b28\" (UID: \"d19f51ba-3f7e-45a0-ad35-f6abb0cf0b28\") "
Dec 24 07:36:46 Server k3s[5772]: E1224 07:36:46.215926    5772 projected.go:292] Couldn't get configMap ix-pihole/kube-root-ca.crt: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:46 Server k3s[5772]: E1224 07:36:46.215953    5772 projected.go:198] Error preparing data for projected volume kube-api-access-8l46l for pod ix-pihole/pihole-ff558c9db-z9l5k: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:46 Server k3s[5772]: E1224 07:36:46.215987    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/70cfd1eb-640c-41b3-a322-48303d01ca4d-kube-api-access-8l46l podName:70cfd1eb-640c-41b3-a322-48303d01ca4d nodeName:}" failed. No retries permitted until 2023-12-24 07:36:48.215976336 +0100 CET m=+12.825931327 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-8l46l" (UniqueName: "kubernetes.io/projected/70cfd1eb-640c-41b3-a322-48303d01ca4d-kube-api-access-8l46l") pod "pihole-ff558c9db-z9l5k" (UID: "70cfd1eb-640c-41b3-a322-48303d01ca4d") : object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:46 Server k3s[5772]: I1224 07:36:46.218539    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d19f51ba-3f7e-45a0-ad35-f6abb0cf0b28-kube-api-access-tw289" (OuterVolumeSpecName: "kube-api-access-tw289") pod "d19f51ba-3f7e-45a0-ad35-f6abb0cf0b28" (UID: "d19f51ba-3f7e-45a0-ad35-f6abb0cf0b28"). InnerVolumeSpecName "kube-api-access-tw289". PluginName "kubernetes.io/projected", VolumeGidValue ""
Dec 24 07:36:46 Server k3s[5772]: E1224 07:36:46.273445    5772 projected.go:292] Couldn't get configMap ix-pihole/kube-root-ca.crt: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:46 Server k3s[5772]: E1224 07:36:46.273461    5772 projected.go:198] Error preparing data for projected volume kube-api-access-kdgpk for pod ix-pihole/pihole-ff558c9db-7wnhn: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:46 Server k3s[5772]: E1224 07:36:46.273495    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d12c8291-0283-4e24-b00a-251ce079dfcc-kube-api-access-kdgpk podName:d12c8291-0283-4e24-b00a-251ce079dfcc nodeName:}" failed. No retries permitted until 2023-12-24 07:36:46.773483247 +0100 CET m=+11.383438233 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-kdgpk" (UniqueName: "kubernetes.io/projected/d12c8291-0283-4e24-b00a-251ce079dfcc-kube-api-access-kdgpk") pod "pihole-ff558c9db-7wnhn" (UID: "d12c8291-0283-4e24-b00a-251ce079dfcc") : object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:46 Server k3s[5772]: I1224 07:36:46.316423    5772 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-tw289\" (UniqueName: \"kubernetes.io/projected/d19f51ba-3f7e-45a0-ad35-f6abb0cf0b28-kube-api-access-tw289\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:46 Server k3s[5772]: E1224 07:36:46.612742    5772 remote_runtime.go:205] "StopPodSandbox from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find sandbox \"4c513fd7fcec6ed9f9c4909205e29230d2724d75c3562c5fcf944727e1138559\": not found" podSandboxID="4c513fd7fcec6ed9f9c4909205e29230d2724d75c3562c5fcf944727e1138559"
Dec 24 07:36:46 Server k3s[5772]: I1224 07:36:46.819257    5772 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24wjw\" (UniqueName: \"kubernetes.io/projected/5e2592f9-ea55-4817-8872-c5d937a57ef6-kube-api-access-24wjw\") pod \"5e2592f9-ea55-4817-8872-c5d937a57ef6\" (UID: \"5e2592f9-ea55-4817-8872-c5d937a57ef6\") "
Dec 24 07:36:46 Server k3s[5772]: E1224 07:36:46.819570    5772 projected.go:292] Couldn't get configMap ix-pihole/kube-root-ca.crt: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:46 Server k3s[5772]: E1224 07:36:46.819599    5772 projected.go:198] Error preparing data for projected volume kube-api-access-kdgpk for pod ix-pihole/pihole-ff558c9db-7wnhn: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:46 Server k3s[5772]: E1224 07:36:46.819636    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d12c8291-0283-4e24-b00a-251ce079dfcc-kube-api-access-kdgpk podName:d12c8291-0283-4e24-b00a-251ce079dfcc nodeName:}" failed. No retries permitted until 2023-12-24 07:36:47.819625747 +0100 CET m=+12.429580738 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-kdgpk" (UniqueName: "kubernetes.io/projected/d12c8291-0283-4e24-b00a-251ce079dfcc-kube-api-access-kdgpk") pod "pihole-ff558c9db-7wnhn" (UID: "d12c8291-0283-4e24-b00a-251ce079dfcc") : object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:46 Server k3s[5772]: I1224 07:36:46.820761    5772 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2592f9-ea55-4817-8872-c5d937a57ef6-kube-api-access-24wjw" (OuterVolumeSpecName: "kube-api-access-24wjw") pod "5e2592f9-ea55-4817-8872-c5d937a57ef6" (UID: "5e2592f9-ea55-4817-8872-c5d937a57ef6"). InnerVolumeSpecName "kube-api-access-24wjw". PluginName "kubernetes.io/projected", VolumeGidValue ""
Dec 24 07:36:46 Server k3s[5772]: I1224 07:36:46.920313    5772 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-24wjw\" (UniqueName: \"kubernetes.io/projected/5e2592f9-ea55-4817-8872-c5d937a57ef6-kube-api-access-24wjw\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:47 Server k3s[5772]: E1224 07:36:47.020798    5772 configmap.go:199] Couldn't get configMap kube-system/openebs-zfspv-bin: object "kube-system"/"openebs-zfspv-bin" not registered
Dec 24 07:36:47 Server k3s[5772]: E1224 07:36:47.020799    5772 configmap.go:199] Couldn't get configMap kube-system/coredns: object "kube-system"/"coredns" not registered
Dec 24 07:36:47 Server k3s[5772]: E1224 07:36:47.020827    5772 configmap.go:199] Couldn't get configMap kube-system/coredns-custom: object "kube-system"/"coredns-custom" not registered
Dec 24 07:36:47 Server k3s[5772]: E1224 07:36:47.020960    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/013e3d9d-e200-4b53-b56a-e762a9decad5-chroot-zfs podName:013e3d9d-e200-4b53-b56a-e762a9decad5 nodeName:}" failed. No retries permitted until 2023-12-24 07:36:51.020948264 +0100 CET m=+15.630903256 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "chroot-zfs" (UniqueName: "kubernetes.io/configmap/013e3d9d-e200-4b53-b56a-e762a9decad5-chroot-zfs") pod "openebs-zfs-node-bx29q" (UID: "013e3d9d-e200-4b53-b56a-e762a9decad5") : object "kube-system"/"openebs-zfspv-bin" not registered
Dec 24 07:36:47 Server k3s[5772]: E1224 07:36:47.021036    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-config-volume podName:72100aad-f257-4df2-89aa-369c77e3439b nodeName:}" failed. No retries permitted until 2023-12-24 07:36:51.02102506 +0100 CET m=+15.630980047 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-config-volume") pod "coredns-59b4f5bbd5-tbbkx" (UID: "72100aad-f257-4df2-89aa-369c77e3439b") : object "kube-system"/"coredns" not registered
Dec 24 07:36:47 Server k3s[5772]: E1224 07:36:47.021047    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-custom-config-volume podName:72100aad-f257-4df2-89aa-369c77e3439b nodeName:}" failed. No retries permitted until 2023-12-24 07:36:51.021041538 +0100 CET m=+15.630996524 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "custom-config-volume" (UniqueName: "kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-custom-config-volume") pod "coredns-59b4f5bbd5-tbbkx" (UID: "72100aad-f257-4df2-89aa-369c77e3439b") : object "kube-system"/"coredns-custom" not registered
Dec 24 07:36:47 Server k3s[5772]: E1224 07:36:47.825547    5772 projected.go:292] Couldn't get configMap ix-pihole/kube-root-ca.crt: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:47 Server k3s[5772]: E1224 07:36:47.825569    5772 projected.go:198] Error preparing data for projected volume kube-api-access-kdgpk for pod ix-pihole/pihole-ff558c9db-7wnhn: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:47 Server k3s[5772]: E1224 07:36:47.825623    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d12c8291-0283-4e24-b00a-251ce079dfcc-kube-api-access-kdgpk podName:d12c8291-0283-4e24-b00a-251ce079dfcc nodeName:}" failed. No retries permitted until 2023-12-24 07:36:49.825611596 +0100 CET m=+14.435566586 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-kdgpk" (UniqueName: "kubernetes.io/projected/d12c8291-0283-4e24-b00a-251ce079dfcc-kube-api-access-kdgpk") pod "pihole-ff558c9db-7wnhn" (UID: "d12c8291-0283-4e24-b00a-251ce079dfcc") : object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:48 Server k3s[5772]: E1224 07:36:48.227074    5772 projected.go:292] Couldn't get configMap ix-pihole/kube-root-ca.crt: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:48 Server k3s[5772]: E1224 07:36:48.227202    5772 projected.go:198] Error preparing data for projected volume kube-api-access-8l46l for pod ix-pihole/pihole-ff558c9db-z9l5k: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:48 Server k3s[5772]: E1224 07:36:48.227315    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/70cfd1eb-640c-41b3-a322-48303d01ca4d-kube-api-access-8l46l podName:70cfd1eb-640c-41b3-a322-48303d01ca4d nodeName:}" failed. No retries permitted until 2023-12-24 07:36:52.227302555 +0100 CET m=+16.837257546 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-8l46l" (UniqueName: "kubernetes.io/projected/70cfd1eb-640c-41b3-a322-48303d01ca4d-kube-api-access-8l46l") pod "pihole-ff558c9db-z9l5k" (UID: "70cfd1eb-640c-41b3-a322-48303d01ca4d") : object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:49 Server k3s[5772]: E1224 07:36:49.836615    5772 projected.go:292] Couldn't get configMap ix-pihole/kube-root-ca.crt: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:49 Server k3s[5772]: E1224 07:36:49.836632    5772 projected.go:198] Error preparing data for projected volume kube-api-access-kdgpk for pod ix-pihole/pihole-ff558c9db-7wnhn: object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:49 Server k3s[5772]: E1224 07:36:49.836665    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d12c8291-0283-4e24-b00a-251ce079dfcc-kube-api-access-kdgpk podName:d12c8291-0283-4e24-b00a-251ce079dfcc nodeName:}" failed. No retries permitted until 2023-12-24 07:36:53.836654966 +0100 CET m=+18.446609959 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-kdgpk" (UniqueName: "kubernetes.io/projected/d12c8291-0283-4e24-b00a-251ce079dfcc-kube-api-access-kdgpk") pod "pihole-ff558c9db-7wnhn" (UID: "d12c8291-0283-4e24-b00a-251ce079dfcc") : object "ix-pihole"/"kube-root-ca.crt" not registered
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.846037    5772 range_allocator.go:109] No Secondary Service CIDR provided. Skipping filtering out secondary service addresses.
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.846080    5772 controllermanager.go:622] Started "nodeipam"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.846191    5772 node_ipam_controller.go:155] Starting ipam controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.846200    5772 shared_informer.go:270] Waiting for caches to sync for node
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.848086    5772 controllermanager.go:622] Started "ephemeral-volume"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.848889    5772 controller.go:169] Starting ephemeral volume controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.848913    5772 shared_informer.go:270] Waiting for caches to sync for ephemeral
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.850126    5772 controllermanager.go:622] Started "deployment"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.850437    5772 deployment_controller.go:154] "Starting controller" controller="deployment"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.850447    5772 shared_informer.go:270] Waiting for caches to sync for deployment
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.852066    5772 controllermanager.go:622] Started "persistentvolume-binder"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.852205    5772 pv_controller_base.go:318] Starting persistent volume controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.852214    5772 shared_informer.go:270] Waiting for caches to sync for persistent volume
Dec 24 07:36:50 Server k3s[5772]: W1224 07:36:50.852309    5772 core.go:221] configure-cloud-routes is set, but no cloud provider specified. Will not configure cloud provider routes.
Dec 24 07:36:50 Server k3s[5772]: W1224 07:36:50.852319    5772 controllermanager.go:600] Skipping "route"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.853947    5772 controllermanager.go:622] Started "attachdetach"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.854170    5772 attach_detach_controller.go:328] Starting attach detach controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.854196    5772 shared_informer.go:270] Waiting for caches to sync for attach detach
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.855748    5772 controllermanager.go:622] Started "clusterrole-aggregation"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.857312    5772 controllermanager.go:622] Started "pv-protection"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.857447    5772 pv_protection_controller.go:75] Starting PV protection controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.857458    5772 shared_informer.go:270] Waiting for caches to sync for PV protection
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.859127    5772 controllermanager.go:622] Started "daemonset"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.859405    5772 daemon_controller.go:267] Starting daemon sets controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.859416    5772 shared_informer.go:270] Waiting for caches to sync for daemon sets
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.862209    5772 controllermanager.go:622] Started "disruption"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.863878    5772 controllermanager.go:622] Started "ttl"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.864037    5772 ttl_controller.go:120] Starting TTL controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.864047    5772 shared_informer.go:270] Waiting for caches to sync for TTL
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.865682    5772 controllermanager.go:622] Started "root-ca-cert-publisher"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.867203    5772 controllermanager.go:622] Started "pvc-protection"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.867276    5772 pvc_protection_controller.go:99] "Starting PVC protection controller"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.867283    5772 shared_informer.go:270] Waiting for caches to sync for PVC protection
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.868970    5772 controllermanager.go:622] Started "endpoint"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.869210    5772 endpoints_controller.go:178] Starting endpoint controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.869219    5772 shared_informer.go:270] Waiting for caches to sync for endpoint
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.870709    5772 controllermanager.go:622] Started "replicationcontroller"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.870756    5772 replica_set.go:201] Starting replicationcontroller controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.870765    5772 shared_informer.go:270] Waiting for caches to sync for ReplicationController
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.889293    5772 controllermanager.go:622] Started "csrsigning"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.922666    5772 disruption.go:424] Sending events to api server.
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.922702    5772 disruption.go:435] Starting disruption controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.922708    5772 shared_informer.go:270] Waiting for caches to sync for disruption
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.922718    5772 clusterroleaggregation_controller.go:188] Starting ClusterRoleAggregator
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.922722    5772 shared_informer.go:270] Waiting for caches to sync for ClusterRoleAggregator
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.924053    5772 controllermanager.go:622] Started "csrcleaner"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984076    5772 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/mnt/Metzwerkspeicher/ix-applications/k3s/server/tls/server-ca.nochain.crt::/mnt/Metzwerkspeicher/ix-applications/k3s/server/tls/server-ca.key"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984150    5772 certificate_controller.go:112] Starting certificate controller "csrsigning-kube-apiserver-client"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984150    5772 certificate_controller.go:112] Starting certificate controller "csrsigning-kubelet-client"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984168    5772 publisher.go:101] Starting root CA certificate configmap publisher
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984174    5772 shared_informer.go:270] Waiting for caches to sync for certificate-csrsigning-kubelet-client
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984177    5772 shared_informer.go:270] Waiting for caches to sync for crt configmap
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984085    5772 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/mnt/Metzwerkspeicher/ix-applications/k3s/server/tls/client-ca.nochain.crt::/mnt/Metzwerkspeicher/ix-applications/k3s/server/tls/client-ca.key"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984094    5772 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/mnt/Metzwerkspeicher/ix-applications/k3s/server/tls/server-ca.nochain.crt::/mnt/Metzwerkspeicher/ix-applications/k3s/server/tls/server-ca.key"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984151    5772 certificate_controller.go:112] Starting certificate controller "csrsigning-kubelet-serving"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984246    5772 shared_informer.go:270] Waiting for caches to sync for certificate-csrsigning-kubelet-serving
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984165    5772 shared_informer.go:270] Waiting for caches to sync for certificate-csrsigning-kube-apiserver-client
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984164    5772 certificate_controller.go:112] Starting certificate controller "csrsigning-legacy-unknown"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984336    5772 shared_informer.go:270] Waiting for caches to sync for certificate-csrsigning-legacy-unknown
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984079    5772 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/mnt/Metzwerkspeicher/ix-applications/k3s/server/tls/client-ca.nochain.crt::/mnt/Metzwerkspeicher/ix-applications/k3s/server/tls/client-ca.key"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.984388    5772 cleaner.go:82] Starting CSR cleaner controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.985714    5772 controllermanager.go:622] Started "tokencleaner"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.985841    5772 tokencleaner.go:111] Starting token cleaner controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.985852    5772 shared_informer.go:270] Waiting for caches to sync for token_cleaner
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.985863    5772 shared_informer.go:277] Caches are synced for token_cleaner
Dec 24 07:36:50 Server k3s[5772]: E1224 07:36:50.987324    5772 core.go:92] Failed to start service controller: WARNING: no cloud provider provided, services of type LoadBalancer will fail
Dec 24 07:36:50 Server k3s[5772]: W1224 07:36:50.987352    5772 controllermanager.go:600] Skipping "service"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.988870    5772 controllermanager.go:622] Started "persistentvolume-expander"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.988969    5772 expand_controller.go:340] Starting expand controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.988979    5772 shared_informer.go:270] Waiting for caches to sync for expand
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.999310    5772 controllermanager.go:622] Started "namespace"
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.999445    5772 namespace_controller.go:195] Starting namespace controller
Dec 24 07:36:50 Server k3s[5772]: I1224 07:36:50.999454    5772 shared_informer.go:270] Waiting for caches to sync for namespace
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.001310    5772 controllermanager.go:622] Started "job"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.001461    5772 job_controller.go:191] Starting job controller
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.001472    5772 shared_informer.go:270] Waiting for caches to sync for job
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014020    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for cronjobs.batch
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014558    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for zfsvolumes.zfs.openebs.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014583    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for daemonsets.apps
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014600    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for roles.rbac.authorization.k8s.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014615    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for csistoragecapacities.storage.k8s.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014630    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for network-attachment-definitions.k8s.cni.cncf.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014643    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for zfsnodes.zfs.openebs.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014656    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for zfsbackups.zfs.openebs.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014673    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for controllerrevisions.apps
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014688    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for rolebindings.rbac.authorization.k8s.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014700    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for helmchartconfigs.helm.cattle.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014859    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for ingresses.networking.k8s.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014876    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for leases.coordination.k8s.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014899    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for zfsrestores.zfs.openebs.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.014912    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for zfssnapshots.zfs.openebs.io
Dec 24 07:36:51 Server k3s[5772]: W1224 07:36:51.015045    5772 shared_informer.go:548] resyncPeriod 12h14m32.519844705s is smaller than resyncCheckPeriod 14h40m33.458129017s and the informer has already started. Changing it to 14h40m33.458129017s
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015076    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for limitranges
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015129    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for horizontalpodautoscalers.autoscaling
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015142    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for endpointslices.discovery.k8s.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015154    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for endpoints
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015189    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for serviceaccounts
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015204    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for podtemplates
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015232    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for deployments.apps
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015244    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for statefulsets.apps
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015513    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for jobs.batch
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015530    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for addons.k3s.cattle.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015545    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for volumesnapshots.snapshot.storage.k8s.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015557    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for poddisruptionbudgets.policy
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015570    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for helmcharts.helm.cattle.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015592    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for replicasets.apps
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015605    5772 resource_quota_monitor.go:218] QuotaMonitor created object count evaluator for networkpolicies.networking.k8s.io
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015617    5772 controllermanager.go:622] Started "resourcequota"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015968    5772 resource_quota_controller.go:277] Starting resource quota controller
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.015995    5772 shared_informer.go:270] Waiting for caches to sync for resource quota
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.016018    5772 resource_quota_monitor.go:295] QuotaMonitor running
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.023211    5772 controllermanager.go:622] Started "horizontalpodautoscaling"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.023274    5772 horizontal.go:181] Starting HPA controller
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.023341    5772 shared_informer.go:270] Waiting for caches to sync for HPA
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.040574    5772 controllermanager.go:622] Started "cronjob"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.040668    5772 cronjob_controllerv2.go:137] "Starting cronjob controller v2"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.040687    5772 shared_informer.go:270] Waiting for caches to sync for cronjob
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.084904    5772 configmap.go:199] Couldn't get configMap kube-system/coredns-custom: object "kube-system"/"coredns-custom" not registered
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.084952    5772 configmap.go:199] Couldn't get configMap kube-system/openebs-zfspv-bin: object "kube-system"/"openebs-zfspv-bin" not registered
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.084963    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-custom-config-volume podName:72100aad-f257-4df2-89aa-369c77e3439b nodeName:}" failed. No retries permitted until 2023-12-24 07:36:59.084952481 +0100 CET m=+23.694907473 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "custom-config-volume" (UniqueName: "kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-custom-config-volume") pod "coredns-59b4f5bbd5-tbbkx" (UID: "72100aad-f257-4df2-89aa-369c77e3439b") : object "kube-system"/"coredns-custom" not registered
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.084906    5772 configmap.go:199] Couldn't get configMap kube-system/coredns: object "kube-system"/"coredns" not registered
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.084983    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/013e3d9d-e200-4b53-b56a-e762a9decad5-chroot-zfs podName:013e3d9d-e200-4b53-b56a-e762a9decad5 nodeName:}" failed. No retries permitted until 2023-12-24 07:36:59.084973398 +0100 CET m=+23.694928385 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "chroot-zfs" (UniqueName: "kubernetes.io/configmap/013e3d9d-e200-4b53-b56a-e762a9decad5-chroot-zfs") pod "openebs-zfs-node-bx29q" (UID: "013e3d9d-e200-4b53-b56a-e762a9decad5") : object "kube-system"/"openebs-zfspv-bin" not registered
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.084995    5772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-config-volume podName:72100aad-f257-4df2-89aa-369c77e3439b nodeName:}" failed. No retries permitted until 2023-12-24 07:36:59.084989514 +0100 CET m=+23.694944502 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-config-volume") pod "coredns-59b4f5bbd5-tbbkx" (UID: "72100aad-f257-4df2-89aa-369c77e3439b") : object "kube-system"/"coredns" not registered
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.090924    5772 controllermanager.go:622] Started "csrapproving"
Dec 24 07:36:51 Server k3s[5772]: W1224 07:36:51.090935    5772 controllermanager.go:587] "bootstrapsigner" is disabled
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.091005    5772 certificate_controller.go:112] Starting certificate controller "csrapproving"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.091016    5772 shared_informer.go:270] Waiting for caches to sync for certificate-csrapproving
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.095303    5772 shared_informer.go:270] Waiting for caches to sync for resource quota
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.115813    5772 shared_informer.go:270] Waiting for caches to sync for garbage collector
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.123497    5772 shared_informer.go:277] Caches are synced for ClusterRoleAggregator
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.123505    5772 shared_informer.go:277] Caches are synced for disruption
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.133977    5772 shared_informer.go:277] Caches are synced for endpoint_slice_mirroring
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.137292    5772 shared_informer.go:277] Caches are synced for ReplicaSet
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.149496    5772 shared_informer.go:277] Caches are synced for ephemeral
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.150619    5772 shared_informer.go:277] Caches are synced for deployment
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.157861    5772 shared_informer.go:277] Caches are synced for PV protection
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.167317    5772 shared_informer.go:277] Caches are synced for PVC protection
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.170160    5772 shared_informer.go:277] Caches are synced for endpoint
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.171407    5772 shared_informer.go:277] Caches are synced for ReplicationController
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.184596    5772 shared_informer.go:277] Caches are synced for certificate-csrsigning-legacy-unknown
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.184615    5772 shared_informer.go:277] Caches are synced for certificate-csrsigning-kube-apiserver-client
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.184658    5772 shared_informer.go:277] Caches are synced for certificate-csrsigning-kubelet-client
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.185121    5772 shared_informer.go:277] Caches are synced for certificate-csrsigning-kubelet-serving
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.185141    5772 shared_informer.go:277] Caches are synced for crt configmap
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.189429    5772 shared_informer.go:277] Caches are synced for expand
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.192028    5772 shared_informer.go:277] Caches are synced for certificate-csrapproving
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.198926    5772 shared_informer.go:277] Caches are synced for stateful set
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.200168    5772 event.go:294] "Event occurred" object="kube-system/snapshot-controller-546868dfb4" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: snapshot-controller-546868dfb4-9ss2x"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.201923    5772 shared_informer.go:277] Caches are synced for job
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.203105    5772 event.go:294] "Event occurred" object="kube-system/csi-smb-controller-7fbbb8fb6f" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: csi-smb-controller-7fbbb8fb6f-qpzrp"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.203333    5772 event.go:294] "Event occurred" object="kube-system/openebs-zfs-controller" fieldPath="" kind="StatefulSet" apiVersion="apps/v1" type="Warning" reason="RecreatingFailedPod" message="StatefulSet kube-system/openebs-zfs-controller is recreating failed Pod openebs-zfs-controller-0"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.204850    5772 event.go:294] "Event occurred" object="kube-system/csi-nfs-controller-7b74694749" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: csi-nfs-controller-7b74694749-47g2k"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.205789    5772 event.go:294] "Event occurred" object="kube-system/coredns-59b4f5bbd5" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: coredns-59b4f5bbd5-rnshv"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.209320    5772 event.go:294] "Event occurred" object="kube-system/snapshot-controller-546868dfb4" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: snapshot-controller-546868dfb4-qtbkj"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.224964    5772 shared_informer.go:277] Caches are synced for HPA
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.225510    5772 shared_informer.go:277] Caches are synced for TTL after finished
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.241097    5772 shared_informer.go:277] Caches are synced for cronjob
Dec 24 07:36:51 Server k3s[5772]: W1224 07:36:51.295613    5772 actual_state_of_world.go:541] Failed to update statusUpdateNeeded field in actual state of world: Failed to set statusUpdateNeeded to needed true, because nodeName="ix-truenas" does not exist
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.295835    5772 shared_informer.go:277] Caches are synced for resource quota
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.299509    5772 shared_informer.go:277] Caches are synced for namespace
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.311227    5772 shared_informer.go:277] Caches are synced for taint
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.311553    5772 node_lifecycle_controller.go:1438] Initializing eviction metric for zone:
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.311686    5772 taint_manager.go:206] "Starting NoExecuteTaintManager"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.311726    5772 taint_manager.go:211] "Sending events to api server"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.311852    5772 event.go:294] "Event occurred" object="ix-truenas" fieldPath="" kind="Node" apiVersion="v1" type="Normal" reason="RegisteredNode" message="Node ix-truenas event: Registered Node ix-truenas in Controller"
Dec 24 07:36:51 Server k3s[5772]: W1224 07:36:51.311893    5772 node_lifecycle_controller.go:1053] Missing timestamp for Node ix-truenas. Assuming now as a timestamp.
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.311926    5772 node_lifecycle_controller.go:1254] Controller detected that zone  is now in state Normal.
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.316235    5772 shared_informer.go:277] Caches are synced for resource quota
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.322773    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323049    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b8bec248-e32a-4e2c-b01d-a10a132f32e9" containerName="intel-gpu-plugin"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323060    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ec0104ce-ec27-4b74-87d8-78cec5066e59" containerName="csi-snapshotter"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323067    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ec0104ce-ec27-4b74-87d8-78cec5066e59" containerName="openebs-zfs-plugin"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323074    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="281b688e-7c75-47b8-aeb8-234d997059af" containerName="nfs"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323080    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="0cadcc5e-31e6-4b41-ab31-6f13326debe1" containerName="smb"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323087    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d12c8291-0283-4e24-b00a-251ce079dfcc" containerName="pihole"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323093    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="0cadcc5e-31e6-4b41-ab31-6f13326debe1" containerName="csi-provisioner"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323100    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ec0104ce-ec27-4b74-87d8-78cec5066e59" containerName="csi-resizer"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323106    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="013e3d9d-e200-4b53-b56a-e762a9decad5" containerName="csi-node-driver-registrar"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323113    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="281b688e-7c75-47b8-aeb8-234d997059af" containerName="csi-provisioner"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323119    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="f7daa893-7b06-453f-9300-2a183e5f6d33" containerName="snapshot-controller"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323125    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="72100aad-f257-4df2-89aa-369c77e3439b" containerName="coredns"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323131    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b8bec248-e32a-4e2c-b01d-a10a132f32e9" containerName="intel-gpu-initcontainer"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323138    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="281b688e-7c75-47b8-aeb8-234d997059af" containerName="csi-snapshotter"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323143    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="0cadcc5e-31e6-4b41-ab31-6f13326debe1" containerName="liveness-probe"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323149    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b65ae8ba-644a-4b60-b64c-eb90ef7c09ad" containerName="snapshot-controller"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323155    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="281b688e-7c75-47b8-aeb8-234d997059af" containerName="liveness-probe"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323161    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="013e3d9d-e200-4b53-b56a-e762a9decad5" containerName="openebs-zfs-plugin"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323168    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ec0104ce-ec27-4b74-87d8-78cec5066e59" containerName="snapshot-controller"
Dec 24 07:36:51 Server k3s[5772]: E1224 07:36:51.323174    5772 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ec0104ce-ec27-4b74-87d8-78cec5066e59" containerName="csi-provisioner"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323542    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="ec0104ce-ec27-4b74-87d8-78cec5066e59" containerName="csi-snapshotter"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323552    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="013e3d9d-e200-4b53-b56a-e762a9decad5" containerName="openebs-zfs-plugin"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323559    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="b65ae8ba-644a-4b60-b64c-eb90ef7c09ad" containerName="snapshot-controller"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323565    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="0cadcc5e-31e6-4b41-ab31-6f13326debe1" containerName="smb"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323570    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="ec0104ce-ec27-4b74-87d8-78cec5066e59" containerName="csi-provisioner"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323576    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="b8bec248-e32a-4e2c-b01d-a10a132f32e9" containerName="intel-gpu-plugin"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323584    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="281b688e-7c75-47b8-aeb8-234d997059af" containerName="csi-provisioner"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323589    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="b8bec248-e32a-4e2c-b01d-a10a132f32e9" containerName="intel-gpu-initcontainer"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323595    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="013e3d9d-e200-4b53-b56a-e762a9decad5" containerName="csi-node-driver-registrar"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323602    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="0cadcc5e-31e6-4b41-ab31-6f13326debe1" containerName="liveness-probe"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323609    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="ec0104ce-ec27-4b74-87d8-78cec5066e59" containerName="snapshot-controller"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323616    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="281b688e-7c75-47b8-aeb8-234d997059af" containerName="liveness-probe"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323621    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="72100aad-f257-4df2-89aa-369c77e3439b" containerName="coredns"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323627    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="f7daa893-7b06-453f-9300-2a183e5f6d33" containerName="snapshot-controller"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323634    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="ec0104ce-ec27-4b74-87d8-78cec5066e59" containerName="csi-resizer"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323639    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="d12c8291-0283-4e24-b00a-251ce079dfcc" containerName="pihole"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323645    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="0cadcc5e-31e6-4b41-ab31-6f13326debe1" containerName="csi-provisioner"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323651    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="ec0104ce-ec27-4b74-87d8-78cec5066e59" containerName="openebs-zfs-plugin"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323657    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="281b688e-7c75-47b8-aeb8-234d997059af" containerName="csi-snapshotter"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.323662    5772 memory_manager.go:346] "RemoveStaleState removing state" podUID="281b688e-7c75-47b8-aeb8-234d997059af" containerName="nfs"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.327251    5772 shared_informer.go:277] Caches are synced for GC
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.327444    5772 gc_controller.go:207] "Garbage collecting pods" numPods=8
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.327555    5772 gc_controller.go:329] "PodGC is force deleting Pod" pod="kube-system/coredns-59b4f5bbd5-tbbkx"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.327688    5772 gc_controller.go:329] "PodGC is force deleting Pod" pod="kube-system/csi-nfs-controller-7b74694749-h72f9"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.328068    5772 gc_controller.go:329] "PodGC is force deleting Pod" pod="kube-system/snapshot-controller-546868dfb4-2z962"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.328075    5772 gc_controller.go:329] "PodGC is force deleting Pod" pod="kube-system/csi-smb-controller-7fbbb8fb6f-rcntv"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.328068    5772 gc_controller.go:329] "PodGC is force deleting Pod" pod="ix-pihole/pihole-ff558c9db-7wnhn"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.328081    5772 gc_controller.go:329] "PodGC is force deleting Pod" pod="kube-system/snapshot-controller-546868dfb4-7xxcd"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.328083    5772 gc_controller.go:329] "PodGC is force deleting Pod" pod="ix-pihole/pihole-ff558c9db-z9l5k"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.328084    5772 gc_controller.go:329] "PodGC is force deleting Pod" pod="kube-system/snapshot-controller-546868dfb4-vxcw2"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.329900    5772 shared_informer.go:277] Caches are synced for service account
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.334179    5772 shared_informer.go:277] Caches are synced for endpoint_slice
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.347240    5772 shared_informer.go:277] Caches are synced for node
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.347268    5772 range_allocator.go:167] Sending events to api server.
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.347284    5772 range_allocator.go:171] Starting range CIDR allocator
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.347288    5772 shared_informer.go:270] Waiting for caches to sync for cidrallocator
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.347295    5772 shared_informer.go:277] Caches are synced for cidrallocator
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.353031    5772 shared_informer.go:277] Caches are synced for persistent volume
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.354224    5772 shared_informer.go:277] Caches are synced for attach detach
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.359725    5772 shared_informer.go:277] Caches are synced for daemon sets
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.360868    5772 event.go:294] "Event occurred" object="kube-system/openebs-zfs-node" fieldPath="" kind="DaemonSet" apiVersion="apps/v1" type="Warning" reason="FailedDaemonPod" message="Found failed daemon pod kube-system/openebs-zfs-node-bx29q on node ix-truenas, will try to kill it"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.364371    5772 shared_informer.go:277] Caches are synced for TTL
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.386177    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/f5b1b779-e3e1-4236-9202-ecdef6f6c5cd-config\") pod \"pihole-ff558c9db-kv4rw\" (UID: \"f5b1b779-e3e1-4236-9202-ecdef6f6c5cd\") " pod="ix-pihole/pihole-ff558c9db-kv4rw"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.386231    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6h4p\" (UniqueName: \"kubernetes.io/projected/f5b1b779-e3e1-4236-9202-ecdef6f6c5cd-kube-api-access-g6h4p\") pod \"pihole-ff558c9db-kv4rw\" (UID: \"f5b1b779-e3e1-4236-9202-ecdef6f6c5cd\") " pod="ix-pihole/pihole-ff558c9db-kv4rw"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.386295    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dnsmasq\" (UniqueName: \"kubernetes.io/host-path/f5b1b779-e3e1-4236-9202-ecdef6f6c5cd-dnsmasq\") pod \"pihole-ff558c9db-kv4rw\" (UID: \"f5b1b779-e3e1-4236-9202-ecdef6f6c5cd\") " pod="ix-pihole/pihole-ff558c9db-kv4rw"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.394057    5772 controller.go:615] quota admission added evaluator for: endpoints
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.397688    5772 event.go:294] "Event occurred" object="kube-system/kube-dns" fieldPath="" kind="Endpoints" apiVersion="v1" type="Warning" reason="FailedToUpdateEndpoint" message="Failed to update endpoint kube-system/kube-dns: Operation cannot be fulfilled on endpoints \"kube-dns\": the object has been modified; please apply your changes to the latest version and try again"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.448894    5772 event.go:294] "Event occurred" object="kube-system/openebs-zfs-controller" fieldPath="" kind="StatefulSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulDelete" message="delete Pod openebs-zfs-controller-0 in StatefulSet openebs-zfs-controller successful"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.452781    5772 event.go:294] "Event occurred" object="kube-system/openebs-zfs-controller" fieldPath="" kind="StatefulSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="create Pod openebs-zfs-controller-0 in StatefulSet openebs-zfs-controller successful"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.456742    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.586736    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/0731db04-c2ff-4c70-9ca0-824dae4a395b-socket-dir\") pod \"openebs-zfs-controller-0\" (UID: \"0731db04-c2ff-4c70-9ca0-824dae4a395b\") " pod="kube-system/openebs-zfs-controller-0"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.586766    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jh82\" (UniqueName: \"kubernetes.io/projected/0731db04-c2ff-4c70-9ca0-824dae4a395b-kube-api-access-7jh82\") pod \"openebs-zfs-controller-0\" (UID: \"0731db04-c2ff-4c70-9ca0-824dae4a395b\") " pod="kube-system/openebs-zfs-controller-0"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.715938    5772 shared_informer.go:277] Caches are synced for garbage collector
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.735297    5772 shared_informer.go:277] Caches are synced for garbage collector
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.735313    5772 garbagecollector.go:163] Garbage collector: all resource monitors have synced. Proceeding to collect garbage
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.745572    5772 controller.go:615] quota admission added evaluator for: endpointslices.discovery.k8s.io
Dec 24 07:36:51 Server k3s[5772]: W1224 07:36:51.747040    5772 endpointslice_controller.go:302] Error syncing endpoint slices for service "kube-system/kube-dns", retrying. Error: EndpointSlice informer cache is out of date
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.788468    5772 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-8l46l\" (UniqueName: \"kubernetes.io/projected/70cfd1eb-640c-41b3-a322-48303d01ca4d-kube-api-access-8l46l\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.788493    5772 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-kdgpk\" (UniqueName: \"kubernetes.io/projected/d12c8291-0283-4e24-b00a-251ce079dfcc-kube-api-access-kdgpk\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.788502    5772 reconciler_common.go:295] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-config-volume\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.788535    5772 reconciler_common.go:295] "Volume detached for volume \"custom-config-volume\" (UniqueName: \"kubernetes.io/configmap/72100aad-f257-4df2-89aa-369c77e3439b-custom-config-volume\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.797355    5772 event.go:294] "Event occurred" object="kube-system/intel-gpu-plugin" fieldPath="" kind="DaemonSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: intel-gpu-plugin-jfgct"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.803247    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.807226    5772 event.go:294] "Event occurred" object="kube-system/openebs-zfs-node" fieldPath="" kind="DaemonSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulDelete" message="Deleted pod: openebs-zfs-node-bx29q"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.814924    5772 event.go:294] "Event occurred" object="kube-system/openebs-zfs-node" fieldPath="" kind="DaemonSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: openebs-zfs-node-5cw4c"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.820207    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989018    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1f566830-7526-4b6c-b1c4-286801b4dca6-device-dir\") pod \"openebs-zfs-node-5cw4c\" (UID: \"1f566830-7526-4b6c-b1c4-286801b4dca6\") " pod="kube-system/openebs-zfs-node-5cw4c"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989047    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"chroot-zfs\" (UniqueName: \"kubernetes.io/configmap/1f566830-7526-4b6c-b1c4-286801b4dca6-chroot-zfs\") pod \"openebs-zfs-node-5cw4c\" (UID: \"1f566830-7526-4b6c-b1c4-286801b4dca6\") " pod="kube-system/openebs-zfs-node-5cw4c"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989067    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pods-mount-dir\" (UniqueName: \"kubernetes.io/host-path/1f566830-7526-4b6c-b1c4-286801b4dca6-pods-mount-dir\") pod \"openebs-zfs-node-5cw4c\" (UID: \"1f566830-7526-4b6c-b1c4-286801b4dca6\") " pod="kube-system/openebs-zfs-node-5cw4c"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989092    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sysfs\" (UniqueName: \"kubernetes.io/host-path/ced713ed-e86e-4722-a029-3fc87251323d-sysfs\") pod \"intel-gpu-plugin-jfgct\" (UID: \"ced713ed-e86e-4722-a029-3fc87251323d\") " pod="kube-system/intel-gpu-plugin-jfgct"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989109    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-root\" (UniqueName: \"kubernetes.io/host-path/1f566830-7526-4b6c-b1c4-286801b4dca6-host-root\") pod \"openebs-zfs-node-5cw4c\" (UID: \"1f566830-7526-4b6c-b1c4-286801b4dca6\") " pod="kube-system/openebs-zfs-node-5cw4c"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989177    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1f566830-7526-4b6c-b1c4-286801b4dca6-plugin-dir\") pod \"openebs-zfs-node-5cw4c\" (UID: \"1f566830-7526-4b6c-b1c4-286801b4dca6\") " pod="kube-system/openebs-zfs-node-5cw4c"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989209    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"devfs\" (UniqueName: \"kubernetes.io/host-path/ced713ed-e86e-4722-a029-3fc87251323d-devfs\") pod \"intel-gpu-plugin-jfgct\" (UID: \"ced713ed-e86e-4722-a029-3fc87251323d\") " pod="kube-system/intel-gpu-plugin-jfgct"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989243    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1f566830-7526-4b6c-b1c4-286801b4dca6-registration-dir\") pod \"openebs-zfs-node-5cw4c\" (UID: \"1f566830-7526-4b6c-b1c4-286801b4dca6\") " pod="kube-system/openebs-zfs-node-5cw4c"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989304    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8lqt\" (UniqueName: \"kubernetes.io/projected/1f566830-7526-4b6c-b1c4-286801b4dca6-kube-api-access-c8lqt\") pod \"openebs-zfs-node-5cw4c\" (UID: \"1f566830-7526-4b6c-b1c4-286801b4dca6\") " pod="kube-system/openebs-zfs-node-5cw4c"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989332    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeletsockets\" (UniqueName: \"kubernetes.io/host-path/ced713ed-e86e-4722-a029-3fc87251323d-kubeletsockets\") pod \"intel-gpu-plugin-jfgct\" (UID: \"ced713ed-e86e-4722-a029-3fc87251323d\") " pod="kube-system/intel-gpu-plugin-jfgct"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989383    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kszm5\" (UniqueName: \"kubernetes.io/projected/ced713ed-e86e-4722-a029-3fc87251323d-kube-api-access-kszm5\") pod \"intel-gpu-plugin-jfgct\" (UID: \"ced713ed-e86e-4722-a029-3fc87251323d\") " pod="kube-system/intel-gpu-plugin-jfgct"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989414    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encr-keys\" (UniqueName: \"kubernetes.io/host-path/1f566830-7526-4b6c-b1c4-286801b4dca6-encr-keys\") pod \"openebs-zfs-node-5cw4c\" (UID: \"1f566830-7526-4b6c-b1c4-286801b4dca6\") " pod="kube-system/openebs-zfs-node-5cw4c"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989450    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nfd-source-hooks\" (UniqueName: \"kubernetes.io/host-path/ced713ed-e86e-4722-a029-3fc87251323d-nfd-source-hooks\") pod \"intel-gpu-plugin-jfgct\" (UID: \"ced713ed-e86e-4722-a029-3fc87251323d\") " pod="kube-system/intel-gpu-plugin-jfgct"
Dec 24 07:36:51 Server k3s[5772]: I1224 07:36:51.989471    5772 reconciler_common.go:295] "Volume detached for volume \"chroot-zfs\" (UniqueName: \"kubernetes.io/configmap/013e3d9d-e200-4b53-b56a-e762a9decad5-chroot-zfs\") on node \"ix-truenas\" DevicePath \"\""
Dec 24 07:36:52 Server k3s[5772]: I1224 07:36:52.417919    5772 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: nfs.csi.k8s.io endpoint: /var/lib/kubelet/plugins/csi-nfsplugin/csi.sock versions: 1.0.0
Dec 24 07:36:52 Server k3s[5772]: I1224 07:36:52.417948    5772 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: nfs.csi.k8s.io at endpoint: /var/lib/kubelet/plugins/csi-nfsplugin/csi.sock
Dec 24 07:36:52 Server k3s[5772]: I1224 07:36:52.422728    5772 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=013e3d9d-e200-4b53-b56a-e762a9decad5 path="/var/lib/kubelet/pods/013e3d9d-e200-4b53-b56a-e762a9decad5/volumes"
Dec 24 07:36:52 Server k3s[5772]: I1224 07:36:52.424112    5772 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=14f02546-7f11-4aa1-aeda-d8e9f10ecbf2 path="/var/lib/kubelet/pods/14f02546-7f11-4aa1-aeda-d8e9f10ecbf2/volumes"
Dec 24 07:36:52 Server k3s[5772]: I1224 07:36:52.425613    5772 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=5e2592f9-ea55-4817-8872-c5d937a57ef6 path="/var/lib/kubelet/pods/5e2592f9-ea55-4817-8872-c5d937a57ef6/volumes"
Dec 24 07:36:52 Server k3s[5772]: I1224 07:36:52.426682    5772 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=70cfd1eb-640c-41b3-a322-48303d01ca4d path="/var/lib/kubelet/pods/70cfd1eb-640c-41b3-a322-48303d01ca4d/volumes"
Dec 24 07:36:52 Server k3s[5772]: I1224 07:36:52.427943    5772 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=72100aad-f257-4df2-89aa-369c77e3439b path="/var/lib/kubelet/pods/72100aad-f257-4df2-89aa-369c77e3439b/volumes"
Dec 24 07:36:52 Server k3s[5772]: I1224 07:36:52.429094    5772 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=8794e834-b432-4b66-872d-5406ffc0ea25 path="/var/lib/kubelet/pods/8794e834-b432-4b66-872d-5406ffc0ea25/volumes"
Dec 24 07:36:52 Server k3s[5772]: I1224 07:36:52.461369    5772 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=b65ae8ba-644a-4b60-b64c-eb90ef7c09ad path="/var/lib/kubelet/pods/b65ae8ba-644a-4b60-b64c-eb90ef7c09ad/volumes"
Dec 24 07:36:52 Server k3s[5772]: I1224 07:36:52.462347    5772 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=d12c8291-0283-4e24-b00a-251ce079dfcc path="/var/lib/kubelet/pods/d12c8291-0283-4e24-b00a-251ce079dfcc/volumes"
Dec 24 07:36:52 Server k3s[5772]: I1224 07:36:52.463368    5772 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=d19f51ba-3f7e-45a0-ad35-f6abb0cf0b28 path="/var/lib/kubelet/pods/d19f51ba-3f7e-45a0-ad35-f6abb0cf0b28/volumes"
Dec 24 07:36:52 Server k3s[5772]: I1224 07:36:52.464486    5772 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=ec0104ce-ec27-4b74-87d8-78cec5066e59 path="/var/lib/kubelet/pods/ec0104ce-ec27-4b74-87d8-78cec5066e59/volumes"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.027378    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.028618    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.030115    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.032054    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.039748    5772 topology_manager.go:210] "Topology Admit Handler"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.197377    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztk5t\" (UniqueName: \"kubernetes.io/projected/ee05a938-532a-413c-b2df-8c900ee36ff3-kube-api-access-ztk5t\") pod \"coredns-59b4f5bbd5-rnshv\" (UID: \"ee05a938-532a-413c-b2df-8c900ee36ff3\") " pod="kube-system/coredns-59b4f5bbd5-rnshv"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.197473    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk6m5\" (UniqueName: \"kubernetes.io/projected/40fb22d1-ed17-475f-8c30-4280894f9388-kube-api-access-jk6m5\") pod \"snapshot-controller-546868dfb4-qtbkj\" (UID: \"40fb22d1-ed17-475f-8c30-4280894f9388\") " pod="kube-system/snapshot-controller-546868dfb4-qtbkj"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.197498    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dphwv\" (UniqueName: \"kubernetes.io/projected/9b81aa20-c725-47ff-9cf4-1dccdb65ddde-kube-api-access-dphwv\") pod \"csi-nfs-controller-7b74694749-47g2k\" (UID: \"9b81aa20-c725-47ff-9cf4-1dccdb65ddde\") " pod="kube-system/csi-nfs-controller-7b74694749-47g2k"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.197517    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/9b81aa20-c725-47ff-9cf4-1dccdb65ddde-socket-dir\") pod \"csi-nfs-controller-7b74694749-47g2k\" (UID: \"9b81aa20-c725-47ff-9cf4-1dccdb65ddde\") " pod="kube-system/csi-nfs-controller-7b74694749-47g2k"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.197552    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/46aa44d3-4607-4bfd-9570-abbba694ce15-socket-dir\") pod \"csi-smb-controller-7fbbb8fb6f-qpzrp\" (UID: \"46aa44d3-4607-4bfd-9570-abbba694ce15\") " pod="kube-system/csi-smb-controller-7fbbb8fb6f-qpzrp"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.197573    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcqfp\" (UniqueName: \"kubernetes.io/projected/46aa44d3-4607-4bfd-9570-abbba694ce15-kube-api-access-rcqfp\") pod \"csi-smb-controller-7fbbb8fb6f-qpzrp\" (UID: \"46aa44d3-4607-4bfd-9570-abbba694ce15\") " pod="kube-system/csi-smb-controller-7fbbb8fb6f-qpzrp"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.197606    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee05a938-532a-413c-b2df-8c900ee36ff3-config-volume\") pod \"coredns-59b4f5bbd5-rnshv\" (UID: \"ee05a938-532a-413c-b2df-8c900ee36ff3\") " pod="kube-system/coredns-59b4f5bbd5-rnshv"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.197635    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-config-volume\" (UniqueName: \"kubernetes.io/configmap/ee05a938-532a-413c-b2df-8c900ee36ff3-custom-config-volume\") pod \"coredns-59b4f5bbd5-rnshv\" (UID: \"ee05a938-532a-413c-b2df-8c900ee36ff3\") " pod="kube-system/coredns-59b4f5bbd5-rnshv"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.197682    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pods-mount-dir\" (UniqueName: \"kubernetes.io/host-path/9b81aa20-c725-47ff-9cf4-1dccdb65ddde-pods-mount-dir\") pod \"csi-nfs-controller-7b74694749-47g2k\" (UID: \"9b81aa20-c725-47ff-9cf4-1dccdb65ddde\") " pod="kube-system/csi-nfs-controller-7b74694749-47g2k"
Dec 24 07:36:53 Server k3s[5772]: I1224 07:36:53.197727    5772 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh2kl\" (UniqueName: \"kubernetes.io/projected/50ae894d-c083-4969-8f3c-4979a824f081-kube-api-access-xh2kl\") pod \"snapshot-controller-546868dfb4-9ss2x\" (UID: \"50ae894d-c083-4969-8f3c-4979a824f081\") " pod="kube-system/snapshot-controller-546868dfb4-9ss2x"
Dec 24 07:36:54 Server k3s[5772]: I1224 07:36:54.416627    5772 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: smb.csi.k8s.io endpoint: /var/lib/kubelet/plugins/smb.csi.k8s.io/csi.sock versions: 1.0.0
Dec 24 07:36:54 Server k3s[5772]: I1224 07:36:54.416645    5772 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: smb.csi.k8s.io at endpoint: /var/lib/kubelet/plugins/smb.csi.k8s.io/csi.sock
Dec 24 07:36:54 Server k3s[5772]: I1224 07:36:54.869085    5772 request.go:690] Waited for 1.570012353s due to client-side throttling, not priority and fairness, request: POST:https://127.0.0.1:6443/api/v1/namespaces/kube-system/serviceaccounts/csi-nfs-controller-sa/token
Dec 24 07:36:58 Server k3s[5772]: I1224 07:36:58.273574    5772 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="ix-pihole/pihole-ff558c9db-kv4rw" podStartSLOduration=24149.273548962 pod.CreationTimestamp="2023-12-24 00:54:29 +0100 CET" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2023-12-24 07:36:58.273435485 +0100 CET m=+22.883390476" watchObservedRunningTime="2023-12-24 07:36:58.273548962 +0100 CET m=+22.883503948"
Dec 24 07:37:02 Server k3s[5772]: I1224 07:37:02.620994    5772 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/openebs-zfs-node-5cw4c" podStartSLOduration=11.620969553 pod.CreationTimestamp="2023-12-24 07:36:51 +0100 CET" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2023-12-24 07:37:02.620629185 +0100 CET m=+27.230584183" watchObservedRunningTime="2023-12-24 07:37:02.620969553 +0100 CET m=+27.230924538"
Dec 24 07:37:05 Server k3s[5772]: I1224 07:37:05.632722    5772 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-59b4f5bbd5-rnshv" podStartSLOduration=14.632696333 pod.CreationTimestamp="2023-12-24 07:36:51 +0100 CET" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2023-12-24 07:37:05.632286402 +0100 CET m=+30.242241390" watchObservedRunningTime="2023-12-24 07:37:05.632696333 +0100 CET m=+30.242651321"
Dec 24 07:37:05 Server k3s[5772]: I1224 07:37:05.641907    5772 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/snapshot-controller-546868dfb4-qtbkj" podStartSLOduration=14.641881087 pod.CreationTimestamp="2023-12-24 07:36:51 +0100 CET" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2023-12-24 07:37:05.641813711 +0100 CET m=+30.251768704" watchObservedRunningTime="2023-12-24 07:37:05.641881087 +0100 CET m=+30.251836079"
Dec 24 07:37:06 Server k3s[5772]: I1224 07:37:06.638429    5772 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/snapshot-controller-546868dfb4-9ss2x" podStartSLOduration=15.638403001 pod.CreationTimestamp="2023-12-24 07:36:51 +0100 CET" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2023-12-24 07:37:06.637916942 +0100 CET m=+31.247871933" watchObservedRunningTime="2023-12-24 07:37:06.638403001 +0100 CET m=+31.248357987"
Dec 24 07:37:08 Server k3s[5772]: I1224 07:37:08.425530    5772 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: zfs.csi.openebs.io endpoint: /var/lib/kubelet/plugins/zfs-localpv/csi.sock versions: 1.0.0
Dec 24 07:37:08 Server k3s[5772]: I1224 07:37:08.425549    5772 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: zfs.csi.openebs.io at endpoint: /var/lib/kubelet/plugins/zfs-localpv/csi.sock
Dec 24 07:37:10 Server k3s[5772]: I1224 07:37:10.027429    5772 controller.go:615] quota admission added evaluator for: zfsnodes.zfs.openebs.io
Dec 24 07:37:11 Server k3s[5772]: I1224 07:37:11.865395    5772 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/intel-gpu-plugin-jfgct" podStartSLOduration=20.865367936 pod.CreationTimestamp="2023-12-24 07:36:51 +0100 CET" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2023-12-24 07:37:09.651539102 +0100 CET m=+34.261494092" watchObservedRunningTime="2023-12-24 07:37:11.865367936 +0100 CET m=+36.475322929"
Dec 24 07:37:11 Server k3s[5772]: I1224 07:37:11.866951    5772 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/openebs-zfs-controller-0" podStartSLOduration=20.866927965 pod.CreationTimestamp="2023-12-24 07:36:51 +0100 CET" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2023-12-24 07:37:11.865273704 +0100 CET m=+36.475228690" watchObservedRunningTime="2023-12-24 07:37:11.866927965 +0100 CET m=+36.476882952"
Dec 24 07:37:13 Server k3s[5772]: I1224 07:37:13.292395    5772 server.go:144] "Got registration request from device plugin with resource" resourceName="gpu.intel.com/i915"
Dec 24 07:37:13 Server k3s[5772]: I1224 07:37:13.873159    5772 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/csi-smb-controller-7fbbb8fb6f-qpzrp" podStartSLOduration=22.873134252 pod.CreationTimestamp="2023-12-24 07:36:51 +0100 CET" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2023-12-24 07:37:13.872122678 +0100 CET m=+38.482077670" watchObservedRunningTime="2023-12-24 07:37:13.873134252 +0100 CET m=+38.483089238"
Dec 24 07:37:14 Server k3s[5772]: I1224 07:37:14.875649    5772 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/csi-nfs-controller-7b74694749-47g2k" podStartSLOduration=23.875609157 pod.CreationTimestamp="2023-12-24 07:36:51 +0100 CET" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2023-12-24 07:37:14.875466586 +0100 CET m=+39.485421586" watchObservedRunningTime="2023-12-24 07:37:14.875609157 +0100 CET m=+39.485564143"
Dec 24 07:37:21 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:37:21.324+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:37:42 Server k3s[5772]: I1224 07:37:42.706400    5772 scope.go:115] "RemoveContainer" containerID="b06467bc5f3df2c5a2b3b2852667516183f544b47c9c129cf8d1e9befaf6e571"
Dec 24 07:37:42 Server k3s[5772]: I1224 07:37:42.772669    5772 scope.go:115] "RemoveContainer" containerID="0248bc1587a637558e6c06d9fc542ab8d0ca0e8c2dafc2963cf25c5def0e49fa"
Dec 24 07:37:42 Server k3s[5772]: I1224 07:37:42.821155    5772 scope.go:115] "RemoveContainer" containerID="f626df0ae306d415783964f0d24b6f35f477236c06d90c000f5bdbc3447820ca"
Dec 24 07:37:42 Server k3s[5772]: I1224 07:37:42.868492    5772 scope.go:115] "RemoveContainer" containerID="3180324c6891d40a398802e5e3c14b13814bda38e0db7e2e4eb9ea8bc25da9f7"
Dec 24 07:37:42 Server k3s[5772]: I1224 07:37:42.973124    5772 scope.go:115] "RemoveContainer" containerID="54083da4fb8935ccfa4669652041e978f2c62793a1ec18a659b0599cc73e7dbd"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.062670    5772 scope.go:115] "RemoveContainer" containerID="b0a2cd8e2d727bc3ca542cfff748984ba16d1fef113e8dd6846834b6114c0cfd"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.143403    5772 scope.go:115] "RemoveContainer" containerID="f9cff543038b1fb044e782db0df1ac27993367390d30b8ee6dc6833e0b7b210a"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.188806    5772 scope.go:115] "RemoveContainer" containerID="01513eee160b623bf279fe6ea918e490b24b658f7208e73ce13df97e52caa50e"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.233911    5772 scope.go:115] "RemoveContainer" containerID="5db8139cbdf159d79aa8dbf231ccb71d4176eace9505271ecb9630550040d553"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.279131    5772 scope.go:115] "RemoveContainer" containerID="099d1a3147f97bde074b39344b836ad27d039ba8c29eea8c72a495c667a8319c"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.325532    5772 scope.go:115] "RemoveContainer" containerID="9abc733695013dd26170a49e08f350df70df55bd0bf3a2e9502a8cfeb73339cb"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.371016    5772 scope.go:115] "RemoveContainer" containerID="3fac6753fa3b6b072a865129174d6da3afb139ad829d73886dad7c98b599802e"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.439905    5772 scope.go:115] "RemoveContainer" containerID="d67b9e202dd96fbc24cd40fb0f14b22a9c450faab02c692179059437e998c743"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.485734    5772 scope.go:115] "RemoveContainer" containerID="e21327b5a08297fcbe194bd5e877e0ca943bda5b4a33b7894a519e9c88f62fff"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.542235    5772 scope.go:115] "RemoveContainer" containerID="2a2d23ba6920bb03008c9c633b9f6854f308f2d6ce76f224de161ca50fee9e88"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.654173    5772 scope.go:115] "RemoveContainer" containerID="1d566032ab8f5caefaaacd2bfe27288021cf9e64d454ff32e96ffe832607ffd3"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.699417    5772 scope.go:115] "RemoveContainer" containerID="7a72b00aaa4a7125e8daaa7693045ac91b4bdfdc77edde0bbd418a44b5fbdcd3"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.797251    5772 scope.go:115] "RemoveContainer" containerID="702d5aa102c7873262a1373057ed1cfd6e2a755333d40d0797a9f3cf53844262"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.845784    5772 scope.go:115] "RemoveContainer" containerID="f391ade3c76db661a2527fead2bb1096e482c6e9db34f18dfc6d677466ab4518"
Dec 24 07:37:43 Server k3s[5772]: I1224 07:37:43.890874    5772 scope.go:115] "RemoveContainer" containerID="2dc499c78a5cad93acb3601c4ba4bd61fdf0c502a90f1df323fe47b2bbf65875"
Dec 24 07:37:54 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:37:54.888+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:38:32 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:38:32.076+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:39:15 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:39:15.787+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:39:54 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:39:54.178+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:40:29 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:40:29.514+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:41:11 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:41:11.191+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:41:36 Server k3s[5772]: time="2023-12-24T07:41:36+01:00" level=info msg="COMPACT compactRev=16158 targetCompactRev=17158 currentRev=18256"
Dec 24 07:41:36 Server k3s[5772]: time="2023-12-24T07:41:36+01:00" level=info msg="COMPACT deleted 982 rows from 1000 revisions in 46.247086ms - compacted to 17158/18256"
Dec 24 07:41:36 Server k3s[5772]: time="2023-12-24T07:41:36+01:00" level=info msg="COMPACT compactRev=17158 targetCompactRev=17256 currentRev=18256"
Dec 24 07:41:36 Server k3s[5772]: time="2023-12-24T07:41:36+01:00" level=info msg="COMPACT deleted 98 rows from 98 revisions in 12.030851ms - compacted to 17256/18256"
Dec 24 07:41:42 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:41:42.258+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:42:14 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:42:14.274+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:42:49 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:42:49.584+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:43:22 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:43:22.843+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:44:03 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:44:03.069+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:44:33 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:44:33.846+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:45:14 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:45:14.709+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:45:50 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:45:50.083+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:46:30 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:46:30.705+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:46:36 Server k3s[5772]: time="2023-12-24T07:46:36+01:00" level=info msg="COMPACT compactRev=17256 targetCompactRev=17771 currentRev=18771"
Dec 24 07:46:36 Server k3s[5772]: time="2023-12-24T07:46:36+01:00" level=info msg="COMPACT deleted 337 rows from 515 revisions in 12.005409ms - compacted to 17771/18771"
Dec 24 07:47:10 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:47:10.279+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:47:48 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:47:48.570+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:48:19 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:48:19.433+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:48:58 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:48:58.698+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:49:33 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:49:33.005+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:50:06 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:50:06.619+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:50:47 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:50:47.633+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:51:31 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:51:31.033+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:51:36 Server k3s[5772]: time="2023-12-24T07:51:36+01:00" level=info msg="COMPACT compactRev=17771 targetCompactRev=18286 currentRev=19286"
Dec 24 07:51:36 Server k3s[5772]: time="2023-12-24T07:51:36+01:00" level=info msg="COMPACT deleted 497 rows from 515 revisions in 6.923349ms - compacted to 18286/19286"
Dec 24 07:52:13 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:52:13.902+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:52:54 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:52:54.690+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:53:29 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:53:29.782+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:54:13 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:54:13.759+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:54:45 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:54:45.734+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:55:17 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:55:17.211+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:55:49 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:55:49.332+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:56:22 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:56:22.549+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:56:36 Server k3s[5772]: time="2023-12-24T07:56:36+01:00" level=info msg="COMPACT compactRev=18286 targetCompactRev=18800 currentRev=19800"
Dec 24 07:56:36 Server k3s[5772]: time="2023-12-24T07:56:36+01:00" level=info msg="COMPACT deleted 514 rows from 514 revisions in 6.393778ms - compacted to 18800/19800"
Dec 24 07:56:55 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:56:55.657+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:57:30 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:57:30.821+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:58:12 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:58:12.940+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:58:48 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:58:48.939+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 07:59:26 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T07:59:26.336+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:00:09 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:00:09.157+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:00:41 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:00:41.862+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:01:14 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:01:14.408+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:01:36 Server k3s[5772]: time="2023-12-24T08:01:36+01:00" level=info msg="COMPACT compactRev=18800 targetCompactRev=19314 currentRev=20314"
Dec 24 08:01:36 Server k3s[5772]: time="2023-12-24T08:01:36+01:00" level=info msg="COMPACT deleted 514 rows from 514 revisions in 4.951969ms - compacted to 19314/20314"
Dec 24 08:01:55 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:01:55.233+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:02:29 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:02:29.468+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:03:02 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:03:02.524+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:03:43 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:03:43.590+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:04:17 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:04:17.194+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:04:49 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:04:49.756+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:05:24 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:05:24.284+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:05:59 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:05:59.324+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:06:36 Server k3s[5772]: time="2023-12-24T08:06:36+01:00" level=info msg="COMPACT compactRev=19314 targetCompactRev=19828 currentRev=20828"
Dec 24 08:06:36 Server k3s[5772]: time="2023-12-24T08:06:36+01:00" level=info msg="COMPACT deleted 514 rows from 514 revisions in 4.032555ms - compacted to 19828/20828"
Dec 24 08:06:36 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:06:36.740+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:07:07 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:07:07.374+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:07:42 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:07:42.568+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:08:20 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:08:20.132+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:08:55 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:08:55.550+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:09:31 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:09:31.762+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:10:06 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:10:06.955+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:10:47 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:10:47.569+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:11:27 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:11:27.762+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:11:36 Server k3s[5772]: time="2023-12-24T08:11:36+01:00" level=info msg="COMPACT compactRev=19828 targetCompactRev=20343 currentRev=21343"
Dec 24 08:11:36 Server k3s[5772]: time="2023-12-24T08:11:36+01:00" level=info msg="COMPACT deleted 515 rows from 515 revisions in 5.495936ms - compacted to 20343/21343"
Dec 24 08:12:09 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:12:09.903+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:12:45 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:12:45.099+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:13:20 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:13:20.293+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:13:53 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:13:53.569+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:14:37 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:14:37.382+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:15:15 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:15:15.629+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:15:49 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:15:49.706+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:16:33 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:16:33.341+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:16:36 Server k3s[5772]: time="2023-12-24T08:16:36+01:00" level=info msg="COMPACT compactRev=20343 targetCompactRev=20858 currentRev=21858"
Dec 24 08:16:36 Server k3s[5772]: time="2023-12-24T08:16:36+01:00" level=info msg="COMPACT deleted 515 rows from 515 revisions in 5.371431ms - compacted to 20858/21858"
Dec 24 08:17:07 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:17:07.899+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:17:48 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:17:48.987+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:18:20 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:18:20.326+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:18:57 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:18:57.761+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:19:38 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:19:38.427+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:20:11 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:20:11.995+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:20:43 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:20:43.438+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:21:27 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:21:27.904+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:21:36 Server k3s[5772]: time="2023-12-24T08:21:36+01:00" level=info msg="COMPACT compactRev=20858 targetCompactRev=21371 currentRev=22371"
Dec 24 08:21:36 Server k3s[5772]: time="2023-12-24T08:21:36+01:00" level=info msg="COMPACT deleted 513 rows from 513 revisions in 8.273271ms - compacted to 21371/22371"
Dec 24 08:22:05 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:22:05.815+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:22:50 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:22:50.740+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:23:24 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:23:24.311+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:24:01 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:24:01.878+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:24:41 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:24:41.731+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:25:22 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:25:22.778+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:25:52 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:25:52.920+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:26:34 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:26:34.399+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:26:36 Server k3s[5772]: time="2023-12-24T08:26:36+01:00" level=info msg="COMPACT compactRev=21371 targetCompactRev=21885 currentRev=22885"
Dec 24 08:26:36 Server k3s[5772]: time="2023-12-24T08:26:36+01:00" level=info msg="COMPACT deleted 514 rows from 514 revisions in 10.697769ms - compacted to 21885/22885"
Dec 24 08:27:18 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:27:18.504+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:27:50 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:27:50.122+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:28:31 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:28:31.016+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:29:09 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:29:09.869+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:29:51 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:29:51.828+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:30:28 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:30:28.149+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:31:07 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:31:07.682+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:31:36 Server k3s[5772]: time="2023-12-24T08:31:36+01:00" level=info msg="COMPACT compactRev=21885 targetCompactRev=22399 currentRev=23399"
Dec 24 08:31:36 Server k3s[5772]: time="2023-12-24T08:31:36+01:00" level=info msg="COMPACT deleted 514 rows from 514 revisions in 25.572201ms - compacted to 22399/23399"
Dec 24 08:31:49 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:31:49.336+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:32:30 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:32:30.962+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:33:13 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:33:13.946+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:33:52 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:33:52.757+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:34:23 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:34:23.463+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:35:05 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:35:05.542+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:35:40 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:35:40.985+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:36:23 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:36:23.970+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:36:36 Server k3s[5772]: time="2023-12-24T08:36:36+01:00" level=info msg="COMPACT compactRev=22399 targetCompactRev=22913 currentRev=23913"
Dec 24 08:36:36 Server k3s[5772]: time="2023-12-24T08:36:36+01:00" level=info msg="COMPACT deleted 514 rows from 514 revisions in 9.047706ms - compacted to 22913/23913"
Dec 24 08:36:55 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:36:55.486+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:37:31 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:37:31.776+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:38:15 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:38:15.744+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:38:55 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:38:55.638+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:39:33 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:39:33.159+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:40:12 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:40:12.879+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:40:56 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:40:56.388+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:41:36 Server k3s[5772]: time="2023-12-24T08:41:36+01:00" level=info msg="COMPACT compactRev=22913 targetCompactRev=23823 currentRev=24823"
Dec 24 08:41:36 Server k3s[5772]: time="2023-12-24T08:41:36+01:00" level=info msg="COMPACT deleted 910 rows from 910 revisions in 9.504879ms - compacted to 23823/24823"
Dec 24 08:41:40 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:41:40.447+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:42:18 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:42:18.480+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:42:51 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:42:51.281+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:43:35 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:43:35.728+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:44:08 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:44:08.060+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:44:51 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:44:51.066+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:45:22 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:45:22.978+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:46:05 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:46:05.070+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:46:36 Server k3s[5772]: time="2023-12-24T08:46:36+01:00" level=info msg="COMPACT compactRev=23823 targetCompactRev=24336 currentRev=25336"
Dec 24 08:46:36 Server k3s[5772]: time="2023-12-24T08:46:36+01:00" level=info msg="COMPACT deleted 809 rows from 513 revisions in 11.047991ms - compacted to 24336/25336"
Dec 24 08:46:39 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:46:39.458+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:47:19 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:47:19.189+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:47:54 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:47:54.383+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:48:28 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:48:28.074+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:49:01 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:49:01.912+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:49:44 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:49:44.884+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:50:15 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:50:15.346+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:50:48 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:50:48.293+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:51:27 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:51:27.175+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:51:36 Server k3s[5772]: time="2023-12-24T08:51:36+01:00" level=info msg="COMPACT compactRev=24336 targetCompactRev=24851 currentRev=25851"
Dec 24 08:51:36 Server k3s[5772]: time="2023-12-24T08:51:36+01:00" level=info msg="COMPACT deleted 615 rows from 515 revisions in 22.503406ms - compacted to 24851/25851"
Dec 24 08:52:11 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:52:11.697+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:52:48 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:52:48.696+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:53:26 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:53:26.422+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:54:10 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:54:10.676+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:54:54 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:54:54.561+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:55:34 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:55:34.927+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:56:16 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:56:16.230+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:56:36 Server k3s[5772]: time="2023-12-24T08:56:36+01:00" level=info msg="COMPACT compactRev=24851 targetCompactRev=25363 currentRev=26363"
Dec 24 08:56:36 Server k3s[5772]: time="2023-12-24T08:56:36+01:00" level=info msg="COMPACT deleted 512 rows from 512 revisions in 11.282129ms - compacted to 25363/26363"
Dec 24 08:56:54 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:56:54.580+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:57:37 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:57:37.041+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:58:14 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:58:14.243+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:58:45 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:58:45.601+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 08:59:17 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T08:59:17.093+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:00:00 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:00:00.713+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:00:43 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:00:43.524+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:01:16 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:01:16.632+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:01:36 Server k3s[5772]: time="2023-12-24T09:01:36+01:00" level=info msg="COMPACT compactRev=25363 targetCompactRev=25877 currentRev=26877"
Dec 24 09:01:36 Server k3s[5772]: time="2023-12-24T09:01:36+01:00" level=info msg="COMPACT deleted 514 rows from 514 revisions in 21.435439ms - compacted to 25877/26877"
Dec 24 09:01:59 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:01:59.479+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:02:43 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:02:43.861+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:03:18 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:03:18.592+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:03:58 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:03:58.884+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:04:42 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:04:42.178+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:05:14 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:05:14.098+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:05:53 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:05:53.197+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:06:35 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:06:35.385+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:06:36 Server k3s[5772]: time="2023-12-24T09:06:36+01:00" level=info msg="COMPACT compactRev=25877 targetCompactRev=26392 currentRev=27392"
Dec 24 09:06:36 Server k3s[5772]: time="2023-12-24T09:06:36+01:00" level=info msg="COMPACT deleted 515 rows from 515 revisions in 10.320178ms - compacted to 26392/27392"
Dec 24 09:07:13 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:07:13.019+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:07:57 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:07:57.807+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:08:31 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:08:31.845+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:09:13 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:09:13.501+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:09:53 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:09:53.383+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:10:33 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:10:33.715+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:11:06 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:11:06.838+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:11:36 Server k3s[5772]: time="2023-12-24T09:11:36+01:00" level=info msg="COMPACT compactRev=26392 targetCompactRev=26905 currentRev=27905"
Dec 24 09:11:36 Server k3s[5772]: time="2023-12-24T09:11:36+01:00" level=info msg="COMPACT deleted 513 rows from 513 revisions in 9.387455ms - compacted to 26905/27905"
Dec 24 09:11:50 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:11:50.921+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:12:33 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:12:33.310+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:13:12 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:13:12.102+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:13:49 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:13:49.673+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:14:31 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:14:31.732+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:15:14 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:15:14.491+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:15:45 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:15:45.135+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:16:17 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:16:17.067+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:16:36 Server k3s[5772]: time="2023-12-24T09:16:36+01:00" level=info msg="COMPACT compactRev=26905 targetCompactRev=27418 currentRev=28418"
Dec 24 09:16:36 Server k3s[5772]: time="2023-12-24T09:16:36+01:00" level=info msg="COMPACT deleted 513 rows from 513 revisions in 9.085758ms - compacted to 27418/28418"
Dec 24 09:16:55 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:16:55.420+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:17:39 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:17:39.009+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:18:20 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:18:20.859+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:18:55 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:18:55.635+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:19:33 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:19:33.078+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:20:04 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:20:04.048+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:20:42 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:20:42.267+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:21:13 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:21:13.152+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:21:36 Server k3s[5772]: time="2023-12-24T09:21:36+01:00" level=info msg="COMPACT compactRev=27418 targetCompactRev=27933 currentRev=28933"
Dec 24 09:21:36 Server k3s[5772]: time="2023-12-24T09:21:36+01:00" level=info msg="COMPACT deleted 515 rows from 515 revisions in 7.77238ms - compacted to 27933/28933"
Dec 24 09:21:44 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:21:44.163+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:22:23 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:22:23.325+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:22:58 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:22:58.830+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:23:37 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:23:37.796+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:24:15 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:24:15.620+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:24:53 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:24:53.145+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:25:26 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:25:26.124+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:25:59 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:25:59.283+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:26:36 Server k3s[5772]: time="2023-12-24T09:26:36+01:00" level=info msg="COMPACT compactRev=27933 targetCompactRev=28446 currentRev=29446"
Dec 24 09:26:36 Server k3s[5772]: time="2023-12-24T09:26:36+01:00" level=info msg="COMPACT deleted 513 rows from 513 revisions in 18.007699ms - compacted to 28446/29446"
Dec 24 09:26:41 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:26:41.460+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:27:26 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:27:26.286+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:28:02 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:28:02.209+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:28:34 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:28:34.671+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:29:08 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:29:08.716+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:29:43 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:29:43.961+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:30:21 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:30:21.724+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:30:59 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:30:59.593+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:31:33 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:31:33.002+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:31:36 Server k3s[5772]: time="2023-12-24T09:31:36+01:00" level=info msg="COMPACT compactRev=28446 targetCompactRev=28960 currentRev=29960"
Dec 24 09:31:36 Server k3s[5772]: time="2023-12-24T09:31:36+01:00" level=info msg="COMPACT deleted 514 rows from 514 revisions in 7.847062ms - compacted to 28960/29960"
Dec 24 09:32:10 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:32:10.108+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:32:54 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:32:54.730+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:33:37 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:33:37.775+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:34:11 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:34:11.927+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:34:47 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:34:47.831+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:35:19 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:35:19.919+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:36:00 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:36:00.647+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:36:36 Server k3s[5772]: time="2023-12-24T09:36:36+01:00" level=info msg="COMPACT compactRev=28960 targetCompactRev=29474 currentRev=30474"
Dec 24 09:36:36 Server k3s[5772]: time="2023-12-24T09:36:36+01:00" level=info msg="COMPACT deleted 514 rows from 514 revisions in 7.953239ms - compacted to 29474/30474"
Dec 24 09:36:40 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:36:40.247+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:37:11 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:37:11.777+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:37:46 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:37:46.602+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:38:20 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:38:20.983+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:38:56 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:38:56.139+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:39:30 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:39:30.704+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:40:01 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:40:01.369+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:40:46 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:40:46.142+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:41:21 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:41:21.071+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:41:36 Server k3s[5772]: time="2023-12-24T09:41:36+01:00" level=info msg="COMPACT compactRev=29474 targetCompactRev=29988 currentRev=30988"
Dec 24 09:41:36 Server k3s[5772]: time="2023-12-24T09:41:36+01:00" level=info msg="COMPACT deleted 514 rows from 514 revisions in 7.662127ms - compacted to 29988/30988"
Dec 24 09:41:58 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:41:58.744+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:42:35 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:42:35.121+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:43:16 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:43:16.222+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:43:52 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:43:52.378+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:44:25 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:44:25.583+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:45:05 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:45:05.548+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:45:47 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:45:47.573+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:46:21 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:46:21.484+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:46:36 Server k3s[5772]: time="2023-12-24T09:46:36+01:00" level=info msg="COMPACT compactRev=29988 targetCompactRev=30502 currentRev=31502"
Dec 24 09:46:36 Server k3s[5772]: time="2023-12-24T09:46:36+01:00" level=info msg="COMPACT deleted 514 rows from 514 revisions in 16.976766ms - compacted to 30502/31502"
Dec 24 09:46:55 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:46:55.587+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:47:32 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:47:32.155+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:48:13 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:48:13.116+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:48:51 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:48:51.191+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:49:31 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:49:31.198+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:50:06 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:50:06.382+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:50:39 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:50:39.105+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:51:13 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:51:13.404+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:51:36 Server k3s[5772]: time="2023-12-24T09:51:36+01:00" level=info msg="COMPACT compactRev=30502 targetCompactRev=31014 currentRev=32014"
Dec 24 09:51:36 Server k3s[5772]: time="2023-12-24T09:51:36+01:00" level=info msg="COMPACT deleted 512 rows from 512 revisions in 6.687195ms - compacted to 31014/32014"
Dec 24 09:51:56 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:51:56.273+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:52:36 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:52:36.744+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:53:16 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:53:16.650+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:53:50 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:53:50.928+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:54:29 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:54:29.794+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:55:05 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:55:05.431+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
Dec 24 09:55:43 Server k3s[5772]: {"level":"warn","ts":"2023-12-24T09:55:43.866+0100","logger":"etcd-client","caller":"v3@v3.5.7-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0007e1dc0/kine.sock","attempt":0,"error":"rpc error: code = Unknown desc = no such table: dbstat"}
