
------------------------------------
1) Describe 'Daemonsets' resources
------------------------------------

Name:           csi-nfs-node
Selector:       app=csi-nfs-node
Node-Selector:  kubernetes.io/os=linux
Labels:         objectset.rio.cattle.io/hash=e20aeb27d825538dc5fa9f4d82c2fe2058f8374c
Annotations:    deprecated.daemonset.template.generation: 1
                objectset.rio.cattle.io/applied:
                  H4sIAAAAAAAA/8xW328auRP/V74a9aFfaRcIhCtdiYc00DbqhaCQ9nSqImTsWZjDa+/Z3m24aP/3k70LoSlJuOoe7gHJ6/n9mc8MvgeW0xc0lrSCBFie23Z5AhGsSQlIYMQw02qGDi...
                objectset.rio.cattle.io/id: 
                objectset.rio.cattle.io/owner-gvk: k3s.cattle.io/v1, Kind=Addon
                objectset.rio.cattle.io/owner-name: nfs-csi
                objectset.rio.cattle.io/owner-namespace: kube-system
Desired Number of Nodes Scheduled: 1
Current Number of Nodes Scheduled: 1
Number of Nodes Scheduled with Up-to-date Pods: 1
Number of Nodes Scheduled with Available Pods: 1
Number of Nodes Misscheduled: 0
Pods Status:  1 Running / 0 Waiting / 0 Succeeded / 0 Failed
Pod Template:
  Labels:           app=csi-nfs-node
  Service Account:  csi-nfs-node-sa
  Containers:
   liveness-probe:
    Image:      registry.k8s.io/sig-storage/livenessprobe:v2.10.0
    Port:       <none>
    Host Port:  <none>
    Args:
      --csi-address=/csi/csi.sock
      --probe-timeout=3s
      --health-port=29653
      --v=2
    Limits:
      memory:  100Mi
    Requests:
      cpu:        10m
      memory:     20Mi
    Environment:  <none>
    Mounts:
      /csi from socket-dir (rw)
   node-driver-registrar:
    Image:      registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.8.0
    Port:       <none>
    Host Port:  <none>
    Args:
      --v=2
      --csi-address=/csi/csi.sock
      --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)
    Limits:
      memory:  100Mi
    Requests:
      cpu:     10m
      memory:  20Mi
    Liveness:  exec [/csi-node-driver-registrar --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH) --mode=kubelet-registration-probe] delay=30s timeout=15s period=10s #success=1 #failure=3
    Environment:
      DRIVER_REG_SOCK_PATH:  /var/lib/kubelet/plugins/csi-nfsplugin/csi.sock
      KUBE_NODE_NAME:         (v1:spec.nodeName)
    Mounts:
      /csi from socket-dir (rw)
      /registration from registration-dir (rw)
   nfs:
    Image:      registry.k8s.io/sig-storage/nfsplugin:v4.4.0
    Port:       29653/TCP
    Host Port:  29653/TCP
    Args:
      -v=5
      --nodeid=$(NODE_ID)
      --endpoint=$(CSI_ENDPOINT)
    Limits:
      memory:  300Mi
    Requests:
      cpu:     10m
      memory:  20Mi
    Liveness:  http-get http://:healthz/healthz delay=30s timeout=10s period=30s #success=1 #failure=5
    Environment:
      NODE_ID:        (v1:spec.nodeName)
      CSI_ENDPOINT:  unix:///csi/csi.sock
    Mounts:
      /csi from socket-dir (rw)
      /var/lib/kubelet/pods from pods-mount-dir (rw)
  Volumes:
   socket-dir:
    Type:          HostPath (bare host directory volume)
    Path:          /var/lib/kubelet/plugins/csi-nfsplugin
    HostPathType:  DirectoryOrCreate
   pods-mount-dir:
    Type:          HostPath (bare host directory volume)
    Path:          /var/lib/kubelet/pods
    HostPathType:  Directory
   registration-dir:
    Type:               HostPath (bare host directory volume)
    Path:               /var/lib/kubelet/plugins_registry
    HostPathType:       Directory
  Priority Class Name:  system-node-critical
Events:
  Type     Reason            Age   From                  Message
  ----     ------            ----  ----                  -------
  Warning  FailedDaemonPod   42m   daemonset-controller  Found failed daemon pod kube-system/csi-nfs-node-pv7lm on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  42m   daemonset-controller  Deleted pod: csi-nfs-node-pv7lm
  Normal   SuccessfulCreate  42m   daemonset-controller  Created pod: csi-nfs-node-lcfxf
  Warning  FailedDaemonPod   42m   daemonset-controller  Found failed daemon pod kube-system/csi-nfs-node-lcfxf on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  42m   daemonset-controller  Deleted pod: csi-nfs-node-lcfxf
  Normal   SuccessfulCreate  42m   daemonset-controller  Created pod: csi-nfs-node-rgxsk
  Warning  FailedDaemonPod   41m   daemonset-controller  Found failed daemon pod kube-system/csi-nfs-node-rgxsk on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  41m   daemonset-controller  Deleted pod: csi-nfs-node-rgxsk
  Normal   SuccessfulCreate  41m   daemonset-controller  Created pod: csi-nfs-node-hl5qn
  Warning  FailedDaemonPod   24m   daemonset-controller  Found failed daemon pod kube-system/csi-nfs-node-hl5qn on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  24m   daemonset-controller  Deleted pod: csi-nfs-node-hl5qn
  Normal   SuccessfulCreate  24m   daemonset-controller  Created pod: csi-nfs-node-8j9mw


Name:           csi-smb-node
Selector:       app=csi-smb-node
Node-Selector:  kubernetes.io/os=linux
Labels:         objectset.rio.cattle.io/hash=dcff83f97210cf176ea098fd305c31dad9e74663
Annotations:    deprecated.daemonset.template.generation: 1
                objectset.rio.cattle.io/applied:
                  H4sIAAAAAAAA/8xWTW8bNxD9K8XABwfY1YcVO/YCOjiWkhhpZEFy0kNgCBQ5kqbikluSu5Fq7H8vyF19xJUSFe2hBwErDjl8897jkM/AMvqCxpJWkADLMtss2hDBkpSABHoMU63G6C...
                objectset.rio.cattle.io/id: 
                objectset.rio.cattle.io/owner-gvk: k3s.cattle.io/v1, Kind=Addon
                objectset.rio.cattle.io/owner-name: smb-csi
                objectset.rio.cattle.io/owner-namespace: kube-system
Desired Number of Nodes Scheduled: 1
Current Number of Nodes Scheduled: 1
Number of Nodes Scheduled with Up-to-date Pods: 1
Number of Nodes Scheduled with Available Pods: 1
Number of Nodes Misscheduled: 0
Pods Status:  1 Running / 0 Waiting / 0 Succeeded / 0 Failed
Pod Template:
  Labels:           app=csi-smb-node
  Service Account:  csi-smb-node-sa
  Containers:
   liveness-probe:
    Image:      registry.k8s.io/sig-storage/livenessprobe:v2.10.0
    Port:       <none>
    Host Port:  <none>
    Args:
      --csi-address=/csi/csi.sock
      --probe-timeout=3s
      --health-port=29643
      --v=2
    Limits:
      memory:  100Mi
    Requests:
      cpu:        10m
      memory:     20Mi
    Environment:  <none>
    Mounts:
      /csi from socket-dir (rw)
   node-driver-registrar:
    Image:      registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.8.0
    Port:       <none>
    Host Port:  <none>
    Args:
      --csi-address=$(ADDRESS)
      --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)
      --v=2
    Limits:
      memory:  100Mi
    Requests:
      cpu:     10m
      memory:  20Mi
    Liveness:  exec [/csi-node-driver-registrar --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH) --mode=kubelet-registration-probe] delay=30s timeout=15s period=10s #success=1 #failure=3
    Environment:
      ADDRESS:               /csi/csi.sock
      DRIVER_REG_SOCK_PATH:  /var/lib/kubelet/plugins/smb.csi.k8s.io/csi.sock
    Mounts:
      /csi from socket-dir (rw)
      /registration from registration-dir (rw)
   smb:
    Image:      registry.k8s.io/sig-storage/smbplugin:v1.11.0
    Port:       29643/TCP
    Host Port:  29643/TCP
    Args:
      --v=5
      --endpoint=$(CSI_ENDPOINT)
      --nodeid=$(KUBE_NODE_NAME)
    Limits:
      memory:  200Mi
    Requests:
      cpu:     10m
      memory:  20Mi
    Liveness:  http-get http://:healthz/healthz delay=30s timeout=10s period=30s #success=1 #failure=5
    Environment:
      CSI_ENDPOINT:    unix:///csi/csi.sock
      KUBE_NODE_NAME:   (v1:spec.nodeName)
    Mounts:
      /csi from socket-dir (rw)
      /var/lib/kubelet/ from mountpoint-dir (rw)
  Volumes:
   socket-dir:
    Type:          HostPath (bare host directory volume)
    Path:          /var/lib/kubelet/plugins/smb.csi.k8s.io
    HostPathType:  DirectoryOrCreate
   mountpoint-dir:
    Type:          HostPath (bare host directory volume)
    Path:          /var/lib/kubelet/
    HostPathType:  DirectoryOrCreate
   registration-dir:
    Type:               HostPath (bare host directory volume)
    Path:               /var/lib/kubelet/plugins_registry/
    HostPathType:       DirectoryOrCreate
  Priority Class Name:  system-node-critical
Events:
  Type     Reason            Age   From                  Message
  ----     ------            ----  ----                  -------
  Warning  FailedDaemonPod   42m   daemonset-controller  Found failed daemon pod kube-system/csi-smb-node-swr2p on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  42m   daemonset-controller  Deleted pod: csi-smb-node-swr2p
  Normal   SuccessfulCreate  42m   daemonset-controller  Created pod: csi-smb-node-lwcht
  Warning  FailedDaemonPod   42m   daemonset-controller  Found failed daemon pod kube-system/csi-smb-node-lwcht on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  42m   daemonset-controller  Deleted pod: csi-smb-node-lwcht
  Normal   SuccessfulCreate  42m   daemonset-controller  Created pod: csi-smb-node-lwsb9
  Warning  FailedDaemonPod   41m   daemonset-controller  Found failed daemon pod kube-system/csi-smb-node-lwsb9 on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  41m   daemonset-controller  Deleted pod: csi-smb-node-lwsb9
  Normal   SuccessfulCreate  41m   daemonset-controller  Created pod: csi-smb-node-txtfq
  Warning  FailedDaemonPod   24m   daemonset-controller  Found failed daemon pod kube-system/csi-smb-node-txtfq on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  24m   daemonset-controller  Deleted pod: csi-smb-node-txtfq
  Normal   SuccessfulCreate  24m   daemonset-controller  Created pod: csi-smb-node-r7glx


Name:           svclb-makemkv-84687e74
Selector:       app=svclb-makemkv-84687e74
Node-Selector:  <none>
Labels:         objectset.rio.cattle.io/hash=d701fc556db97b97b47a28f34a1d369ffed0b6a9
                svccontroller.k3s.cattle.io/nodeselector=false
                svccontroller.k3s.cattle.io/svcname=makemkv
                svccontroller.k3s.cattle.io/svcnamespace=ix-makemkv
Annotations:    deprecated.daemonset.template.generation: 1
                objectset.rio.cattle.io/applied:
                  H4sIAAAAAAAA/7xUUU/jOBD+K6d5TtNm6bZg6R4QoBO621K13XtZVWhiT6mvjm3ZkywVyn8/OaVc0JYF3cMKpCoz33y2v/lmngC9/ptC1M6CAPQ+DpsCMthpq0DANVLl7JIYMqiIUS...
                objectset.rio.cattle.io/id: 
                objectset.rio.cattle.io/owner-gvk: /v1, Kind=Service
                objectset.rio.cattle.io/owner-name: makemkv
                objectset.rio.cattle.io/owner-namespace: ix-makemkv
Desired Number of Nodes Scheduled: 1
Current Number of Nodes Scheduled: 1
Number of Nodes Scheduled with Up-to-date Pods: 1
Number of Nodes Scheduled with Available Pods: 1
Number of Nodes Misscheduled: 0
Pods Status:  1 Running / 0 Waiting / 0 Succeeded / 0 Failed
Pod Template:
  Labels:           app=svclb-makemkv-84687e74
                    svccontroller.k3s.cattle.io/svcname=makemkv
                    svccontroller.k3s.cattle.io/svcnamespace=ix-makemkv
  Service Account:  svclb
  Containers:
   lb-tcp-10180:
    Image:      rancher/klipper-lb:v0.4.4
    Port:       10180/TCP
    Host Port:  10180/TCP
    Environment:
      SRC_PORT:    10180
      SRC_RANGES:  0.0.0.0/0
      DEST_PROTO:  TCP
      DEST_PORT:   10180
      DEST_IPS:    172.17.166.94
    Mounts:        <none>
  Volumes:         <none>
Events:
  Type     Reason            Age   From                  Message
  ----     ------            ----  ----                  -------
  Warning  FailedDaemonPod   60m   daemonset-controller  Found failed daemon pod kube-system/svclb-makemkv-84687e74-lhbxq on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  60m   daemonset-controller  Deleted pod: svclb-makemkv-84687e74-lhbxq
  Normal   SuccessfulCreate  55m   daemonset-controller  Created pod: svclb-makemkv-84687e74-5ddzx
  Warning  FailedDaemonPod   42m   daemonset-controller  Found failed daemon pod kube-system/svclb-makemkv-84687e74-5ddzx on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  42m   daemonset-controller  Deleted pod: svclb-makemkv-84687e74-5ddzx
  Normal   SuccessfulCreate  41m   daemonset-controller  Created pod: svclb-makemkv-84687e74-fzzpk
  Warning  FailedDaemonPod   25m   daemonset-controller  Found failed daemon pod kube-system/svclb-makemkv-84687e74-fzzpk on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  25m   daemonset-controller  Deleted pod: svclb-makemkv-84687e74-fzzpk
  Normal   SuccessfulCreate  23m   daemonset-controller  Created pod: svclb-makemkv-84687e74-8kljd


Name:           svclb-makemkv-vnc-ad0786b1
Selector:       app=svclb-makemkv-vnc-ad0786b1
Node-Selector:  <none>
Labels:         objectset.rio.cattle.io/hash=e8d165e9e5cc834077c8efe796cfbd0f62c13e69
                svccontroller.k3s.cattle.io/nodeselector=false
                svccontroller.k3s.cattle.io/svcname=makemkv-vnc
                svccontroller.k3s.cattle.io/svcnamespace=ix-makemkv
Annotations:    deprecated.daemonset.template.generation: 1
                objectset.rio.cattle.io/applied:
                  H4sIAAAAAAAA/7xUTW/jNhD9K8WcJcVKsv4Q0EOQBEXQrmPY3l4WRjAixzEriiTIkTZGoP9eUHZSBXU2QQ+FD4bI9x7JN2/mGdCpP8kHZQ0UgM6FszaHBCplJBRwg1RbsyKGBGpilM...
                objectset.rio.cattle.io/id: 
                objectset.rio.cattle.io/owner-gvk: /v1, Kind=Service
                objectset.rio.cattle.io/owner-name: makemkv-vnc
                objectset.rio.cattle.io/owner-namespace: ix-makemkv
Desired Number of Nodes Scheduled: 1
Current Number of Nodes Scheduled: 1
Number of Nodes Scheduled with Up-to-date Pods: 1
Number of Nodes Scheduled with Available Pods: 1
Number of Nodes Misscheduled: 0
Pods Status:  1 Running / 0 Waiting / 0 Succeeded / 0 Failed
Pod Template:
  Labels:           app=svclb-makemkv-vnc-ad0786b1
                    svccontroller.k3s.cattle.io/svcname=makemkv-vnc
                    svccontroller.k3s.cattle.io/svcnamespace=ix-makemkv
  Service Account:  svclb
  Containers:
   lb-tcp-10181:
    Image:      rancher/klipper-lb:v0.4.4
    Port:       10181/TCP
    Host Port:  10181/TCP
    Environment:
      SRC_PORT:    10181
      SRC_RANGES:  0.0.0.0/0
      DEST_PROTO:  TCP
      DEST_PORT:   10181
      DEST_IPS:    172.17.99.26
    Mounts:        <none>
  Volumes:         <none>
Events:
  Type     Reason            Age   From                  Message
  ----     ------            ----  ----                  -------
  Warning  FailedDaemonPod   60m   daemonset-controller  Found failed daemon pod kube-system/svclb-makemkv-vnc-ad0786b1-x778h on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  60m   daemonset-controller  Deleted pod: svclb-makemkv-vnc-ad0786b1-x778h
  Normal   SuccessfulCreate  55m   daemonset-controller  Created pod: svclb-makemkv-vnc-ad0786b1-98d79
  Warning  FailedDaemonPod   42m   daemonset-controller  Found failed daemon pod kube-system/svclb-makemkv-vnc-ad0786b1-98d79 on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  42m   daemonset-controller  Deleted pod: svclb-makemkv-vnc-ad0786b1-98d79
  Normal   SuccessfulCreate  41m   daemonset-controller  Created pod: svclb-makemkv-vnc-ad0786b1-tffhx
  Warning  FailedDaemonPod   25m   daemonset-controller  Found failed daemon pod kube-system/svclb-makemkv-vnc-ad0786b1-tffhx on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  25m   daemonset-controller  Deleted pod: svclb-makemkv-vnc-ad0786b1-tffhx
  Normal   SuccessfulCreate  23m   daemonset-controller  Created pod: svclb-makemkv-vnc-ad0786b1-l48xl


Name:           svclb-openspeedtest-5fe4cae5
Selector:       app=svclb-openspeedtest-5fe4cae5
Node-Selector:  <none>
Labels:         objectset.rio.cattle.io/hash=b893a029f715aa01e2a252ba4450dc3c3bda63ce
                svccontroller.k3s.cattle.io/nodeselector=false
                svccontroller.k3s.cattle.io/svcname=openspeedtest
                svccontroller.k3s.cattle.io/svcnamespace=ix-openspeedtest
Annotations:    deprecated.daemonset.template.generation: 1
                objectset.rio.cattle.io/applied:
                  H4sIAAAAAAAA/8RUT2/jthP9Kj/MWZYt/9n8QqCHIAmKoF3HsL29LIxgRI5j1hRJkCNtjEDfvaDspDKabIJeCh8MkW/ecN68mWdAr/+gELWzIAC9j8OmgAz22ioQcINUObsihgwqYl...
                objectset.rio.cattle.io/id: 
                objectset.rio.cattle.io/owner-gvk: /v1, Kind=Service
                objectset.rio.cattle.io/owner-name: openspeedtest
                objectset.rio.cattle.io/owner-namespace: ix-openspeedtest
Desired Number of Nodes Scheduled: 1
Current Number of Nodes Scheduled: 1
Number of Nodes Scheduled with Up-to-date Pods: 1
Number of Nodes Scheduled with Available Pods: 1
Number of Nodes Misscheduled: 0
Pods Status:  1 Running / 0 Waiting / 0 Succeeded / 0 Failed
Pod Template:
  Labels:           app=svclb-openspeedtest-5fe4cae5
                    svccontroller.k3s.cattle.io/svcname=openspeedtest
                    svccontroller.k3s.cattle.io/svcnamespace=ix-openspeedtest
  Service Account:  svclb
  Containers:
   lb-tcp-10256:
    Image:      rancher/klipper-lb:v0.4.4
    Port:       10256/TCP
    Host Port:  10256/TCP
    Environment:
      SRC_PORT:    10256
      SRC_RANGES:  0.0.0.0/0
      DEST_PROTO:  TCP
      DEST_PORT:   10256
      DEST_IPS:    172.17.139.9
    Mounts:        <none>
  Volumes:         <none>
Events:
  Type     Reason            Age   From                  Message
  ----     ------            ----  ----                  -------
  Warning  FailedDaemonPod   60m   daemonset-controller  Found failed daemon pod kube-system/svclb-openspeedtest-5fe4cae5-w25x6 on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  60m   daemonset-controller  Deleted pod: svclb-openspeedtest-5fe4cae5-w25x6
  Normal   SuccessfulCreate  55m   daemonset-controller  Created pod: svclb-openspeedtest-5fe4cae5-5hck9
  Warning  FailedDaemonPod   42m   daemonset-controller  Found failed daemon pod kube-system/svclb-openspeedtest-5fe4cae5-5hck9 on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  42m   daemonset-controller  Deleted pod: svclb-openspeedtest-5fe4cae5-5hck9
  Normal   SuccessfulCreate  41m   daemonset-controller  Created pod: svclb-openspeedtest-5fe4cae5-zfvl7
  Warning  FailedDaemonPod   25m   daemonset-controller  Found failed daemon pod kube-system/svclb-openspeedtest-5fe4cae5-zfvl7 on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  25m   daemonset-controller  Deleted pod: svclb-openspeedtest-5fe4cae5-zfvl7
  Normal   SuccessfulCreate  23m   daemonset-controller  Created pod: svclb-openspeedtest-5fe4cae5-hxgdq


Name:           openebs-zfs-node
Selector:       app=openebs-zfs-node
Node-Selector:  <none>
Labels:         objectset.rio.cattle.io/hash=4b1efd47d7f37713706abe9f8b0de3f0adf02e2c
                openebs.io/component-name=openebs-zfs-node
                openebs.io/version=2.3.0
Annotations:    deprecated.daemonset.template.generation: 2
                objectset.rio.cattle.io/applied:
                  H4sIAAAAAAAA/6xWUY/aOBD+K5XVk1opIezSdnuR9oFCrl11FxBsew/VCjn2JPhw7Mh20uVW+e8nOwEChG1P2rfEnvlm5pvPYz8hnLPvoDSTAoUI57kOygvkoTUTFIVojCGTYgEGeS...
                objectset.rio.cattle.io/id: 
                objectset.rio.cattle.io/owner-gvk: k3s.cattle.io/v1, Kind=Addon
                objectset.rio.cattle.io/owner-name: zfs-operator
                objectset.rio.cattle.io/owner-namespace: kube-system
Desired Number of Nodes Scheduled: 1
Current Number of Nodes Scheduled: 1
Number of Nodes Scheduled with Up-to-date Pods: 1
Number of Nodes Scheduled with Available Pods: 1
Number of Nodes Misscheduled: 0
Pods Status:  1 Running / 0 Waiting / 0 Succeeded / 0 Failed
Pod Template:
  Labels:           app=openebs-zfs-node
                    openebs.io/component-name=openebs-zfs-node
                    openebs.io/version=2.3.0
                    role=openebs-zfs
  Service Account:  openebs-zfs-node-sa
  Containers:
   csi-node-driver-registrar:
    Image:      registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.8.0
    Port:       <none>
    Host Port:  <none>
    Args:
      --v=5
      --csi-address=$(ADDRESS)
      --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)
    Environment:
      ADDRESS:               /plugin/csi.sock
      DRIVER_REG_SOCK_PATH:  /var/lib/kubelet/plugins/zfs-localpv/csi.sock
      KUBE_NODE_NAME:         (v1:spec.nodeName)
      NODE_DRIVER:           openebs-zfs
    Mounts:
      /plugin from plugin-dir (rw)
      /registration from registration-dir (rw)
   openebs-zfs-plugin:
    Image:      openebs/zfs-driver:2.3.0
    Port:       <none>
    Host Port:  <none>
    Args:
      --nodename=$(OPENEBS_NODE_NAME)
      --endpoint=$(OPENEBS_CSI_ENDPOINT)
      --plugin=$(OPENEBS_NODE_DRIVER)
    Environment:
      OPENEBS_NODE_NAME:      (v1:spec.nodeName)
      OPENEBS_CSI_ENDPOINT:  unix:///plugin/csi.sock
      OPENEBS_NODE_DRIVER:   agent
      OPENEBS_NAMESPACE:     openebs
      ALLOWED_TOPOLOGIES:    All
    Mounts:
      /dev from device-dir (rw)
      /home/keys from encr-keys (rw)
      /host from host-root (ro)
      /plugin from plugin-dir (rw)
      /sbin/zfs from chroot-zfs (rw,path="zfs")
      /var/lib/kubelet/ from pods-mount-dir (rw)
  Volumes:
   device-dir:
    Type:          HostPath (bare host directory volume)
    Path:          /dev
    HostPathType:  Directory
   encr-keys:
    Type:          HostPath (bare host directory volume)
    Path:          /home/keys
    HostPathType:  DirectoryOrCreate
   chroot-zfs:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      openebs-zfspv-bin
    Optional:  false
   host-root:
    Type:          HostPath (bare host directory volume)
    Path:          /
    HostPathType:  Directory
   registration-dir:
    Type:          HostPath (bare host directory volume)
    Path:          /var/lib/kubelet/plugins_registry/
    HostPathType:  DirectoryOrCreate
   plugin-dir:
    Type:          HostPath (bare host directory volume)
    Path:          /var/lib/kubelet/plugins/zfs-localpv/
    HostPathType:  DirectoryOrCreate
   pods-mount-dir:
    Type:               HostPath (bare host directory volume)
    Path:               /var/lib/kubelet/
    HostPathType:       Directory
  Priority Class Name:  openebs-zfs-csi-node-critical
Events:
  Type     Reason            Age   From                  Message
  ----     ------            ----  ----                  -------
  Warning  FailedDaemonPod   55m   daemonset-controller  Found failed daemon pod kube-system/openebs-zfs-node-mjzvt on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  55m   daemonset-controller  Deleted pod: openebs-zfs-node-mjzvt
  Normal   SuccessfulCreate  55m   daemonset-controller  Created pod: openebs-zfs-node-4gd58
  Warning  FailedDaemonPod   42m   daemonset-controller  Found failed daemon pod kube-system/openebs-zfs-node-4gd58 on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  42m   daemonset-controller  Deleted pod: openebs-zfs-node-4gd58
  Normal   SuccessfulCreate  41m   daemonset-controller  Created pod: openebs-zfs-node-8pd46
  Warning  FailedDaemonPod   24m   daemonset-controller  Found failed daemon pod kube-system/openebs-zfs-node-8pd46 on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  24m   daemonset-controller  Deleted pod: openebs-zfs-node-8pd46
  Normal   SuccessfulCreate  23m   daemonset-controller  Created pod: openebs-zfs-node-frcqf


Name:           nvidia-device-plugin-daemonset
Selector:       name=nvidia-device-plugin-ds
Node-Selector:  <none>
Labels:         <none>
Annotations:    deprecated.daemonset.template.generation: 2
Desired Number of Nodes Scheduled: 1
Current Number of Nodes Scheduled: 1
Number of Nodes Scheduled with Up-to-date Pods: 1
Number of Nodes Scheduled with Available Pods: 1
Number of Nodes Misscheduled: 0
Pods Status:  1 Running / 0 Waiting / 0 Succeeded / 0 Failed
Pod Template:
  Labels:       name=nvidia-device-plugin-ds
  Annotations:  scheduler.alpha.kubernetes.io/critical-pod: 
  Containers:
   nvidia-device-plugin-ctr:
    Image:      nvcr.io/nvidia/k8s-device-plugin:v0.13.0
    Port:       <none>
    Host Port:  <none>
    Command:
      nvidia-device-plugin
      --config-file
      /etc/config/nvdefault.yaml
    Environment:  <none>
    Mounts:
      /etc/config from plugin-config (rw)
      /var/lib/kubelet/device-plugins from device-plugin (rw)
  Volumes:
   device-plugin:
    Type:          HostPath (bare host directory volume)
    Path:          /var/lib/kubelet/device-plugins
    HostPathType:  
   plugin-config:
    Type:               ConfigMap (a volume populated by a ConfigMap)
    Name:               nvidia-device-plugin-config
    Optional:           false
  Priority Class Name:  system-node-critical
Events:
  Type     Reason            Age   From                  Message
  ----     ------            ----  ----                  -------
  Warning  FailedDaemonPod   55m   daemonset-controller  Found failed daemon pod kube-system/nvidia-device-plugin-daemonset-zjvgb on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  55m   daemonset-controller  Deleted pod: nvidia-device-plugin-daemonset-zjvgb
  Normal   SuccessfulCreate  55m   daemonset-controller  Created pod: nvidia-device-plugin-daemonset-cbbkm
  Normal   SuccessfulCreate  41m   daemonset-controller  Created pod: nvidia-device-plugin-daemonset-bnzq6
  Warning  FailedDaemonPod   23m   daemonset-controller  Found failed daemon pod kube-system/nvidia-device-plugin-daemonset-bnzq6 on node ix-truenas, will try to kill it
  Normal   SuccessfulDelete  23m   daemonset-controller  Deleted pod: nvidia-device-plugin-daemonset-bnzq6
  Normal   SuccessfulCreate  23m   daemonset-controller  Created pod: nvidia-device-plugin-daemonset-pjd4w
