Node Disk Manager User Guide
OpenEBS Documentation is now migrated to https://openebs.io/docs. The page you are currently viewing is a static snapshot and will be removed in the upcoming releases.
This section provides the operations need to performed by the Admin for configuring NDM.
Admin operations
Create blockdevice CRs for unsupported disks
Admin Operations
Include filters
Users can include only selected block device for the creation of blockdevice CR and then use only the created blockdevice CR for cStor pool creation or using for provisioning Local PV based on device. For including the selected blockdevices, update the operator YAML file with the required blockdevices under NDM configuration section so that only these blockdevice will be taken for the creation of blockdevice CR. Add the blockdevice path in the following configuration for specifying particular disks. This configuration must be added in openebs-ndm-config under Configmap in openebs-operator.yaml.
This change must be done in the openebs-operator.yaml file that you have downloaded before OpenEBS installation. If the change is performed after the OpenEBS installation, then user must restart corresponding NDM DaemonSet pods to update the NDM configuration.
filterconfigs:
- key: path-filter
name: path filter
state: true
include: "/dev/sda"
exclude: ""
When the above configuration is used, only the block device /dev/sda will be used for creating the block device custom resource. All other disks will be excluded.
Note:
- Regex support is not available on the
filterconfigsin NDMConfigmapand theConfigmapis applicable to cluster level. This means, if user provide/dev/sdbinfilterconfigsas an include filter, then all/dev/sdbblockdevices from all nodes in the cluster will be used for the creation of blockdevice CR by NDM.
Exclude filters
NDM do some filtering on the disks to exclude, for example boot disk. By default, NDM excludes the following device path while creating block device CR. This configuration is added in openebs-ndm-config under Configmap in openebs-operator.yaml.
/dev/loop - loop devices.
/dev/fd - file descriptors.
/dev/sr - CD-ROM devices.
/dev/ram - ramdisks.
/dev/dm -lvm.
/dev/md -multiple device ( software RAID devices).
/dev/rbd - ceph RBD devices
/dev/zd - zfs volumes
The following is the snippet of NDM configuration file from openebs operator YAML which excludes the provided disks/paths.
filterconfigs:
- key: os-disk-exclude-filter
name: os disk exclude filter
state: true
exclude: "/,/etc/hosts,/boot"
- key: vendor-filter
name: vendor filter
state: true
include: ""
exclude: "CLOUDBYT,OpenEBS"
- key: path-filter
name: path filter
state: true
include: ""
exclude: "loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/rbd"
It is also possible to customize by adding more disk types associated with your nodes. For example, used disks, unwanted disks and so on.
This change must be done in the openebs-operator.yaml file that you have downloaded before OpenEBS installation. If the change is performed after the OpenEBS installation, then user must restart corresponding NDM DaemonSet pods to update the NDM configuration.
filterconfigs:
- key: path-filter
name: path filter
state: true
include: ""
exclude: "loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/sdb"
Note:
Regex support is not available on the
filterconfigsin NDMConfigmapand the Configmap is applicable to cluster level. This means, if user provide/dev/sdbin configmap as an excluded filter, then all/dev/sdbblockdevices from all nodes in the cluster will be excluded by NDM.It is recommended to use OpenEBS provisioner alone in the cluster. If you are using other storage provider provisioner like
gce-pdalong with OpenEBS, use exclude filters to avoid those disks from being consumed by OpenEBS. For example, if you are using thestandardstorage class in GKE with storage provisioner as kubernetes.io/gce-pd, and when it creates a PVC, a GPD is attached to the node. This GPD will be detected by NDM and it may be used by OpenEBS for provisioning volume. To avoid this scenario, it is recommended to put the associated device path created on the node in the exclude field under path-filter. If GPD is attached as/dev/sdc, then add/dev/sdcin the above mentioned field.Example snippet:
In the downloaded openebs-operator.yaml, find openebs-ndm-config configmap and update the values for path-filter and any other filters if required.
--- # This is the node-disk-manager related config. # It can be used to customize the disks probes and filters apiVersion: v1 kind: ConfigMap metadata: name: openebs-ndm-config namespace: openebs data: # udev-probe is default or primary probe which should be enabled to run ndm # filterconfigs contains configs of filters - in their form fo include # and exclude comma separated strings node-disk-manager.config: | probeconfigs: - key: udev-probe name: udev probe state: true - key: seachest-probe name: seachest probe state: false - key: smart-probe name: smart probe state: true filterconfigs: - key: os-disk-exclude-filter name: os disk exclude filter state: true exclude: "/,/etc/hosts,/boot" - key: vendor-filter name: vendor filter state: true include: "" exclude: "CLOUDBYT,OpenEBS" - key: path-filter name: path filter state: true include: "" exclude: "loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/sdc" ---
Create blockdevice CRs for unsupported disks
Currently, NDM out of the box manages disks, partitions, lvm, crypt and other dm devices. If the user need to have blockdevice for other device types like md array or any other unsupported device types, the blockdevice resource can be manually created using the following steps:
Create the sample block device CR YAML using the following spec. Following is the sample block device CR YAML.
apiVersion: openebs.io/v1alpha1 kind: BlockDevice metadata: name: example-blockdevice-1 labels: kubernetes.io/hostname: <host name in which disk/blockdevice is attached> # like gke-openebs-user-default-pool-044afcb8-bmc0 ndm.io/managed: "false" # for manual blockdevice creation put false ndm.io/blockdevice-type: blockdevice status: claimState: Unclaimed state: Active spec: capacity: logicalSectorSize: 512 storage: <total capacity in bytes> #like 53687091200 details: deviceType: <device type> # like disk, partition, lvm, crypt, md firmwareRevision: <firmware revision> model: <model name of blockdevice> # like PersistentDisk serial: <serial no of disk> # like google-disk-2 compliance: <compliance of disk> #like "SPC-4" vendor: <vendor of disk> #like Google devlinks: - kind: by-id links: - <link1> # like /dev/disk/by-id/scsi-0Google_PersistentDisk_disk-2 - <link2> # like /dev/disk/by-id/google-disk-2 - kind: by-path links: - <link1> # like /dev/disk/by-path/virtio-pci-0000:00:03.0-scsi-0:0:2:0 nodeAttributes: nodeName: <node name> # output of `kubectl get nodes` can be used path: <devpath> # like /dev/md0Modify the created block device CR sample YAML with the disk information. In the above block device CR sample spec, following fields must be filled before applying the YAML.
- name
- Provide unique name for the blockdevice CR. In the above YAML spec, given name for the blockdevice CR is
example-blockdevice-1
- Provide unique name for the blockdevice CR. In the above YAML spec, given name for the blockdevice CR is
- kubernetes.io/hostname
- Hostname of the node where the blockdevice is attached.
- storage
- Provide the storage capacity in
byteslike53687091200.
- Provide the storage capacity in
- logicalSectorSize
- logical sector size of blockdevice. For example, 512, 4096 etc. Provided 512 in the above example snippet. This value can be changed as per the logical sector size of the device.
- deviceType
- Type of the device. This can be obtained from
lsblkoutput. eg: lvm, crypt, nbd, md etc
- Type of the device. This can be obtained from
- links
- This field should be filled for by-id and by-path. These details can be obtained from worker node by running the following command
udevadm info -q property -n <device_path>
- This field should be filled for by-id and by-path. These details can be obtained from worker node by running the following command
- nodeName
- Name of the Node where the blockdevice is attached. The output of
kubectl get nodescan be used to obtain this value.
- Name of the Node where the blockdevice is attached. The output of
- path
- The value should be like
/dev/dm-0or/dev/md0.
- The value should be like
- name
Apply the modified YAML file to create the blockdevice CR for the provided device path.
kubectl apply -f <blockdevice-cr.yaml> -n <openebs_namespace>Note: The blockdevice CR should be created on the same namespace where openebs is installed.
Repeat the same steps for each unsupported device and create blockdevice CR for the devices.
Verify if the blockdevice is created by running the following
kubectl get blockdevice -n openebscommand.
Note: If you are creating a block device CR for an unsupported device, then you must add the corresponding disk under exclude filter so that NDM will not select the particular disk for BD creation. See here for customizing the exclude filter in NDM configuration.
See Also:
Understanding Node Disk Manager
Feedback
Was this page helpful?
Thanks for the feedback. Open an issue in the GitHub repo if you want to report a problem or suggest an improvement. Engage and get additional help on https://kubernetes.slack.com/messages/openebs/.