5. IDXD DMA Device Driver

The idxd dmadev driver provides a poll-mode driver (PMD) for Intel® Data Streaming Accelerator (Intel DSA). This PMD can be used in conjunction with Intel® DSA devices to offload data operations, such as data copies, to hardware, freeing up CPU cycles for other tasks.

5.1. Hardware Requirements

The dpdk-devbind.py script, included with DPDK, can be used to show the presence of supported hardware. Running dpdk-devbind.py --status-dev dma will show all the DMA devices on the system, including IDXD supported devices. Intel® DSA devices, are currently (at time of writing) appearing as devices with type “0b25”, due to the absence of pci-id database entries for them at this point.

5.2. Compilation

For builds using meson and ninja, the driver will be built when the target platform is x86-based. No additional compilation steps are necessary.

5.3. Device Setup

Intel® DSA devices can use the IDXD kernel driver or DPDK-supported drivers, such as vfio-pci. Both are supported by the IDXD PMD.

Note

To use Intel® DSA devices in DPDK multi-process applications, the devices should be bound to the vfio-pci driver. Multi-process is not supported when using the kernel IDXD driver.

5.3.1. Intel® DSA devices using IDXD kernel driver

To use an Intel® DSA device bound to the IDXD kernel driver, the device must first be configured. The accel-config utility library can be used for configuration.

Note

The device configuration can also be done by directly interacting with the sysfs nodes. An example of how this may be done can be seen in the script dpdk_idxd_cfg.py included in the driver source directory.

There are some mandatory configuration steps before being able to use a device with an application. The internal engines, which do the copies or other operations, and the work-queues, which are used by applications to assign work to the device, need to be assigned to groups, and the various other configuration options, such as priority or queue depth, need to be set for each queue.

To assign an engine to a group:

$ accel-config config-engine dsa0/engine0.0 --group-id=0

To assign work queues to groups for passing descriptors to the engines a similar accel-config command can be used. However, the work queues also need to be configured depending on the use case. Some configuration options include:

  • mode (Dedicated/Shared): Indicates whether a WQ may accept jobs from multiple queues simultaneously.

  • priority: WQ priority between 1 and 15. Larger value means higher priority.

  • wq-size: the size of the WQ. Sum of all WQ sizes must be less that the total-size defined by the device.

  • type: WQ type (kernel/mdev/user). Determines how the device is presented.

  • name: identifier given to the WQ.

Example configuration for a work queue:

$ accel-config config-wq dsa0/wq0.0 --group-id=0 \
   --mode=dedicated --priority=10 --wq-size=8 \
   --max-batch-size=512 --type=user --name=dpdk_app1

Once the devices have been configured, they need to be enabled:

$ accel-config enable-device dsa0
$ accel-config enable-wq dsa0/wq0.0

Check the device configuration:

$ accel-config list

Every Intel® DSA instance supports multiple queues and each should be similarly configured. As a further example, the following set of commands will configure and enable 4 queues on instance 0, giving each an equal share of resources:

# configure 4 groups, each with one engine
accel-config config-engine dsa0/engine0.0 --group-id=0
accel-config config-engine dsa0/engine0.1 --group-id=1
accel-config config-engine dsa0/engine0.2 --group-id=2
accel-config config-engine dsa0/engine0.3 --group-id=3

# configure 4 queues, putting each in a different group, so each
# is backed by a single engine
accel-config config-wq dsa0/wq0.0 --group-id=0 --type=user --wq-size=32 \
    --priority=10 --max-batch-size=1024 --mode=dedicated --name=dpdk_app1
accel-config config-wq dsa0/wq0.1 --group-id=1 --type=user --wq-size=32 \
    --priority=10 --max-batch-size=1024 --mode=dedicated --name=dpdk_app1
accel-config config-wq dsa0/wq0.2 --group-id=2 --type=user --wq-size=32 \
    --priority=10 --max-batch-size=1024 --mode=dedicated --name=dpdk_app1
accel-config config-wq dsa0/wq0.3 --group-id=3 --type=user --wq-size=32 \
    --priority=10 --max-batch-size=1024 --mode=dedicated --name=dpdk_app1

# enable device and queues
accel-config enable-device dsa0
accel-config enable-wq dsa0/wq0.0 dsa0/wq0.1 dsa0/wq0.2 dsa0/wq0.3

5.3.2. Devices using VFIO/UIO drivers

The HW devices to be used will need to be bound to a user-space IO driver for use. The dpdk-devbind.py script can be used to view the state of the devices and to bind them to a suitable DPDK-supported driver, such as vfio-pci. For example:

$ dpdk-devbind.py -b vfio-pci 6a:01.0

Note

Since each individual queue on the HW device is its own separate dmadev instance, the internal DMA device name includes the HW queue ID as a suffix on the PCI address. The above device when used by a DPDK application will be accessible via dmadevs with names: 0000:6a:01.0-q0, 00006a:01.0-q1, etc.

5.3.3. Device Probing and Initialization

For devices bound to a suitable DPDK-supported VFIO/UIO driver, the HW devices will be found as part of the device scan done at application initialization time without the need to pass parameters to the application.

For Intel® DSA devices, DPDK will automatically configure the device with the maximum number of workqueues available on it, partitioning all resources equally among the queues. If fewer workqueues are required, then the max_queues parameter may be passed to the device driver on the EAL commandline, via the allowlist or -a flag e.g.:

$ dpdk-test -a <b:d:f>,max_queues=4

For devices bound to the IDXD kernel driver, the DPDK IDXD driver will automatically perform a scan for available workqueues to use. Any workqueues found listed in /dev/dsa on the system will be checked in /sys, and any which have dpdk_ prefix in their name will be automatically probed by the driver to make them available to the application. Alternatively, to support use by multiple DPDK processes simultaneously, the value used as the DPDK --file-prefix parameter may be used as a workqueue name prefix, instead of dpdk_, allowing each DPDK application instance to only use a subset of configured queues.

Additionally, the -a (allowlist) or -b (blocklist) commandline parameters are also available to further restrict the device list that will be used. If the -a option is used, then any device that passes the dpdk_ or --file-prefix prefix condition must also be present in the allow list. Similarly, when the block list is used, any device that passes the prefix condition must not be in the block list. For example, to only use wq0.3, assuming the name prefix condition is met:

$ dpdk-test -a wq0.3

Once probed successfully, irrespective of kernel driver, the device will appear as a dmadev, that is a “DMA device type” inside DPDK, and can be accessed using APIs from the rte_dmadev library.

5.4. Using IDXD DMAdev Devices

To use the devices from an application, the dmadev API can be used.

5.4.1. Device Configuration

IDXD configuration requirements:

  • ring_size must be a power of two, between 64 and 4096.

  • Only one vchan is supported per device (work queue).

  • IDXD devices do not support silent mode.

  • The transfer direction must be set to RTE_DMA_DIR_MEM_TO_MEM to copy from memory to memory.

Once configured, the device can then be made ready for use by calling the rte_dma_start() API.

5.4.2. Performing Data Copies

Refer to the Enqueue / Dequeue APIs section of the dmadev library documentation for details on operation enqueue, submission and completion API usage.

It is expected that, for efficiency reasons, a burst of operations will be enqueued to the device via multiple enqueue calls between calls to the rte_dma_submit() function.

When gathering completions, rte_dma_completed() should be used, up until the point an error occurs in an operation. If an error was encountered, rte_dma_completed_status() must be used to kick the device off to continue processing operations and also to gather the status of each individual operations which is filled in to the status array provided as parameter by the application.

The following status codes are supported by IDXD:

  • RTE_DMA_STATUS_SUCCESSFUL: The operation was successful.

  • RTE_DMA_STATUS_INVALID_OPCODE: The operation failed due to an invalid operation code.

  • RTE_DMA_STATUS_INVALID_LENGTH: The operation failed due to an invalid data length.

  • RTE_DMA_STATUS_NOT_ATTEMPTED: The operation was not attempted.

  • RTE_DMA_STATUS_ERROR_UNKNOWN: The operation failed due to an unspecified error.

The following code shows how to retrieve the number of successfully completed copies within a burst and then using rte_dma_completed_status() to check which operation failed and kick off the device to continue processing operations:

enum rte_dma_status_code status[COMP_BURST_SZ];
uint16_t count, idx, status_count;
bool error = 0;

count = rte_dma_completed(dev_id, vchan, COMP_BURST_SZ, &idx, &error);

if (error){
   status_count = rte_dma_completed_status(dev_id, vchan, COMP_BURST_SZ, &idx, status);
}