5. IDXD DMA Device Driver
The idxd
dmadev driver provides a poll-mode driver (PMD) for Intel®
Data Streaming Accelerator (Intel DSA).
This PMD can be used in conjunction with Intel® DSA devices to offload
data operations, such as data copies, to hardware, freeing up CPU cycles for
other tasks.
5.1. Hardware Requirements
The dpdk-devbind.py
script, included with DPDK, can be used to show the
presence of supported hardware. Running dpdk-devbind.py --status-dev dma
will show all the DMA devices on the system, including IDXD supported devices.
Intel® DSA devices, are currently (at time of writing) appearing
as devices with type “0b25”, due to the absence of pci-id database entries for
them at this point.
5.2. Compilation
For builds using meson
and ninja
, the driver will be built when the
target platform is x86-based. No additional compilation steps are necessary.
5.3. Device Setup
Intel® DSA devices can use the IDXD kernel driver or DPDK-supported drivers,
such as vfio-pci
. Both are supported by the IDXD PMD.
Note
To use Intel® DSA devices in DPDK multi-process applications, the devices should be bound to the vfio-pci driver. Multi-process is not supported when using the kernel IDXD driver.
5.3.1. Intel® DSA devices using IDXD kernel driver
To use an Intel® DSA device bound to the IDXD kernel driver, the device must first be configured. The accel-config utility library can be used for configuration.
Note
The device configuration can also be done by directly interacting with the sysfs nodes.
An example of how this may be done can be seen in the script dpdk_idxd_cfg.py
included in the driver source directory.
There are some mandatory configuration steps before being able to use a device with an application. The internal engines, which do the copies or other operations, and the work-queues, which are used by applications to assign work to the device, need to be assigned to groups, and the various other configuration options, such as priority or queue depth, need to be set for each queue.
To assign an engine to a group:
$ accel-config config-engine dsa0/engine0.0 --group-id=0
To assign work queues to groups for passing descriptors to the engines a similar accel-config command can be used. However, the work queues also need to be configured depending on the use case. Some configuration options include:
- mode (Dedicated/Shared): Indicates whether a WQ may accept jobs from multiple queues simultaneously.
- priority: WQ priority between 1 and 15. Larger value means higher priority.
- wq-size: the size of the WQ. Sum of all WQ sizes must be less that the total-size defined by the device.
- type: WQ type (kernel/mdev/user). Determines how the device is presented.
- name: identifier given to the WQ.
Example configuration for a work queue:
$ accel-config config-wq dsa0/wq0.0 --group-id=0 \
--mode=dedicated --priority=10 --wq-size=8 \
--max-batch-size=512 --type=user --name=dpdk_app1
Once the devices have been configured, they need to be enabled:
$ accel-config enable-device dsa0
$ accel-config enable-wq dsa0/wq0.0
Check the device configuration:
$ accel-config list
Every Intel® DSA instance supports multiple queues and each should be similarly configured. As a further example, the following set of commands will configure and enable 4 queues on instance 0, giving each an equal share of resources:
# configure 4 groups, each with one engine
accel-config config-engine dsa0/engine0.0 --group-id=0
accel-config config-engine dsa0/engine0.1 --group-id=1
accel-config config-engine dsa0/engine0.2 --group-id=2
accel-config config-engine dsa0/engine0.3 --group-id=3
# configure 4 queues, putting each in a different group, so each
# is backed by a single engine
accel-config config-wq dsa0/wq0.0 --group-id=0 --type=user --wq-size=32 \
--priority=10 --max-batch-size=1024 --mode=dedicated --name=dpdk_app1
accel-config config-wq dsa0/wq0.1 --group-id=1 --type=user --wq-size=32 \
--priority=10 --max-batch-size=1024 --mode=dedicated --name=dpdk_app1
accel-config config-wq dsa0/wq0.2 --group-id=2 --type=user --wq-size=32 \
--priority=10 --max-batch-size=1024 --mode=dedicated --name=dpdk_app1
accel-config config-wq dsa0/wq0.3 --group-id=3 --type=user --wq-size=32 \
--priority=10 --max-batch-size=1024 --mode=dedicated --name=dpdk_app1
# enable device and queues
accel-config enable-device dsa0
accel-config enable-wq dsa0/wq0.0 dsa0/wq0.1 dsa0/wq0.2 dsa0/wq0.3
5.3.2. Devices using VFIO/UIO drivers
The HW devices to be used will need to be bound to a user-space IO driver for use.
The dpdk-devbind.py
script can be used to view the state of the devices
and to bind them to a suitable DPDK-supported driver, such as vfio-pci
.
For example:
$ dpdk-devbind.py -b vfio-pci 6a:01.0
Note
Since each individual queue on the HW device is its own separate dmadev instance,
the internal DMA device name includes the HW queue ID as a suffix on the PCI address.
The above device when used by a DPDK application will be accessible via dmadevs with names:
0000:6a:01.0-q0
, 00006a:01.0-q1
, etc.
5.3.3. Device Probing and Initialization
For devices bound to a suitable DPDK-supported VFIO/UIO driver, the HW devices will be found as part of the device scan done at application initialization time without the need to pass parameters to the application.
For Intel® DSA devices, DPDK will automatically configure the device with the
maximum number of workqueues available on it, partitioning all resources equally
among the queues.
If fewer workqueues are required, then the max_queues
parameter may be passed to
the device driver on the EAL commandline, via the allowlist
or -a
flag e.g.:
$ dpdk-test -a <b:d:f>,max_queues=4
For devices bound to the IDXD kernel driver,
the DPDK IDXD driver will automatically perform a scan for available workqueues
to use. Any workqueues found listed in /dev/dsa
on the system will be checked
in /sys
, and any which have dpdk_
prefix in their name will be automatically
probed by the driver to make them available to the application.
Alternatively, to support use by multiple DPDK processes simultaneously,
the value used as the DPDK --file-prefix
parameter may be used as a workqueue
name prefix, instead of dpdk_
, allowing each DPDK application instance to only
use a subset of configured queues.
Additionally, the -a (allowlist) or -b (blocklist) commandline parameters
are also available to further restrict the device list that will be used.
If the -a option is used, then any device that passes the dpdk_
or --file-prefix
prefix condition must also be present in the allow list.
Similarly, when the block list is used,
any device that passes the prefix condition must not be in the block list.
For example, to only use wq0.3
, assuming the name prefix condition is met:
$ dpdk-test -a wq0.3
Once probed successfully, irrespective of kernel driver, the device will appear as a dmadev
,
that is a “DMA device type” inside DPDK, and can be accessed using APIs from the
rte_dmadev
library.
5.4. Using IDXD DMAdev Devices
To use the devices from an application, the dmadev API can be used.
5.4.1. Device Configuration
IDXD configuration requirements:
ring_size
must be a power of two, between 64 and 4096.- Only one
vchan
is supported per device (work queue). - IDXD devices do not support silent mode.
- The transfer direction must be set to
RTE_DMA_DIR_MEM_TO_MEM
to copy from memory to memory.
Once configured, the device can then be made ready for use by calling the
rte_dma_start()
API.
5.4.2. Performing Data Copies
Refer to the Enqueue / Dequeue APIs section of the dmadev library documentation for details on operation enqueue, submission and completion API usage.
It is expected that, for efficiency reasons, a burst of operations will be enqueued to the
device via multiple enqueue calls between calls to the rte_dma_submit()
function.
When gathering completions, rte_dma_completed()
should be used, up until the point an error
occurs in an operation. If an error was encountered, rte_dma_completed_status()
must be used
to kick the device off to continue processing operations and also to gather the status of each
individual operations which is filled in to the status
array provided as parameter by the
application.
The following status codes are supported by IDXD:
RTE_DMA_STATUS_SUCCESSFUL
: The operation was successful.RTE_DMA_STATUS_INVALID_OPCODE
: The operation failed due to an invalid operation code.RTE_DMA_STATUS_INVALID_LENGTH
: The operation failed due to an invalid data length.RTE_DMA_STATUS_NOT_ATTEMPTED
: The operation was not attempted.RTE_DMA_STATUS_ERROR_UNKNOWN
: The operation failed due to an unspecified error.
The following code shows how to retrieve the number of successfully completed
copies within a burst and then using rte_dma_completed_status()
to check
which operation failed and kick off the device to continue processing operations:
enum rte_dma_status_code status[COMP_BURST_SZ];
uint16_t count, idx, status_count;
bool error = 0;
count = rte_dma_completed(dev_id, vchan, COMP_BURST_SZ, &idx, &error);
if (error){
status_count = rte_dma_completed_status(dev_id, vchan, COMP_BURST_SZ, &idx, status);
}