13. dpdk-test-eventdev Application

The dpdk-test-eventdev tool is a Data Plane Development Kit (DPDK) application that allows exercising various eventdev use cases. This application has a generic framework to add new eventdev based test cases to verify functionality and measure the performance parameters of DPDK eventdev devices.

13.1. Running the Application

The application has a number of command line options:

dpdk-test-eventdev [EAL Options] -- [application options]

13.1.1. EAL Options

The following are the EAL command-line options that can be used in conjunction with the dpdk-test-eventdev application. See the DPDK Getting Started Guides for more information on these options.

  • -c <COREMASK> or -l <CORELIST>

    Set the hexadecimal bitmask of the cores to run on. The corelist is a list of cores to use.

  • --vdev <driver><id>

    Add a virtual eventdev device.

13.1.2. Application Options

The following are the application command-line options:

  • --verbose

    Set verbose level. Default is 1. Value > 1 displays more details.

  • --dev <n>

    Set the device id of the event device.

  • --test <name>

    Set test name, where name is one of the following:

    order_queue
    order_atq
    perf_queue
    perf_atq
    pipeline_atq
    pipeline_queue
    
  • --socket_id <n>

    Set the socket id of the application resources.

  • --pool-sz <n>

    Set the number of mbufs to be allocated from the mempool.

  • --plcores <CORELIST>

    Set the list of cores to be used as producers.

  • --wlcores <CORELIST>

    Set the list of cores to be used as workers.

  • --stlist <type_list>

    Set the scheduled type of each stage where type_list size determines the number of stages used in the test application. Each type_list member can be one of the following:

    P or p : Parallel schedule type
    O or o : Ordered schedule type
    A or a : Atomic schedule type
    

    Application expects the type_list in comma separated form (i.e. --stlist o,a,a,a)

  • --nb_flows <n>

    Set the number of flows to produce.

  • --nb_pkts <n>

    Set the number of packets to produce. 0 implies no limit.

  • --worker_deq_depth <n>

    Set the dequeue depth of the worker.

  • --fwd_latency

    Perform forward latency measurement.

  • --queue_priority

    Enable queue priority.

  • --prod_type_ethdev

    Use ethernet device as producer.

  • --prod_type_timerdev

    Use event timer adapter as producer.

  • --prod_type_timerdev_burst

    Use burst mode event timer adapter as producer.

  • --prod_type_cryptodev

    Use crypto device as producer.

  • --timer_tick_nsec

    Used to dictate number of nano seconds between bucket traversal of the event timer adapter. Refer rte_event_timer_adapter_conf.

  • --max_tmo_nsec

    Used to configure event timer adapter max arm timeout in nano seconds.

  • --expiry_nsec

    Dictate the number of nano seconds after which the event timer expires.

  • --nb_timers

    Number of event timers each producer core will generate.

  • --nb_timer_adptrs

    Number of event timer adapters to be used. Each adapter is used in round robin manner by the producer cores.

  • --deq_tmo_nsec

    Global dequeue timeout for all the event ports if the provided dequeue timeout is out of the supported range of event device it will be adjusted to the highest/lowest supported dequeue timeout supported.

  • --crypto_adptr_mode

    Set crypto adapter mode. Use 0 for OP_NEW (default) and 1 for OP_FORWARD mode.

  • --crypto_op_type

    Set crypto operation type. Use 0 for symmetric crypto ops (default) and 1 for asymmetric crypto ops.

  • --crypto_cipher_alg

    Cipher algorithm to be used. Default algorithm is NULL.

  • --crypto_cipher_key

    Key for the cipher algorithm selected.

  • --crypto_cipher_iv_sz

    IV size for the cipher algorithm

  • --mbuf_sz

    Set packet mbuf size. Can be used to configure Jumbo Frames. Only applicable for pipeline_atq and pipeline_queue tests.

  • --max_pkt_sz

    Set max packet mbuf size. Can be used to configure Rx/Tx scatter gather. Only applicable for pipeline_atq and pipeline_queue tests.

  • --prod_enq_burst_sz

    Set producer enqueue burst size. Can be used to configure the number of events the producer(s) will enqueue as a burst to the event device. Only applicable for perf_queue and perf_atq test in combination with CPU (default) or crypto device (--prod_type_cryptodev) producers.

  • --nb_eth_queues

    Configure multiple Rx queues per each ethernet port. Only applicable for pipeline_atq and pipeline_queue tests.

  • --enable_vector

    Enable event vector for Rx/Tx/crypto adapters. Only applicable for pipeline_* and perf_* tests.

  • --vector_size

    Vector size to configure for the Rx/crypto adapter. Only applicable for pipeline_* and perf_* tests.

  • --vector_tmo_ns

    Vector timeout nanoseconds to be configured for the Rx/crypto adapter. Only applicable for pipeline_* and perf_* tests.

  • --per_port_pool

    Configure unique mempool per ethernet device, the size of each pool is equal to pool_sz. Only applicable for pipeline_atq and pipeline_queue tests.

  • --tx_first

    Transmit given number of packets across all the ethernet device that are enabled in the test. Only applicable for pipeline_atq and pipeline_queue tests.

  • --tx_pkt_sz

    Packet size to use for –tx_first. Only applicable for pipeline_atq and pipeline_queue tests.

13.2. Eventdev Tests

13.2.1. ORDER_QUEUE Test

This is a functional test case that aims at testing the following:

  1. Verify the ingress order maintenance.
  2. Verify the exclusive(atomic) access to given atomic flow per eventdev port.
Table 13.95 Order queue test eventdev configuration.
# Items Value Comments
1 nb_queues 2 q0(ordered), q1(atomic)
2 nb_producers 1  
3 nb_workers >= 1  
4 nb_ports nb_workers + 1 Workers use port 0 to port n-1. Producer uses port n
../_images/eventdev_order_queue_test.svg

Fig. 13.3 order queue test operation.

The order queue test configures the eventdev with two queues and an event producer to inject the events to q0(ordered) queue. Both q0(ordered) and q1(atomic) are linked to all the workers.

The event producer maintains a sequence number per flow and injects the events to the ordered queue. The worker receives the events from ordered queue and forwards to atomic queue. Since the events from an ordered queue can be processed in parallel on the different workers, the ingress order of events might have changed on the downstream atomic queue enqueue. On enqueue to the atomic queue, the eventdev PMD reorders the event to the original ingress order(i.e producer ingress order).

When the event is dequeued from the atomic queue by the worker, this test verifies the expected sequence number of associated event per flow by comparing the free running expected sequence number per flow.

13.2.1.1. Application options

Supported application command line options are following:

--verbose
--dev
--test
--socket_id
--pool_sz
--plcores
--wlcores
--nb_flows
--nb_pkts
--worker_deq_depth
--deq_tmo_nsec

13.2.1.2. Example

Example command to run order queue test:

sudo <build_dir>/app/dpdk-test-eventdev --vdev=event_sw0 -- \
             --test=order_queue --plcores 1 --wlcores 2,3

13.2.2. ORDER_ATQ Test

This test verifies the same aspects of order_queue test, the difference is the number of queues used, this test operates on a single all types queue(atq) instead of two different queues for ordered and atomic.

Table 13.96 Order all types queue test eventdev configuration.
# Items Value Comments
1 nb_queues 1 q0(all types queue)
2 nb_producers 1  
3 nb_workers >= 1  
4 nb_ports nb_workers + 1 Workers use port 0 to port n-1.Producer uses port n.
../_images/eventdev_order_atq_test.svg

Fig. 13.4 order all types queue test operation.

13.2.2.1. Application options

Supported application command line options are following:

--verbose
--dev
--test
--socket_id
--pool_sz
--plcores
--wlcores
--nb_flows
--nb_pkts
--worker_deq_depth
--deq_tmo_nsec

13.2.2.2. Example

Example command to run order all types queue test:

sudo <build_dir>/app/dpdk-test-eventdev --vdev=event_octeontx -- \
                     --test=order_atq --plcores 1 --wlcores 2,3

13.2.3. PERF_QUEUE Test

This is a performance test case that aims at testing the following:

  1. Measure the number of events can be processed in a second.
  2. Measure the latency to forward an event.
Table 13.97 Perf queue test eventdev configuration.
# Items Value Comments
1 nb_queues nb_producers * nb_stages Queues will be configured based on the user requested sched type list(–stlist)
2 nb_producers >= 1 Selected through –plcores command line argument.
3 nb_workers >= 1 Selected through –wlcores command line argument
4 nb_ports nb_workers + nb_producers Workers use port 0 to port n-1. Producers use port n to port p
../_images/eventdev_perf_queue_test.svg

Fig. 13.5 perf queue test operation.

The perf queue test configures the eventdev with Q queues and P ports, where Q and P is a function of the number of workers, the number of producers and number of stages as mentioned in Table 13.97.

The user can choose the number of workers, the number of producers and number of stages through the --wlcores, --plcores and the --stlist application command line arguments respectively.

The producer(s) injects the events to eventdev based on the first stage sched type list requested by the user through --stlist command line argument. It can inject a burst of events using --prod_enq_burst_sz command line argument.

Based on the number of stages to process(selected through --stlist), The application forwards the event to next upstream queue and terminates when it reaches the last stage in the pipeline. On event termination, application increments the number events processed and print periodically in one second to get the number of events processed in one second.

When --fwd_latency command line option selected, the application inserts the timestamp in the event on the first stage and then on termination, it updates the number of cycles to forward a packet. The application uses this value to compute the average latency to a forward packet.

When --prod_type_ethdev command line option is selected, the application uses the probed ethernet devices as producers by configuring them as Rx adapters instead of using synthetic producers.

13.2.3.1. Application options

Supported application command line options are following:

--verbose
--dev
--test
--socket_id
--pool_sz
--plcores
--wlcores
--stlist
--nb_flows
--nb_pkts
--worker_deq_depth
--fwd_latency
--queue_priority
--prod_type_ethdev
--prod_type_timerdev_burst
--prod_type_timerdev
--prod_type_cryptodev
--prod_enq_burst_sz
--timer_tick_nsec
--max_tmo_nsec
--expiry_nsec
--nb_timers
--nb_timer_adptrs
--deq_tmo_nsec
--crypto_adptr_mode

13.2.3.2. Example

Example command to run perf queue test:

sudo <build_dir>/app/dpdk-test-eventdev -c 0xf -s 0x1 --vdev=event_sw0 -- \
     --test=perf_queue --plcores=2 --wlcore=3 --stlist=p --nb_pkts=0

Example command to run perf queue test with producer enqueuing a burst of events:

sudo <build_dir>/app/dpdk-test-eventdev -c 0xf -s 0x1 --vdev=event_sw0 -- \
     --test=perf_queue --plcores=2 --wlcore=3 --stlist=p --nb_pkts=0 \
     --prod_enq_burst_sz=32

Example command to run perf queue test with ethernet ports:

sudo build/app/dpdk-test-eventdev --vdev=event_sw0 -- \
     --test=perf_queue --plcores=2 --wlcore=3 --stlist=p --prod_type_ethdev

Example command to run perf queue test with event timer adapter:

sudo  <build_dir>/app/dpdk-test-eventdev --vdev="event_octeontx" -- \
             --wlcores 4 --plcores 12 --test perf_queue --stlist=a \
             --prod_type_timerdev --fwd_latency

13.2.4. PERF_ATQ Test

This is a performance test case that aims at testing the following with all types queue eventdev scheme.

  1. Measure the number of events can be processed in a second.
  2. Measure the latency to forward an event.
Table 13.98 Perf all types queue test eventdev configuration.
# Items Value Comments
1 nb_queues nb_producers Queues will be configured based on the user requested sched type list(–stlist)
2 nb_producers >= 1 Selected through –plcores command line argument.
3 nb_workers >= 1 Selected through –wlcores command line argument
4 nb_ports nb_workers + nb_producers Workers use port 0 to port n-1. Producers use port n to port p
../_images/eventdev_perf_atq_test.svg

Fig. 13.6 perf all types queue test operation.

The all types queues(atq) perf test configures the eventdev with Q queues and P ports, where Q and P is a function of the number of workers and number of producers as mentioned in Table 13.98.

The atq queue test functions as same as perf_queue test. The difference is, It uses, all type queue scheme instead of separate queues for each stage and thus reduces the number of queues required to realize the use case and enables flow pinning as the event does not move to the next queue.

13.2.4.1. Application options

Supported application command line options are following:

--verbose
--dev
--test
--socket_id
--pool_sz
--plcores
--wlcores
--stlist
--nb_flows
--nb_pkts
--worker_deq_depth
--fwd_latency
--prod_type_ethdev
--prod_type_timerdev_burst
--prod_type_timerdev
--prod_type_cryptodev
--timer_tick_nsec
--max_tmo_nsec
--expiry_nsec
--nb_timers
--nb_timer_adptrs
--deq_tmo_nsec
--crypto_adptr_mode

13.2.4.2. Example

Example command to run perf all types queue test:

sudo <build_dir>/app/dpdk-test-eventdev --vdev=event_octeontx -- \
             --test=perf_atq --plcores=2 --wlcore=3 --stlist=p --nb_pkts=0

Example command to run perf all types queue test with event timer adapter:

sudo  <build_dir>/app/dpdk-test-eventdev --vdev="event_octeontx" -- \
             --wlcores 4 --plcores 12 --test perf_atq --verbose 20 \
             --stlist=a --prod_type_timerdev --fwd_latency

13.2.5. PIPELINE_QUEUE Test

This is a pipeline test case that aims at testing the following:

  1. Measure the end-to-end performance of an event dev with a ethernet dev.
  2. Maintain packet ordering from Rx to Tx.
Table 13.99 Pipeline queue test eventdev configuration.
# Items Value Comments
1 nb_queues (nb_producers * nb_stages) + nb_producers Queues will be configured based on the user requested sched type list(–stlist) At the last stage of the schedule list the event is enqueued onto per port unique queue which is then Transmitted.
2 nb_producers >= 1 Producers will be configured based on the number of detected ethernet devices. Each ethdev will be configured as an Rx adapter.
3 nb_workers >= 1 Selected through –wlcores command line argument
4 nb_ports nb_workers + (nb_produces * 2) Workers use port 0 to port n. Producers use port n+1 to port n+m, depending on the Rx adapter capability. Consumers use port n+m+1 to port n+o depending on the Tx adapter capability.
../_images/eventdev_pipeline_queue_test_generic.svg
../_images/eventdev_pipeline_queue_test_internal_port.svg

Fig. 13.8 pipeline queue test operation.

The pipeline queue test configures the eventdev with Q queues and P ports, where Q and P is a function of the number of workers, the number of producers and number of stages as mentioned in Table 13.99.

The user can choose the number of workers and number of stages through the --wlcores and the --stlist application command line arguments respectively.

The number of producers depends on the number of ethernet devices detected and each ethernet device is configured as a event_eth_rx_adapter that acts as a producer.

The producer(s) injects the events to eventdev based the first stage sched type list requested by the user through --stlist the command line argument.

Based on the number of stages to process(selected through --stlist), The application forwards the event to next upstream queue and when it reaches the last stage in the pipeline if the event type is atomic it is enqueued onto ethdev Tx queue else to maintain ordering the event type is set to atomic and enqueued onto the last stage queue.

If the ethdev and eventdev pair have RTE_EVENT_ETH_TX_ADAPTER_CAP_INTERNAL_PORT capability then the worker cores enqueue the packets to the eventdev directly using rte_event_eth_tx_adapter_enqueue else the worker cores enqueue the packet onto the SINGLE_LINK_QUEUE that is managed by the Tx adapter. The Tx adapter dequeues the packet and transmits it.

On packet Tx, application increments the number events processed and print periodically in one second to get the number of events processed in one second.

13.2.5.1. Application options

Supported application command line options are following:

--verbose
--dev
--test
--socket_id
--pool_sz
--wlcores
--stlist
--worker_deq_depth
--prod_type_ethdev
--deq_tmo_nsec
--nb_eth_queues
--enable_vector
--vector_size
--vector_tmo_ns
--per_port_pool
--tx_first
--tx_pkt_sz

Note

  • The --prod_type_ethdev is mandatory for running this test.

13.2.5.2. Example

Example command to run pipeline queue test:

sudo <build_dir>/app/dpdk-test-eventdev -c 0xf -s 0x8 --vdev=event_sw0 -- \
    --test=pipeline_queue --wlcore=1 --prod_type_ethdev --stlist=a

Example command to run pipeline atq test with vector events:

sudo <build_dir>/app/dpdk-test-eventdev -c 0xf -s 0x8 --vdev=event_sw0 -- \
    --test=pipeline_queue --wlcore=1 --prod_type_ethdev --stlist=a \
    --enable_vector  --vector_size 512

13.2.6. PIPELINE_ATQ Test

This is a pipeline test case that aims at testing the following with all types queue eventdev scheme.

  1. Measure the end-to-end performance of an event dev with a ethernet dev.
  2. Maintain packet ordering from Rx to Tx.
Table 13.100 Pipeline atq test eventdev configuration.
# Items Value Comments
1 nb_queues nb_producers + x Queues will be configured based on the user requested sched type list(–stlist) where x = nb_producers in generic pipeline and 0 if all the ethdev being used have Internal port capability
2 nb_producers >= 1 Producers will be configured based on the number of detected ethernet devices. Each ethdev will be configured as an Rx adapter.
3 nb_workers >= 1 Selected through –wlcores command line argument
4 nb_ports nb_workers + nb_producers + x Workers use port 0 to port n. Producers use port n+1 to port n+m, depending on the Rx adapter capability. x = nb_producers in generic pipeline and 0 if all the ethdev being used have Internal port capability. Consumers may use port n+m+1 to port n+o depending on the Tx adapter capability.
../_images/eventdev_pipeline_atq_test_generic.svg
../_images/eventdev_pipeline_atq_test_internal_port.svg

Fig. 13.10 pipeline atq test operation.

The pipeline atq test configures the eventdev with Q queues and P ports, where Q and P is a function of the number of workers, the number of producers and number of stages as mentioned in Table 13.100.

The atq queue test functions as same as pipeline_queue test. The difference is, It uses, all type queue scheme instead of separate queues for each stage and thus reduces the number of queues required to realize the use case.

13.2.6.1. Application options

Supported application command line options are following:

--verbose
--dev
--test
--socket_id
--pool_sz
--wlcores
--stlist
--worker_deq_depth
--prod_type_ethdev
--deq_tmo_nsec
--nb_eth_queues
--enable_vector
--vector_size
--vector_tmo_ns
--per_port_pool
--tx_first
--tx_pkt_sz

Note

  • The --prod_type_ethdev is mandatory for running this test.

13.2.6.2. Example

Example command to run pipeline atq test:

sudo <build_dir>/app/dpdk-test-eventdev -c 0xf -s 0x8 --vdev=event_sw0 -- \
    --test=pipeline_atq --wlcore=1 --prod_type_ethdev --stlist=a

Example command to run pipeline atq test with vector events:

sudo <build_dir>/app/dpdk-test-eventdev -c 0xf -s 0x8 --vdev=event_sw0 -- \
    --test=pipeline_atq --wlcore=1 --prod_type_ethdev --stlist=a \
    --enable_vector  --vector_size 512