Searched refs:nb_rx (Results 1 - 25 of 58) sorted by relevance

123

/deb_dpdk/app/test-pmd/
H A Diofwd.c80 uint16_t nb_rx; local
97 nb_rx = rte_eth_rx_burst(fs->rx_port, fs->rx_queue,
99 if (unlikely(nb_rx == 0))
101 fs->rx_packets += nb_rx;
104 fs->rx_burst_stats.pkt_burst_spread[nb_rx]++;
107 pkts_burst, nb_rx);
111 if (unlikely(nb_tx < nb_rx) && fs->retry_enabled) {
113 while (nb_tx < nb_rx && retry++ < burst_tx_retry_num) {
116 &pkts_burst[nb_tx], nb_rx - nb_tx);
123 if (unlikely(nb_tx < nb_rx)) {
[all...]
H A Dmacfwd.c84 uint16_t nb_rx; local
101 nb_rx = rte_eth_rx_burst(fs->rx_port, fs->rx_queue, pkts_burst,
103 if (unlikely(nb_rx == 0))
107 fs->rx_burst_stats.pkt_burst_spread[nb_rx]++;
109 fs->rx_packets += nb_rx;
115 for (i = 0; i < nb_rx; i++) {
116 if (likely(i < nb_rx - 1))
131 nb_tx = rte_eth_tx_burst(fs->tx_port, fs->tx_queue, pkts_burst, nb_rx);
135 if (unlikely(nb_tx < nb_rx) && fs->retry_enabled) {
137 while (nb_tx < nb_rx
[all...]
H A Dmacswap.c83 uint16_t nb_rx; local
101 nb_rx = rte_eth_rx_burst(fs->rx_port, fs->rx_queue, pkts_burst,
103 if (unlikely(nb_rx == 0))
107 fs->rx_burst_stats.pkt_burst_spread[nb_rx]++;
109 fs->rx_packets += nb_rx;
115 for (i = 0; i < nb_rx; i++) {
116 if (likely(i < nb_rx - 1))
133 nb_tx = rte_eth_tx_burst(fs->tx_port, fs->tx_queue, pkts_burst, nb_rx);
137 if (unlikely(nb_tx < nb_rx) && fs->retry_enabled) {
139 while (nb_tx < nb_rx
[all...]
H A Dflowgen.c128 uint16_t nb_rx; local
145 nb_rx = rte_eth_rx_burst(fs->rx_port, fs->rx_queue, pkts_burst,
147 fs->rx_packets += nb_rx;
149 for (i = 0; i < nb_rx; i++)
212 if (unlikely(nb_tx < nb_rx) && fs->retry_enabled) {
214 while (nb_tx < nb_rx && retry++ < burst_tx_retry_num) {
217 &pkts_burst[nb_tx], nb_rx - nb_tx);
H A Drxonly.c92 uint16_t nb_rx; local
112 nb_rx = rte_eth_rx_burst(fs->rx_port, fs->rx_queue, pkts_burst,
114 if (unlikely(nb_rx == 0))
118 fs->rx_burst_stats.pkt_burst_spread[nb_rx]++;
120 fs->rx_packets += nb_rx;
129 (unsigned) nb_rx);
130 for (i = 0; i < nb_rx; i++) {
H A Dicmpecho.c315 uint16_t nb_rx; local
338 nb_rx = rte_eth_rx_burst(fs->rx_port, fs->rx_queue, pkts_burst,
340 if (unlikely(nb_rx == 0))
344 fs->rx_burst_stats.pkt_burst_spread[nb_rx]++;
346 fs->rx_packets += nb_rx;
348 for (i = 0; i < nb_rx; i++) {
349 if (likely(i < nb_rx - 1))
H A Dcsumonly.c653 uint16_t nb_rx; local
674 nb_rx = rte_eth_rx_burst(fs->rx_port, fs->rx_queue, pkts_burst,
676 if (unlikely(nb_rx == 0))
680 fs->rx_burst_stats.pkt_burst_spread[nb_rx]++;
682 fs->rx_packets += nb_rx;
692 for (i = 0; i < nb_rx; i++) {
693 if (likely(i < nb_rx - 1))
864 nb_tx = rte_eth_tx_burst(fs->tx_port, fs->tx_queue, pkts_burst, nb_rx);
868 if (unlikely(nb_tx < nb_rx) && fs->retry_enabled) {
870 while (nb_tx < nb_rx
[all...]
/deb_dpdk/app/test/
H A Dtest_pmd_perf.c378 unsigned i, portid, nb_rx, nb_tx; local
386 nb_rx = rte_eth_rx_burst((uint8_t) portid, 0,
388 if (unlikely(nb_rx == 0)) {
393 count += nb_rx;
394 nb_tx = rte_eth_tx_burst(portid, 0, pkts_burst, nb_rx);
395 if (unlikely(nb_tx < nb_rx)) {
396 drop += (nb_rx - nb_tx);
399 } while (++nb_tx < nb_rx);
416 unsigned i, portid, nb_rx, nb_tx; local
425 nb_rx
454 unsigned i, portid, nb_rx, nb_tx; local
497 unsigned i, portid, nb_rx = 0, nb_tx = 0; local
572 unsigned i, portid, nb_rx = 0; local
[all...]
H A Dtest_kni.c165 unsigned nb_rx, nb_tx, num, i; local
190 for (nb_rx = 0; nb_rx < PKT_BURST_SZ; nb_rx++) {
191 pkts_burst[nb_rx] = rte_pktmbuf_alloc(mp);
192 if (!pkts_burst[nb_rx])
197 nb_rx);
200 if (num < nb_rx) {
201 for (i = num; i < nb_rx; i++) {
/deb_dpdk/examples/l3fwd/
H A Dl3fwd_em.h114 l3fwd_em_no_opt_send_packets(int nb_rx, struct rte_mbuf **pkts_burst, argument
120 for (j = 0; j < PREFETCH_OFFSET && j < nb_rx; j++)
127 for (j = 0; j < (nb_rx - PREFETCH_OFFSET); j++) {
134 for (; j < nb_rx; j++)
H A Dl3fwd_em_sse.h101 l3fwd_em_send_packets(int nb_rx, struct rte_mbuf **pkts_burst, argument
107 for (j = 0; j < nb_rx; j++)
110 send_packets_multi(qconf, pkts_burst, dst_port, nb_rx);
H A Dl3fwd_lpm.h130 l3fwd_lpm_no_opt_send_packets(int nb_rx, struct rte_mbuf **pkts_burst, argument
136 for (j = 0; j < PREFETCH_OFFSET && j < nb_rx; j++)
140 for (j = 0; j < (nb_rx - PREFETCH_OFFSET); j++) {
147 for (; j < nb_rx; j++)
H A Dl3fwd_lpm.c119 int i, nb_rx; local
175 nb_rx = rte_eth_rx_burst(portid, queueid, pkts_burst,
177 if (nb_rx == 0)
181 l3fwd_lpm_send_packets(nb_rx, pkts_burst,
184 l3fwd_lpm_no_opt_send_packets(nb_rx, pkts_burst,
H A Dl3fwd_lpm_sse.h180 l3fwd_lpm_send_packets(int nb_rx, struct rte_mbuf **pkts_burst, argument
187 const int32_t k = RTE_ALIGN_FLOOR(nb_rx, FWDSTEP);
198 switch (nb_rx % FWDSTEP) {
210 send_packets_multi(qconf, pkts_burst, dst_port, nb_rx);
/deb_dpdk/examples/l2fwd-cat/
H A Dl2fwd-cat.c149 const uint16_t nb_rx = rte_eth_rx_burst(port, 0, local
152 if (unlikely(nb_rx == 0))
157 bufs, nb_rx);
160 if (unlikely(nb_tx < nb_rx)) {
162 for (buf = nb_tx; buf < nb_rx; buf++)
/deb_dpdk/examples/skeleton/
H A Dbasicfwd.c147 const uint16_t nb_rx = rte_eth_rx_burst(port, 0, local
150 if (unlikely(nb_rx == 0))
155 bufs, nb_rx);
158 if (unlikely(nb_tx < nb_rx)) {
160 for (buf = nb_tx; buf < nb_rx; buf++)
/deb_dpdk/examples/qos_sched/
H A Dapp_thread.c84 uint32_t i, nb_rx; local
96 nb_rx = rte_eth_rx_burst(conf->rx_port, conf->rx_queue, rx_mbufs,
99 if (likely(nb_rx != 0)) {
100 APP_STATS_ADD(conf->stat.nb_rx, nb_rx);
102 for(i = 0; i < nb_rx; i++) {
110 (void **)rx_mbufs, nb_rx) != 0)) {
111 for(i = 0; i < nb_rx; i++) {
227 APP_STATS_ADD(conf->stat.nb_rx, nb_pkt);
261 APP_STATS_ADD(conf->stat.nb_rx, nb_pk
[all...]
H A Dmain.c209 flow->rx_thread.stat.nb_rx,
212 flow->wt_thread.stat.nb_rx,
214 flow->wt_thread.stat.nb_rx - flow->wt_thread.stat.nb_drop);
/deb_dpdk/examples/rxtx_callbacks/
H A Dmain.c171 const uint16_t nb_rx = rte_eth_rx_burst(port, 0, local
173 if (unlikely(nb_rx == 0))
176 bufs, nb_rx);
177 if (unlikely(nb_tx < nb_rx)) {
180 for (buf = nb_tx; buf < nb_rx; buf++)
/deb_dpdk/drivers/net/virtio/
H A Dvirtio_rxtx.c753 uint16_t nb_used, num, nb_rx; local
775 nb_rx = 0;
817 rx_pkts[nb_rx++] = rxm;
819 rxvq->stats.bytes += rx_pkts[nb_rx - 1]->pkt_len;
823 rxvq->stats.packets += nb_rx;
852 return nb_rx;
864 uint16_t nb_used, num, nb_rx; local
883 nb_rx = 0;
895 if (nb_rx == nb_pkts)
933 rx_pkts[nb_rx]
[all...]
/deb_dpdk/examples/distributor/
H A Dmain.c222 const uint16_t nb_rx = rte_eth_rx_burst(port, 0, bufs, local
224 if (unlikely(nb_rx == 0)) {
229 app_stats.rx.rx_pkts += nb_rx;
231 rte_distributor_process(d, bufs, nb_rx);
331 const uint16_t nb_rx = rte_ring_dequeue_burst(in_r, local
333 app_stats.tx.dequeue_pkts += nb_rx;
336 if (unlikely(nb_rx == 0)) {
346 for (i = 0; i < nb_rx; i++) {
/deb_dpdk/lib/librte_port/
H A Drte_port_sched.c95 uint32_t nb_rx; local
97 nb_rx = rte_sched_port_dequeue(p->sched, pkts, n_pkts);
98 RTE_PORT_SCHED_READER_PKTS_IN_ADD(p, nb_rx);
100 return nb_rx;
H A Drte_port_ring.c112 uint32_t nb_rx; local
114 nb_rx = rte_ring_sc_dequeue_burst(p->ring, (void **) pkts, n_pkts);
115 RTE_PORT_RING_READER_STATS_PKTS_IN_ADD(p, nb_rx);
117 return nb_rx;
125 uint32_t nb_rx; local
127 nb_rx = rte_ring_mc_dequeue_burst(p->ring, (void **) pkts, n_pkts);
128 RTE_PORT_RING_READER_STATS_PKTS_IN_ADD(p, nb_rx);
130 return nb_rx;
/deb_dpdk/examples/qos_meter/
H A Dmain.c190 int i, nb_rx; local
202 nb_rx = rte_eth_rx_burst(port_rx, NIC_RX_QUEUE, pkts_rx, PKT_RX_BURST_MAX);
205 for (i = 0; i < nb_rx; i ++) {
/deb_dpdk/examples/kni/
H A Dmain.c243 unsigned nb_rx, num; local
254 nb_rx = rte_eth_rx_burst(port_id, 0, pkts_burst, PKT_BURST_SZ);
255 if (unlikely(nb_rx > PKT_BURST_SZ)) {
260 num = rte_kni_tx_burst(p->kni[i], pkts_burst, nb_rx);
264 if (unlikely(num < nb_rx)) {
266 kni_burst_free_mbufs(&pkts_burst[num], nb_rx - num);
267 kni_stats[port_id].rx_dropped += nb_rx - num;

Completed in 26 milliseconds

123