Searched refs:rxq (Results 1 - 25 of 89) sorted by relevance

1234

/deb_dpdk/drivers/net/bnxt/
H A Dbnxt_rxq.c52 void bnxt_free_rxq_stats(struct bnxt_rx_queue *rxq) argument
54 struct bnxt_cp_ring_info *cpr = rxq->cp_ring;
67 struct bnxt_rx_queue *rxq; local
82 rxq = bp->eth_dev->data->rx_queues[0];
83 rxq->vnic = vnic;
145 rxq = bp->eth_dev->data->rx_queues[ring_idx];
146 rxq->vnic = vnic;
183 rxq = bp->eth_dev->data->rx_queues[i];
184 rxq->vnic = vnic;
216 static void bnxt_rx_queue_release_mbufs(struct bnxt_rx_queue *rxq __rte_unuse
237 struct bnxt_rx_queue *rxq; local
248 struct bnxt_rx_queue *rxq = (struct bnxt_rx_queue *)rx_queue; local
275 struct bnxt_rx_queue *rxq; local
[all...]
H A Dbnxt_rxr.c61 static inline int bnxt_alloc_rx_data(struct bnxt_rx_queue *rxq, argument
69 data = __bnxt_alloc_rx_data(rxq->mb_pool);
99 struct bnxt_rx_queue *rxq, uint32_t *raw_cons)
101 struct bnxt_cp_ring_info *cpr = rxq->cp_ring;
102 struct bnxt_rx_ring_info *rxr = rxq->rx_ring;
135 mbuf->port = rxq->port_id;
190 if (bnxt_alloc_rx_data(rxq, rxr, prod)) {
217 struct bnxt_rx_queue *rxq = rx_queue; local
218 struct bnxt_cp_ring_info *cpr = rxq->cp_ring;
219 struct bnxt_rx_ring_info *rxr = rxq
98 bnxt_rx_pkt(struct rte_mbuf **rx_pkt, struct bnxt_rx_queue *rxq, uint32_t *raw_cons) argument
270 struct bnxt_rx_queue *rxq = bp->rx_queues[i]; local
288 bnxt_init_rx_ring_struct(struct bnxt_rx_queue *rxq, unsigned int socket_id) argument
358 bnxt_init_one_rx_ring(struct bnxt_rx_queue *rxq) argument
[all...]
/deb_dpdk/drivers/net/i40e/
H A Di40e_rxtx_vec_common.h44 reassemble_packets(struct i40e_rx_queue *rxq, struct rte_mbuf **rx_bufs, argument
48 struct rte_mbuf *start = rxq->pkt_first_seg;
49 struct rte_mbuf *end = rxq->pkt_last_seg;
56 rx_bufs[buf_idx]->data_len += rxq->crc_len;
67 start->pkt_len -= rxq->crc_len;
68 if (end->data_len > rxq->crc_len) {
69 end->data_len -= rxq->crc_len;
77 secondlast->data_len -= (rxq->crc_len -
94 rx_bufs[buf_idx]->data_len += rxq->crc_len;
95 rx_bufs[buf_idx]->pkt_len += rxq
174 _i40e_rx_queue_release_mbufs_vec(struct i40e_rx_queue *rxq) argument
204 i40e_rxq_vec_setup_default(struct i40e_rx_queue *rxq) argument
[all...]
H A Di40e_rxtx.c393 check_rx_burst_bulk_alloc_preconditions(struct i40e_rx_queue *rxq) argument
395 check_rx_burst_bulk_alloc_preconditions(__rte_unused struct i40e_rx_queue *rxq)
401 if (!(rxq->rx_free_thresh >= RTE_PMD_I40E_RX_MAX_BURST)) {
403 "rxq->rx_free_thresh=%d, "
405 rxq->rx_free_thresh, RTE_PMD_I40E_RX_MAX_BURST);
407 } else if (!(rxq->rx_free_thresh < rxq->nb_rx_desc)) {
409 "rxq->rx_free_thresh=%d, "
410 "rxq->nb_rx_desc=%d",
411 rxq
442 i40e_rx_scan_hw_ring(struct i40e_rx_queue *rxq) argument
532 i40e_rx_fill_from_stage(struct i40e_rx_queue *rxq, struct rte_mbuf **rx_pkts, uint16_t nb_pkts) argument
551 i40e_rx_alloc_bufs(struct i40e_rx_queue *rxq) argument
604 struct i40e_rx_queue *rxq = (struct i40e_rx_queue *)rx_queue; local
682 struct i40e_rx_queue *rxq; local
804 struct i40e_rx_queue *rxq = rx_queue; local
1481 struct i40e_rx_queue *rxq; local
1519 struct i40e_rx_queue *rxq; local
1659 struct i40e_rx_queue *rxq; local
1789 i40e_dev_rx_queue_release(void *rxq) argument
1808 struct i40e_rx_queue *rxq; local
1841 struct i40e_rx_queue *rxq = rx_queue; local
2090 i40e_rx_queue_release_mbufs(struct i40e_rx_queue *rxq) argument
2125 i40e_reset_rx_queue(struct i40e_rx_queue *rxq) argument
2290 i40e_alloc_rx_queue_mbufs(struct i40e_rx_queue *rxq) argument
2333 i40e_rx_queue_config(struct i40e_rx_queue *rxq) argument
2390 i40e_rx_queue_init(struct i40e_rx_queue *rxq) argument
2571 struct i40e_rx_queue *rxq; local
2630 struct i40e_rx_queue *rxq; local
2683 struct i40e_rx_queue *rxq = local
2746 struct i40e_rx_queue *rxq = dev->data->rx_queues[i]; local
2836 i40e_rxq_vec_setup(struct i40e_rx_queue __rte_unused *rxq) argument
[all...]
H A Di40e_rxtx_vec_sse.c51 i40e_rxq_rearm(struct i40e_rx_queue *rxq) argument
56 struct i40e_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start];
62 rxdp = rxq->rx_ring + rxq->rxrearm_start;
65 if (rte_mempool_get_bulk(rxq->mp,
68 if (rxq->rxrearm_nb + RTE_I40E_RXQ_REARM_THRESH >=
69 rxq->nb_rx_desc) {
72 rxep[i].mbuf = &rxq->fake_mbuf;
77 rte_eth_devices[rxq
242 _recv_raw_pkts_vec(struct i40e_rx_queue *rxq, struct rte_mbuf **rx_pkts, uint16_t nb_pkts, uint8_t *split_packet) argument
492 struct i40e_rx_queue *rxq = rx_queue; local
614 i40e_rx_queue_release_mbufs_vec(struct i40e_rx_queue *rxq) argument
620 i40e_rxq_vec_setup(struct i40e_rx_queue *rxq) argument
[all...]
H A Di40e_rxtx_vec_altivec.c50 i40e_rxq_rearm(struct i40e_rx_queue *rxq) argument
56 struct i40e_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start];
64 rxdp = rxq->rx_ring + rxq->rxrearm_start;
67 if (rte_mempool_get_bulk(rxq->mp,
70 if (rxq->rxrearm_nb + RTE_I40E_RXQ_REARM_THRESH >=
71 rxq->nb_rx_desc) {
74 rxep[i].mbuf = &rxq->fake_mbuf;
79 rte_eth_devices[rxq
236 _recv_raw_pkts_vec(struct i40e_rx_queue *rxq, struct rte_mbuf **rx_pkts, uint16_t nb_pkts, uint8_t *split_packet) argument
510 struct i40e_rx_queue *rxq = rx_queue; local
633 i40e_rx_queue_release_mbufs_vec(struct i40e_rx_queue *rxq) argument
639 i40e_rxq_vec_setup(struct i40e_rx_queue *rxq) argument
[all...]
H A Di40e_rxtx_vec_neon.c50 i40e_rxq_rearm(struct i40e_rx_queue *rxq) argument
55 struct i40e_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start];
62 rxdp = rxq->rx_ring + rxq->rxrearm_start;
65 if (unlikely(rte_mempool_get_bulk(rxq->mp,
68 if (rxq->rxrearm_nb + RTE_I40E_RXQ_REARM_THRESH >=
69 rxq->nb_rx_desc) {
71 rxep[i].mbuf = &rxq->fake_mbuf;
75 rte_eth_devices[rxq
219 _recv_raw_pkts_vec(struct i40e_rx_queue *rxq, struct rte_mbuf **rx_pkts, uint16_t nb_pkts, uint8_t *split_packet) argument
477 struct i40e_rx_queue *rxq = rx_queue; local
598 i40e_rx_queue_release_mbufs_vec(struct i40e_rx_queue *rxq) argument
604 i40e_rxq_vec_setup(struct i40e_rx_queue *rxq) argument
[all...]
/deb_dpdk/drivers/net/bnx2x/
H A Dbnx2x_rxtx.c53 bnx2x_dev_rx_queue_release(void *rxq) argument
55 bnx2x_rx_queue_release(rxq);
68 struct bnx2x_rx_queue *rxq; local
78 rxq = rte_zmalloc_socket("ethdev RX queue", sizeof(struct bnx2x_rx_queue),
80 if (NULL == rxq) {
81 PMD_DRV_LOG(ERR, sc, "rte_zmalloc for rxq failed!");
84 rxq->sc = sc;
85 rxq->mb_pool = mp;
86 rxq->queue_id = queue_idx;
87 rxq
348 struct bnx2x_rx_queue *rxq = p_rxq; local
480 struct bnx2x_rx_queue *rxq = dev->data->rx_queues[i]; local
[all...]
/deb_dpdk/drivers/net/mlx5/
H A Dmlx5_stats.c66 struct rxq *rxq = (*priv->rxqs)[i]; local
68 if (rxq == NULL)
70 idx = rxq->stats.idx;
73 tmp.q_ipackets[idx] += rxq->stats.ipackets;
74 tmp.q_ibytes[idx] += rxq->stats.ibytes;
76 tmp.q_errors[idx] += (rxq->stats.idropped +
77 rxq->stats.rx_nombuf);
80 tmp.ipackets += rxq->stats.ipackets;
81 tmp.ibytes += rxq
[all...]
H A Dmlx5_rxq.c382 struct rxq_ctrl, rxq);
647 const unsigned int sges_n = 1 << rxq_ctrl->rxq.sges_n;
655 &(*rxq_ctrl->rxq.wqes)[i];
663 buf = rte_pktmbuf_alloc(rxq_ctrl->rxq.mp);
679 PORT(buf) = rxq_ctrl->rxq.port_id;
690 (*rxq_ctrl->rxq.elts)[i] = buf;
693 (void *)rxq_ctrl, elts_n, elts_n / (1 << rxq_ctrl->rxq.sges_n));
700 if ((*rxq_ctrl->rxq.elts)[i] != NULL)
701 rte_pktmbuf_free_seg((*rxq_ctrl->rxq.elts)[i]);
702 (*rxq_ctrl->rxq
1221 struct rxq *rxq = (*priv->rxqs)[idx]; local
1300 struct rxq *rxq = (struct rxq *)dpdk_rxq; local
1345 struct rxq *rxq = dpdk_rxq; local
[all...]
H A Dmlx5_rxtx.c90 mlx5_rx_poll_len(struct rxq *rxq, volatile struct mlx5_cqe *cqe,
95 rxq_cq_to_ol_flags(struct rxq *rxq, volatile struct mlx5_cqe *cqe)
1145 * @param rxq
1157 mlx5_rx_poll_len(struct rxq *rxq, volatile struct mlx5_cqe *cqe, argument
1160 struct rxq_zip *zip = &rxq->zip;
1168 (uintptr_t)(&(*rxq->cqes)[zip->ca & cqe_cnt]);
1181 if (unlikely(rxq
1251 rxq_cq_to_ol_flags(struct rxq *rxq, volatile struct mlx5_cqe *cqe) argument
1290 struct rxq *rxq = dpdk_rxq; local
[all...]
H A Dmlx5_vlan.c153 struct rxq *rxq = (*priv->rxqs)[idx]; local
154 struct rxq_ctrl *rxq_ctrl = container_of(rxq, struct rxq_ctrl, rxq);
162 vlan_offloads, rxq->port_id, idx);
176 rxq->vlan_strip = !!on;
/deb_dpdk/drivers/net/e1000/
H A Dem_rxtx.c655 struct em_rx_queue *rxq; local
668 rxq = rx_queue;
672 rx_id = rxq->rx_tail;
673 rx_ring = rxq->rx_ring;
674 sw_ring = rxq->sw_ring;
718 (unsigned) rxq->port_id, (unsigned) rxq->queue_id,
722 nmb = rte_mbuf_raw_alloc(rxq->mb_pool);
726 (unsigned) rxq->port_id,
727 (unsigned) rxq
827 struct em_rx_queue *rxq; local
1259 em_rx_queue_release_mbufs(struct em_rx_queue *rxq) argument
1274 em_rx_queue_release(struct em_rx_queue *rxq) argument
1284 eth_em_rx_queue_release(void *rxq) argument
1291 em_reset_rx_queue(struct em_rx_queue *rxq) argument
1308 struct em_rx_queue *rxq; local
1391 struct em_rx_queue *rxq; local
1418 struct em_rx_queue *rxq = rx_queue; local
1436 struct em_rx_queue *rxq; local
1543 em_alloc_rx_queue_mbufs(struct em_rx_queue *rxq) argument
1586 struct em_rx_queue *rxq; local
1827 struct em_rx_queue *rxq; local
[all...]
H A Digb_rxtx.c704 rx_desc_hlen_type_rss_to_pkt_flags(struct igb_rx_queue *rxq, uint32_t hl_tp_rs) argument
714 struct rte_eth_dev dev = rte_eth_devices[rxq->port_id];
723 RTE_SET_USED(rxq);
767 struct igb_rx_queue *rxq; local
786 rxq = rx_queue;
787 rx_id = rxq->rx_tail;
788 rx_ring = rxq->rx_ring;
789 sw_ring = rxq->sw_ring;
833 (unsigned) rxq->port_id, (unsigned) rxq
944 struct igb_rx_queue *rxq; local
1377 igb_rx_queue_release_mbufs(struct igb_rx_queue *rxq) argument
1392 igb_rx_queue_release(struct igb_rx_queue *rxq) argument
1402 eth_igb_rx_queue_release(void *rxq) argument
1408 igb_reset_rx_queue(struct igb_rx_queue *rxq) argument
1432 struct igb_rx_queue *rxq; local
1516 struct igb_rx_queue *rxq; local
1543 struct igb_rx_queue *rxq = rx_queue; local
1561 struct igb_rx_queue *rxq; local
1947 igb_alloc_rx_queue_mbufs(struct igb_rx_queue *rxq) argument
2018 struct igb_rx_queue *rxq; local
2318 struct igb_rx_queue *rxq; local
2499 struct igb_rx_queue *rxq; local
[all...]
/deb_dpdk/drivers/net/ixgbe/
H A Dixgbe_rxtx_vec_neon.c47 ixgbe_rxq_rearm(struct ixgbe_rx_queue *rxq) argument
52 struct ixgbe_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start];
59 rxdp = rxq->rx_ring + rxq->rxrearm_start;
62 if (unlikely(rte_mempool_get_bulk(rxq->mb_pool,
65 if (rxq->rxrearm_nb + RTE_IXGBE_RXQ_REARM_THRESH >=
66 rxq->nb_rx_desc) {
68 rxep[i].mbuf = &rxq->fake_mbuf;
73 rte_eth_devices[rxq
192 _recv_raw_pkts_vec(struct ixgbe_rx_queue *rxq, struct rte_mbuf **rx_pkts, uint16_t nb_pkts, uint8_t *split_packet) argument
403 struct ixgbe_rx_queue *rxq = rx_queue; local
526 ixgbe_rx_queue_release_mbufs_vec(struct ixgbe_rx_queue *rxq) argument
550 ixgbe_rxq_vec_setup(struct ixgbe_rx_queue *rxq) argument
[all...]
H A Dixgbe_rxtx_vec_common.h43 reassemble_packets(struct ixgbe_rx_queue *rxq, struct rte_mbuf **rx_bufs, argument
47 struct rte_mbuf *start = rxq->pkt_first_seg;
48 struct rte_mbuf *end = rxq->pkt_last_seg;
55 rx_bufs[buf_idx]->data_len += rxq->crc_len;
66 start->pkt_len -= rxq->crc_len;
67 if (end->data_len > rxq->crc_len)
68 end->data_len -= rxq->crc_len;
76 secondlast->data_len -= (rxq->crc_len -
92 rx_bufs[buf_idx]->data_len += rxq->crc_len;
93 rx_bufs[buf_idx]->pkt_len += rxq
198 _ixgbe_rx_queue_release_mbufs_vec(struct ixgbe_rx_queue *rxq) argument
275 ixgbe_rxq_vec_setup_default(struct ixgbe_rx_queue *rxq) argument
[all...]
H A Dixgbe_rxtx_vec_sse.c49 ixgbe_rxq_rearm(struct ixgbe_rx_queue *rxq) argument
54 struct ixgbe_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start];
62 rxdp = rxq->rx_ring + rxq->rxrearm_start;
65 if (rte_mempool_get_bulk(rxq->mb_pool,
68 if (rxq->rxrearm_nb + RTE_IXGBE_RXQ_REARM_THRESH >=
69 rxq->nb_rx_desc) {
72 rxep[i].mbuf = &rxq->fake_mbuf;
77 rte_eth_devices[rxq
254 _recv_raw_pkts_vec(struct ixgbe_rx_queue *rxq, struct rte_mbuf **rx_pkts, uint16_t nb_pkts, uint8_t *split_packet) argument
498 struct ixgbe_rx_queue *rxq = rx_queue; local
621 ixgbe_rx_queue_release_mbufs_vec(struct ixgbe_rx_queue *rxq) argument
645 ixgbe_rxq_vec_setup(struct ixgbe_rx_queue *rxq) argument
[all...]
H A Dixgbe_rxtx.c1379 ixgbe_rx_scan_hw_ring(struct ixgbe_rx_queue *rxq) argument
1391 uint64_t vlan_flags = rxq->vlan_flags;
1394 rxdp = &rxq->rx_ring[rxq->rx_tail];
1395 rxep = &rxq->sw_ring[rxq->rx_tail];
1429 rxq->crc_len;
1443 (pkt_info[j], rxq->pkt_type_mask);
1459 rxq->rx_stage[i + j] = rxep[j].mbuf;
1469 rxq
1477 ixgbe_rx_alloc_bufs(struct ixgbe_rx_queue *rxq, bool reset_mbuf) argument
1523 ixgbe_rx_fill_from_stage(struct ixgbe_rx_queue *rxq, struct rte_mbuf **rx_pkts, uint16_t nb_pkts) argument
1547 struct ixgbe_rx_queue *rxq = (struct ixgbe_rx_queue *)rx_queue; local
1636 struct ixgbe_rx_queue *rxq; local
1847 ixgbe_fill_cluster_head_buf( struct rte_mbuf *head, union ixgbe_adv_rx_desc *desc, struct ixgbe_rx_queue *rxq, uint32_t staterr) argument
1911 struct ixgbe_rx_queue *rxq = rx_queue; local
2521 ixgbe_rx_queue_release_mbufs(struct ixgbe_rx_queue *rxq) argument
2560 ixgbe_rx_queue_release(struct ixgbe_rx_queue *rxq) argument
2571 ixgbe_dev_rx_queue_release(void *rxq) argument
2585 check_rx_burst_bulk_alloc_preconditions(struct ixgbe_rx_queue *rxq) argument
2632 ixgbe_reset_rx_queue(struct ixgbe_adapter *adapter, struct ixgbe_rx_queue *rxq) argument
2693 struct ixgbe_rx_queue *rxq; local
2863 struct ixgbe_rx_queue *rxq; local
2891 struct ixgbe_rx_queue *rxq = rx_queue; local
2924 struct ixgbe_rx_queue *rxq = dev->data->rx_queues[i]; local
3997 ixgbe_alloc_rx_queue_mbufs(struct ixgbe_rx_queue *rxq) argument
4400 struct ixgbe_rx_queue *rxq = dev->data->rx_queues[i]; local
4480 struct ixgbe_rx_queue *rxq = dev->data->rx_queues[i]; local
4555 struct ixgbe_rx_queue *rxq; local
4837 struct ixgbe_rx_queue *rxq; local
4903 struct ixgbe_rx_queue *rxq; local
4951 struct ixgbe_rx_queue *rxq; local
5094 struct ixgbe_rx_queue *rxq; local
5134 struct ixgbe_rx_queue *rxq; local
5311 struct ixgbe_rx_queue *rxq; local
5393 ixgbe_rxq_vec_setup(struct ixgbe_rx_queue __rte_unused *rxq) argument
[all...]
/deb_dpdk/drivers/net/vmxnet3/
H A Dvmxnet3_rxtx.c90 vmxnet3_rxq_dump(struct vmxnet3_rx_queue *rxq) argument
94 if (rxq == NULL)
99 rxq->cmd_ring[0].base, rxq->cmd_ring[1].base, rxq->comp_ring.base);
102 (unsigned long)rxq->cmd_ring[0].basePA,
103 (unsigned long)rxq->cmd_ring[1].basePA,
104 (unsigned long)rxq->comp_ring.basePA);
106 avail = vmxnet3_cmd_ring_desc_avail(&rxq->cmd_ring[0]);
109 (uint32_t)rxq
203 vmxnet3_dev_rx_queue_release(void *rxq) argument
254 vmxnet3_dev_rx_queue_reset(void *rxq) argument
303 struct vmxnet3_rx_queue *rxq = dev->data->rx_queues[i]; local
530 vmxnet3_renew_desc(vmxnet3_rx_queue_t *rxq, uint8_t ring_id, struct rte_mbuf *mbuf) argument
566 vmxnet3_post_rx_bufs(vmxnet3_rx_queue_t *rxq, uint8_t ring_id) argument
671 vmxnet3_rx_queue_t *rxq; local
946 struct vmxnet3_rx_queue *rxq; local
1063 vmxnet3_rx_queue_t *rxq = dev->data->rx_queues[i]; local
[all...]
/deb_dpdk/drivers/net/qede/
H A Dqede_rxtx.c13 static inline int qede_alloc_rx_buffer(struct qede_rx_queue *rxq) argument
18 uint16_t idx = rxq->sw_rx_prod & NUM_RX_BDS(rxq);
20 new_mb = rte_mbuf_raw_alloc(rxq->mb_pool);
22 PMD_RX_LOG(ERR, rxq,
25 idx, rxq->sw_rx_cons & NUM_RX_BDS(rxq),
26 rte_mempool_avail_count(rxq->mb_pool),
27 rte_mempool_in_use_count(rxq->mb_pool));
30 rxq
41 qede_rx_queue_release_mbufs(struct qede_rx_queue *rxq) argument
57 struct qede_rx_queue *rxq = rx_queue; local
93 struct qede_rx_queue *rxq; local
332 int fp_rx = qdev->fp_num_rx, rxq = 0, txq = 0; local
479 qede_update_rx_prod(struct qede_dev *edev, struct qede_rx_queue *rxq) argument
778 qede_rx_bd_ring_consume(struct qede_rx_queue *rxq) argument
785 qede_reuse_page(struct qede_dev *qdev, struct qede_rx_queue *rxq, struct qede_rx_entry *curr_cons) argument
806 qede_recycle_rx_bd_ring(struct qede_rx_queue *rxq, struct qede_dev *qdev, uint8_t count) argument
839 struct qede_rx_queue *rxq = p_rxq; local
892 struct qede_rx_queue *rxq = p_rxq; local
1265 int rxq = 0, txq = 0; local
[all...]
/deb_dpdk/drivers/net/mlx4/
H A Dmlx4.c174 struct rxq { struct
282 struct rxq rxq_parent; /* Parent queue when RSS is enabled. */
285 struct rxq *(*rxqs)[]; /* RX queues. */
706 rxq_setup(struct rte_eth_dev *dev, struct rxq *rxq, uint16_t desc,
711 rxq_cleanup(struct rxq *rxq);
930 struct rxq *primary_rxq = (*sd->primary_priv->rxqs)[i];
2122 * @param rxq
2134 rxq_alloc_elts_sp(struct rxq *rx argument
2241 rxq_free_elts_sp(struct rxq *rxq) argument
2281 rxq_alloc_elts(struct rxq *rxq, unsigned int elts_n, struct rte_mbuf **pool) argument
2387 rxq_free_elts(struct rxq *rxq) argument
2423 rxq_del_flow(struct rxq *rxq, unsigned int mac_index, unsigned int vlan_index) argument
2450 rxq_mac_addr_del(struct rxq *rxq, unsigned int mac_index) argument
2478 rxq_mac_addrs_del(struct rxq *rxq) argument
2504 rxq_add_flow(struct rxq *rxq, unsigned int mac_index, unsigned int vlan_index) argument
2588 rxq_mac_addr_add(struct rxq *rxq, unsigned int mac_index) argument
2635 rxq_mac_addrs_add(struct rxq *rxq) argument
2775 rxq_allmulticast_enable(struct rxq *rxq) argument
2811 rxq_allmulticast_disable(struct rxq *rxq) argument
2831 rxq_promiscuous_enable(struct rxq *rxq) argument
2869 rxq_promiscuous_disable(struct rxq *rxq) argument
2890 rxq_cleanup(struct rxq *rxq) argument
2998 rxq_cq_to_ol_flags(const struct rxq *rxq, uint32_t flags) argument
3040 struct rxq *rxq = (struct rxq *)dpdk_rxq; local
3268 struct rxq *rxq = (struct rxq *)dpdk_rxq; local
3430 struct rxq *rxq = dpdk_rxq; local
3581 rxq_rehash(struct rte_eth_dev *dev, struct rxq *rxq) argument
3791 rxq_setup(struct rte_eth_dev *dev, struct rxq *rxq, uint16_t desc, unsigned int socket, int inactive, const struct rte_eth_rxconf *conf, struct rte_mempool *mp) argument
4019 struct rxq *rxq = (*priv->rxqs)[idx]; local
4081 struct rxq *rxq = (struct rxq *)dpdk_rxq; local
4124 struct rxq *rxq; local
4190 struct rxq *rxq; local
4520 struct rxq *rxq = (*priv->rxqs)[i]; local
4960 struct rxq *rxq = (*priv->rxqs)[i]; local
[all...]
/deb_dpdk/drivers/net/nfp/
H A Dnfp_net.c76 static int nfp_net_rx_fill_freelist(struct nfp_net_rxq *rxq);
81 static void nfp_net_rx_queue_release(void *rxq);
280 nfp_net_rx_queue_release_mbufs(struct nfp_net_rxq *rxq) argument
284 if (rxq->rxbufs == NULL)
287 for (i = 0; i < rxq->rx_count; i++) {
288 if (rxq->rxbufs[i].mbuf) {
289 rte_pktmbuf_free_seg(rxq->rxbufs[i].mbuf);
290 rxq->rxbufs[i].mbuf = NULL;
298 struct nfp_net_rxq *rxq = rx_queue; local
300 if (rxq) {
308 nfp_net_reset_rx_queue(struct nfp_net_rxq *rxq) argument
1081 struct nfp_net_rxq *rxq; local
1273 struct nfp_net_rxq *rxq; local
1373 nfp_net_rx_fill_freelist(struct nfp_net_rxq *rxq) argument
1561 nfp_net_rx_cksum(struct nfp_net_rxq *rxq, struct nfp_net_rx_desc *rxd, struct rte_mbuf *mb) argument
1597 nfp_net_set_hash(struct nfp_net_rxq *rxq, struct nfp_net_rx_desc *rxd, struct rte_mbuf *mbuf) argument
1673 nfp_net_mbuf_alloc_failed(struct nfp_net_rxq *rxq) argument
1711 struct nfp_net_rxq *rxq; local
[all...]
/deb_dpdk/drivers/net/fm10k/
H A Dfm10k_rxtx_vec.c262 fm10k_rxq_vec_setup(struct fm10k_rx_queue *rxq) argument
272 mb_def.port = rxq->port_id;
278 rxq->mbuf_initializer = *(uint64_t *)p;
283 fm10k_rxq_rearm(struct fm10k_rx_queue *rxq) argument
288 struct rte_mbuf **mb_alloc = &rxq->sw_ring[rxq->rxrearm_start];
298 rxdp = rxq->hw_ring + rxq->rxrearm_start;
301 if (rte_mempool_get_bulk(rxq->mp,
307 mb_alloc[i] = &rxq
381 fm10k_rx_queue_release_mbufs_vec(struct fm10k_rx_queue *rxq) argument
406 struct fm10k_rx_queue *rxq = rx_queue; local
611 fm10k_reassemble_packets(struct fm10k_rx_queue *rxq, struct rte_mbuf **rx_bufs, uint16_t nb_bufs, uint8_t *split_flags) argument
669 struct fm10k_rx_queue *rxq = rx_queue; local
[all...]
/deb_dpdk/drivers/net/thunderx/
H A Dnicvf_rxtx.c371 nicvf_fill_rbdr(struct nicvf_rxq *rxq, int to_fill) argument
375 struct nicvf_rbdr *rbdr = rxq->shared_rbdr;
376 uint64_t mbuf_phys_off = rxq->mbuf_phys_off;
382 if (unlikely(rte_mempool_get_bulk(rxq->pool, obj_p, to_fill) < 0)) {
383 rte_eth_devices[rxq->port_id].data->rx_mbuf_alloc_failed +=
411 nicvf_rx_pkts_to_process(struct nicvf_rxq *rxq, uint16_t nb_pkts, argument
415 rxq->available_space = nicvf_addr_read(rxq->cq_status)
441 struct nicvf_rxq *rxq = rx_queue; local
442 union cq_entry_t *desc = rxq
548 struct nicvf_rxq *rxq = rx_queue; local
589 struct nicvf_rxq *rxq; local
598 struct nicvf_rxq *rxq; local
[all...]
H A Dnicvf_ethdev.c575 struct nicvf_rxq *rxq, uint16_t qidx, uint32_t desc_cnt)
590 rxq->phys = rz->phys_addr;
591 rxq->desc = rz->addr;
592 rxq->qlen_mask = desc_cnt - 1;
669 struct nicvf_rxq *rxq; local
676 rxq = dev->data->rx_queues[qidx];
677 if (rxq->precharge_cnt) {
679 rxq->mbuf_phys_off);
680 rte_mempool_put(rxq->pool, obj);
681 rxq
574 nicvf_qset_cq_alloc(struct rte_eth_dev *dev, struct nicvf *nic, struct nicvf_rxq *rxq, uint16_t qidx, uint32_t desc_cnt) argument
1041 nicvf_rx_queue_release_mbufs(struct rte_eth_dev *dev, struct nicvf_rxq *rxq) argument
1071 nicvf_rx_queue_reset(struct nicvf_rxq *rxq) argument
1082 struct nicvf_rxq *rxq; local
1123 struct nicvf_rxq *rxq; local
1232 struct nicvf_rxq *rxq; local
1387 struct nicvf_rxq *rxq; local
1415 struct nicvf_rxq *rxq; local
1615 struct nicvf_rxq *rxq; local
[all...]

Completed in 49 milliseconds

1234