Searched refs:txq (Results 1 - 25 of 80) sorted by relevance

1234

/deb_dpdk/drivers/net/bnxt/
H A Dbnxt_txq.c48 void bnxt_free_txq_stats(struct bnxt_tx_queue *txq) argument
50 struct bnxt_cp_ring_info *cpr = txq->cp_ring;
56 static void bnxt_tx_queue_release_mbufs(struct bnxt_tx_queue *txq) argument
61 sw_ring = txq->tx_ring->tx_buf_ring;
63 for (i = 0; i < txq->tx_ring->tx_ring_struct->ring_size; i++) {
74 struct bnxt_tx_queue *txq; local
78 txq = bp->tx_queues[i];
79 bnxt_tx_queue_release_mbufs(txq);
85 struct bnxt_tx_queue *txq = (struct bnxt_tx_queue *)tx_queue; local
87 if (txq) {
110 struct bnxt_tx_queue *txq; local
[all...]
H A Dbnxt_txr.c56 struct bnxt_tx_queue *txq = bp->tx_queues[i]; local
58 if (!txq)
61 bnxt_free_ring(txq->tx_ring->tx_ring_struct);
62 rte_free(txq->tx_ring->tx_ring_struct);
63 rte_free(txq->tx_ring);
65 bnxt_free_ring(txq->cp_ring->cp_ring_struct);
66 rte_free(txq->cp_ring->cp_ring_struct);
67 rte_free(txq->cp_ring);
69 rte_free(txq);
74 int bnxt_init_one_tx_ring(struct bnxt_tx_queue *txq) argument
85 bnxt_init_tx_ring_struct(struct bnxt_tx_queue *txq, unsigned int socket_id) argument
144 bnxt_start_xmit(struct rte_mbuf *tx_pkt, struct bnxt_tx_queue *txq) argument
324 bnxt_tx_cmp(struct bnxt_tx_queue *txq, int nr_pkts) argument
349 bnxt_handle_tx_cp(struct bnxt_tx_queue *txq) argument
386 struct bnxt_tx_queue *txq = tx_queue; local
[all...]
/deb_dpdk/drivers/net/ixgbe/
H A Dixgbe_rxtx_vec_common.h105 ixgbe_tx_free_bufs(struct ixgbe_tx_queue *txq) argument
115 status = txq->tx_ring[txq->tx_next_dd].wb.status;
119 n = txq->tx_rs_thresh;
125 txep = &txq->sw_ring_v[txq->tx_next_dd - (n - 1)];
153 txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + txq->tx_rs_thresh);
154 txq
172 _ixgbe_tx_queue_release_mbufs_vec(struct ixgbe_tx_queue *txq) argument
228 _ixgbe_tx_free_swring_vec(struct ixgbe_tx_queue *txq) argument
240 _ixgbe_reset_tx_queue_vec(struct ixgbe_tx_queue *txq) argument
293 ixgbe_txq_vec_setup_default(struct ixgbe_tx_queue *txq, const struct ixgbe_txq_ops *txq_ops) argument
[all...]
H A Dixgbe_rxtx_vec_neon.c457 struct ixgbe_tx_queue *txq = (struct ixgbe_tx_queue *)tx_queue; local
466 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh);
468 if (txq->nb_tx_free < txq->tx_free_thresh)
469 ixgbe_tx_free_bufs(txq);
471 nb_commit = nb_pkts = (uint16_t)RTE_MIN(txq->nb_tx_free, nb_pkts);
475 tx_id = txq->tx_tail;
476 txdp = &txq->tx_ring[tx_id];
477 txep = &txq->sw_ring_v[tx_id];
479 txq
520 ixgbe_tx_queue_release_mbufs_vec(struct ixgbe_tx_queue *txq) argument
532 ixgbe_tx_free_swring(struct ixgbe_tx_queue *txq) argument
538 ixgbe_reset_tx_queue(struct ixgbe_tx_queue *txq) argument
556 ixgbe_txq_vec_setup(struct ixgbe_tx_queue *txq) argument
[all...]
H A Dixgbe_rxtx.c118 ixgbe_tx_free_bufs(struct ixgbe_tx_queue *txq) argument
126 status = txq->tx_ring[txq->tx_next_dd].wb.status;
134 txep = &(txq->sw_ring[txq->tx_next_dd - (txq->tx_rs_thresh - 1)]);
136 for (i = 0; i < txq->tx_rs_thresh; ++i, ++txep) {
158 txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + txq
215 ixgbe_tx_fill_hw_ring(struct ixgbe_tx_queue *txq, struct rte_mbuf **pkts, uint16_t nb_pkts) argument
251 struct ixgbe_tx_queue *txq = (struct ixgbe_tx_queue *)tx_queue; local
360 ixgbe_set_xmit_ctx(struct ixgbe_tx_queue *txq, volatile struct ixgbe_adv_tx_context_desc *ctx_txd, uint64_t ol_flags, union ixgbe_tx_offload tx_offload) argument
473 what_advctx_update(struct ixgbe_tx_queue *txq, uint64_t flags, union ixgbe_tx_offload tx_offload) argument
539 ixgbe_xmit_cleanup(struct ixgbe_tx_queue *txq) argument
601 struct ixgbe_tx_queue *txq; local
2195 ixgbe_tx_queue_release_mbufs(struct ixgbe_tx_queue *txq) argument
2210 ixgbe_tx_free_swring(struct ixgbe_tx_queue *txq) argument
2218 ixgbe_tx_queue_release(struct ixgbe_tx_queue *txq) argument
2228 ixgbe_dev_tx_queue_release(void *txq) argument
2235 ixgbe_reset_tx_queue(struct ixgbe_tx_queue *txq) argument
2285 ixgbe_set_tx_function(struct rte_eth_dev *dev, struct ixgbe_tx_queue *txq) argument
2322 struct ixgbe_tx_queue *txq; local
2915 struct ixgbe_tx_queue *txq = dev->data->tx_queues[i]; local
4738 struct ixgbe_tx_queue *txq; local
4836 struct ixgbe_tx_queue *txq; local
4994 struct ixgbe_tx_queue *txq; local
5036 struct ixgbe_tx_queue *txq; local
5111 struct ixgbe_tx_queue *txq; local
5269 struct ixgbe_tx_queue *txq; local
5310 struct ixgbe_tx_queue *txq; local
[all...]
H A Dixgbe_rxtx_vec_sse.c551 struct ixgbe_tx_queue *txq = (struct ixgbe_tx_queue *)tx_queue; local
560 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh);
562 if (txq->nb_tx_free < txq->tx_free_thresh)
563 ixgbe_tx_free_bufs(txq);
565 nb_commit = nb_pkts = (uint16_t)RTE_MIN(txq->nb_tx_free, nb_pkts);
569 tx_id = txq->tx_tail;
570 txdp = &txq->tx_ring[tx_id];
571 txep = &txq->sw_ring_v[tx_id];
573 txq
615 ixgbe_tx_queue_release_mbufs_vec(struct ixgbe_tx_queue *txq) argument
627 ixgbe_tx_free_swring(struct ixgbe_tx_queue *txq) argument
633 ixgbe_reset_tx_queue(struct ixgbe_tx_queue *txq) argument
651 ixgbe_txq_vec_setup(struct ixgbe_tx_queue *txq) argument
[all...]
/deb_dpdk/drivers/net/mlx5/
H A Dmlx5_rxtx.c78 txq_mp2mr(struct txq *txq, struct rte_mempool *mp)
82 mlx5_tx_dbrec(struct txq *txq, volatile struct mlx5_wqe *wqe)
177 txq_complete(struct txq *txq) __attribute__((always_inline));
184 * @param txq
188 txq_complete(struct txq *txq) argument
190 const unsigned int elts_n = 1 << txq
284 txq_mp2mr(struct txq *txq, struct rte_mempool *mp) argument
316 mlx5_tx_dbrec(struct txq *txq, volatile struct mlx5_wqe *wqe) argument
337 tx_prefetch_cqe(struct txq *txq, uint16_t ci) argument
354 tx_prefetch_wqe(struct txq *txq, uint16_t ci) argument
378 struct txq *txq = (struct txq *)dpdk_txq; local
630 mlx5_mpw_new(struct txq *txq, struct mlx5_mpw *mpw, uint32_t length) argument
669 mlx5_mpw_close(struct txq *txq, struct mlx5_mpw *mpw) argument
703 struct txq *txq = (struct txq *)dpdk_txq; local
840 mlx5_mpw_inline_new(struct txq *txq, struct mlx5_mpw *mpw, uint32_t length) argument
875 mlx5_mpw_inline_close(struct txq *txq, struct mlx5_mpw *mpw) argument
909 struct txq *txq = (struct txq *)dpdk_txq; local
[all...]
H A Dmlx5_stats.c87 struct txq *txq = (*priv->txqs)[i]; local
89 if (txq == NULL)
91 idx = txq->stats.idx;
94 tmp.q_opackets[idx] += txq->stats.opackets;
95 tmp.q_obytes[idx] += txq->stats.obytes;
97 tmp.q_errors[idx] += txq->stats.oerrors;
100 tmp.opackets += txq->stats.opackets;
101 tmp.obytes += txq->stats.obytes;
103 tmp.oerrors += txq
[all...]
H A Dmlx5_txq.c83 (*txq_ctrl->txq.elts)[i] = NULL;
84 for (i = 0; (i != (1u << txq_ctrl->txq.wqe_n)); ++i) {
85 volatile struct mlx5_wqe64 *wqe = &(*txq_ctrl->txq.wqes)[i];
90 txq_ctrl->txq.elts_head = 0;
91 txq_ctrl->txq.elts_tail = 0;
92 txq_ctrl->txq.elts_comp = 0;
104 unsigned int elts_n = 1 << txq_ctrl->txq.elts_n;
105 unsigned int elts_head = txq_ctrl->txq.elts_head;
106 unsigned int elts_tail = txq_ctrl->txq.elts_tail;
107 struct rte_mbuf *(*elts)[elts_n] = txq_ctrl->txq
465 struct txq *txq = (*priv->txqs)[idx]; local
553 struct txq *txq = (struct txq *)dpdk_txq; local
598 struct txq *txq = dpdk_txq; local
[all...]
H A Dmlx5_mr.c170 * txq->mp2mr[]. If mp2mr[] is full, remove an entry first.
174 * @param txq
185 txq_mp2mr_reg(struct txq *txq, struct rte_mempool *mp, unsigned int idx) argument
187 struct txq_ctrl *txq_ctrl = container_of(txq, struct txq_ctrl, txq);
199 if (unlikely(idx == RTE_DIM(txq_ctrl->txq.mp2mr))) {
204 claim_zero(ibv_dereg_mr(txq_ctrl->txq.mp2mr[0].mr));
205 memmove(&txq_ctrl->txq.mp2mr[0], &txq_ctrl->txq
[all...]
/deb_dpdk/drivers/net/e1000/
H A Dem_rxtx.c223 em_set_xmit_ctx(struct em_tx_queue* txq, argument
281 txq->ctx_cache.flags = flags;
282 txq->ctx_cache.cmp_mask = cmp_mask;
283 txq->ctx_cache.hdrlen = hdrlen;
291 what_ctx_update(struct em_tx_queue *txq, uint64_t flags, argument
295 if (likely (txq->ctx_cache.flags == flags &&
296 ((txq->ctx_cache.hdrlen.data ^ hdrlen.data) &
297 txq->ctx_cache.cmp_mask) == 0))
306 em_xmit_cleanup(struct em_tx_queue *txq) argument
308 struct em_tx_entry *sw_ring = txq
378 struct em_tx_queue *txq; local
1072 em_tx_queue_release_mbufs(struct em_tx_queue *txq) argument
1087 em_tx_queue_release(struct em_tx_queue *txq) argument
1097 eth_em_tx_queue_release(void *txq) argument
1104 em_reset_tx_queue(struct em_tx_queue *txq) argument
1145 struct em_tx_queue *txq; local
1435 struct em_tx_queue *txq; local
1775 struct em_tx_queue *txq; local
1841 struct em_tx_queue *txq; local
[all...]
H A Digb_rxtx.c253 igbe_set_xmit_ctx(struct igb_tx_queue* txq, argument
263 ctx_curr = txq->ctx_curr;
264 ctx_idx = ctx_curr + txq->ctx_start;
320 txq->ctx_cache[ctx_curr].flags = ol_flags;
321 txq->ctx_cache[ctx_curr].tx_offload.data =
323 txq->ctx_cache[ctx_curr].tx_offload_mask = tx_offload_mask;
337 what_advctx_update(struct igb_tx_queue *txq, uint64_t flags, argument
341 if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) &&
342 (txq
387 struct igb_tx_queue *txq; local
1204 igb_tx_queue_release_mbufs(struct igb_tx_queue *txq) argument
1219 igb_tx_queue_release(struct igb_tx_queue *txq) argument
1229 eth_igb_tx_queue_release(void *txq) argument
1235 igb_reset_tx_queue_stat(struct igb_tx_queue *txq) argument
1245 igb_reset_tx_queue(struct igb_tx_queue *txq, struct rte_eth_dev *dev) argument
1286 struct igb_tx_queue *txq; local
1560 struct igb_tx_queue *txq; local
2264 struct igb_tx_queue *txq; local
2451 struct igb_tx_queue *txq; local
2515 struct igb_tx_queue *txq; local
[all...]
/deb_dpdk/drivers/net/bnx2x/
H A Dbnx2x_rxtx.c197 bnx2x_dev_tx_queue_release(void *txq) argument
199 bnx2x_tx_queue_release(txq);
205 struct bnx2x_tx_queue *txq; local
212 txq = p_txq;
213 sc = txq->sc;
214 fp = &sc->fp[txq->queue_id];
216 if ((unlikely((txq->nb_tx_desc - txq->nb_tx_avail) >
217 txq->tx_free_thresh)))
220 nb_tx_pkts = RTE_MIN(nb_pkts, txq
254 struct bnx2x_tx_queue *txq; local
472 struct bnx2x_tx_queue *txq = dev->data->tx_queues[i]; local
[all...]
/deb_dpdk/drivers/net/i40e/
H A Di40e_rxtx.c353 i40e_xmit_cleanup(struct i40e_tx_queue *txq) argument
355 struct i40e_tx_entry *sw_ring = txq->sw_ring;
356 volatile struct i40e_tx_desc *txd = txq->tx_ring;
357 uint16_t last_desc_cleaned = txq->last_desc_cleaned;
358 uint16_t nb_tx_desc = txq->nb_tx_desc;
362 desc_to_clean_to = (uint16_t)(last_desc_cleaned + txq->tx_rs_thresh);
372 txq->port_id, txq->queue_id);
385 txq->last_desc_cleaned = desc_to_clean_to;
386 txq
1028 struct i40e_tx_queue *txq; local
1256 i40e_tx_free_bufs(struct i40e_tx_queue *txq) argument
1322 i40e_tx_fill_hw_ring(struct i40e_tx_queue *txq, struct rte_mbuf **pkts, uint16_t nb_pkts) argument
1350 tx_xmit_pkts(struct i40e_tx_queue *txq, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) argument
1549 struct i40e_tx_queue *txq; local
1575 struct i40e_tx_queue *txq; local
1874 struct i40e_tx_queue *txq; local
2057 i40e_dev_tx_queue_release(void *txq) argument
2164 i40e_tx_queue_release_mbufs(struct i40e_tx_queue *txq) argument
2204 i40e_reset_tx_queue(struct i40e_tx_queue *txq) argument
2243 i40e_tx_queue_init(struct i40e_tx_queue *txq) argument
2515 struct i40e_tx_queue *txq; local
2647 struct i40e_tx_queue *txq; local
2755 i40e_set_tx_function_flag(struct rte_eth_dev *dev, struct i40e_tx_queue *txq) argument
2785 struct i40e_tx_queue *txq = local
2842 i40e_txq_vec_setup(struct i40e_tx_queue __rte_unused *txq) argument
[all...]
H A Di40e_rxtx_vec_common.h107 i40e_tx_free_bufs(struct i40e_tx_queue *txq) argument
116 if ((txq->tx_ring[txq->tx_next_dd].cmd_type_offset_bsz &
121 n = txq->tx_rs_thresh;
126 txep = &txq->sw_ring[txq->tx_next_dd - (n - 1)];
155 txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + txq->tx_rs_thresh);
156 txq
[all...]
H A Di40e_rxtx_vec_neon.c534 struct i40e_tx_queue *txq = (struct i40e_tx_queue *)tx_queue; local
543 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh);
545 if (txq->nb_tx_free < txq->tx_free_thresh)
546 i40e_tx_free_bufs(txq);
548 nb_commit = nb_pkts = (uint16_t)RTE_MIN(txq->nb_tx_free, nb_pkts);
552 tx_id = txq->tx_tail;
553 txdp = &txq->tx_ring[tx_id];
554 txep = &txq->sw_ring[tx_id];
556 txq
610 i40e_txq_vec_setup(struct i40e_tx_queue __rte_unused *txq) argument
[all...]
H A Di40e_rxtx_vec_sse.c550 struct i40e_tx_queue *txq = (struct i40e_tx_queue *)tx_queue; local
559 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh);
561 if (txq->nb_tx_free < txq->tx_free_thresh)
562 i40e_tx_free_bufs(txq);
564 nb_commit = nb_pkts = (uint16_t)RTE_MIN(txq->nb_tx_free, nb_pkts);
568 tx_id = txq->tx_tail;
569 txdp = &txq->tx_ring[tx_id];
570 txep = &txq->sw_ring[tx_id];
572 txq
626 i40e_txq_vec_setup(struct i40e_tx_queue __rte_unused *txq) argument
[all...]
/deb_dpdk/drivers/net/vmxnet3/
H A Dvmxnet3_rxtx.c121 vmxnet3_txq_dump(struct vmxnet3_tx_queue *txq) argument
125 if (txq == NULL)
129 txq->cmd_ring.base, txq->comp_ring.base, txq->data_ring.base);
131 (unsigned long)txq->cmd_ring.basePA,
132 (unsigned long)txq->comp_ring.basePA,
133 (unsigned long)txq->data_ring.basePA);
135 avail = vmxnet3_cmd_ring_desc_avail(&txq->cmd_ring);
137 (uint32_t)txq
186 vmxnet3_dev_tx_queue_release(void *txq) argument
226 vmxnet3_dev_tx_queue_reset(void *txq) argument
294 struct vmxnet3_tx_queue *txq = dev->data->tx_queues[i]; local
313 vmxnet3_unmap_pkt(uint16_t eop_idx, vmxnet3_tx_queue_t *txq) argument
342 vmxnet3_tq_tx_complete(vmxnet3_tx_queue_t *txq) argument
365 vmxnet3_tx_queue_t *txq = tx_queue; local
842 struct vmxnet3_tx_queue *txq; local
1088 struct vmxnet3_tx_queue *txq = dev->data->tx_queues[i]; local
[all...]
/deb_dpdk/examples/ethtool/ethtool-app/
H A Dmain.c65 struct txq_port txq; member in struct:app_port
195 struct txq_port *txq; local
214 txq = &ptr_port->txq;
226 &txq->buf_frames[txq->cnt_unsent],
227 RTE_DIM(txq->buf_frames) - txq->cnt_unsent
233 ptr_frame = txq->buf_frames[
234 idx_frame + txq
[all...]
/deb_dpdk/drivers/net/qede/
H A Dqede_rxtx.c68 static void qede_tx_queue_release_mbufs(struct qede_tx_queue *txq) argument
72 PMD_TX_LOG(DEBUG, txq, "releasing %u mbufs\n", txq->nb_tx_desc);
74 if (txq->sw_tx_ring) {
75 for (i = 0; i < txq->nb_tx_desc; i++) {
76 if (txq->sw_tx_ring[i].mbuf) {
77 rte_pktmbuf_free(txq->sw_tx_ring[i].mbuf);
78 txq->sw_tx_ring[i].mbuf = NULL;
228 struct qede_tx_queue *txq = tx_queue; local
230 if (txq !
250 struct qede_tx_queue *txq; local
332 int fp_rx = qdev->fp_num_rx, rxq = 0, txq = 0; local
588 struct qede_tx_queue *txq; local
1055 qede_free_tx_pkt(struct ecore_dev *edev, struct qede_tx_queue *txq) argument
1084 qede_process_tx_compl(struct ecore_dev *edev, struct qede_tx_queue *txq) argument
1115 struct qede_tx_queue *txq = p_txq; local
1161 struct qede_tx_queue *txq = p_txq; local
1265 int rxq = 0, txq = 0; local
1324 qede_drain_txq(struct qede_dev *qdev, struct qede_tx_queue *txq, bool allow_drain) argument
1391 struct qede_tx_queue *txq = fp->txqs[tc]; local
1447 struct qede_tx_queue *txq; local
[all...]
/deb_dpdk/drivers/net/fm10k/
H A Dfm10k_rxtx_vec.c48 fm10k_reset_tx_queue(struct fm10k_tx_queue *txq);
706 fm10k_txq_vec_setup(struct fm10k_tx_queue *txq) argument
708 txq->ops = &vec_txq_ops;
712 fm10k_tx_vec_condition_check(struct fm10k_tx_queue *txq) argument
715 if ((txq->txq_flags & FM10K_SIMPLE_TX_FLAG) != FM10K_SIMPLE_TX_FLAG)
718 if (txq->tx_ftag_en)
745 fm10k_tx_free_bufs(struct fm10k_tx_queue *txq) argument
755 flags = txq->hw_ring[txq->next_dd].flags;
759 n = txq
814 struct fm10k_tx_queue *txq = (struct fm10k_tx_queue *)tx_queue; local
875 fm10k_reset_tx_queue(struct fm10k_tx_queue *txq) argument
[all...]
H A Dfm10k.h91 #define FM10K_TX_FREE_THRESH_DEFAULT(txq) 32
92 #define FM10K_TX_FREE_THRESH_MIN(txq) 1
93 #define FM10K_TX_FREE_THRESH_MAX(txq) ((txq)->nb_desc - 3)
94 #define FM10K_TX_FREE_THRESH_DIV(txq) 0
104 #define FM10K_TX_RS_THRESH_DEFAULT(txq) 32
105 #define FM10K_TX_RS_THRESH_MIN(txq) 1
106 #define FM10K_TX_RS_THRESH_MAX(txq) \
107 RTE_MIN(((txq)->nb_desc - 2), (txq)
[all...]
/deb_dpdk/drivers/net/nfp/
H A Dnfp_net.c86 static int nfp_net_tx_free_bufs(struct nfp_net_txq *txq);
87 static void nfp_net_tx_queue_release(void *txq);
317 nfp_net_tx_queue_release_mbufs(struct nfp_net_txq *txq) argument
321 if (txq->txbufs == NULL)
324 for (i = 0; i < txq->tx_count; i++) {
325 if (txq->txbufs[i].mbuf) {
326 rte_pktmbuf_free_seg(txq->txbufs[i].mbuf);
327 txq->txbufs[i].mbuf = NULL;
335 struct nfp_net_txq *txq = tx_queue; local
337 if (txq) {
345 nfp_net_reset_tx_queue(struct nfp_net_txq *txq) argument
1422 struct nfp_net_txq *txq; local
1531 nfp_net_tx_cksum(struct nfp_net_txq *txq, struct nfp_net_tx_desc *txd, struct rte_mbuf *mb) argument
1878 nfp_net_tx_free_bufs(struct nfp_net_txq *txq) argument
1926 nfp_net_txq_full(struct nfp_net_txq *txq) argument
1934 struct nfp_net_txq *txq; local
[all...]
/deb_dpdk/drivers/net/cxgbe/
H A Dsge.c71 struct sge_eth_txq *txq);
809 struct sge_eth_txq *txq = &adap->sge.ethtxq[0]; local
814 for (i = 0; i < adap->sge.max_ethqsets; i++, txq++) {
815 if (t4_os_trylock(&txq->txq_lock)) {
816 coal_idx = txq->q.coalesce.idx;
818 if (coal_idx == txq->q.last_coal_idx &&
819 txq->q.pidx == txq->q.last_pidx) {
820 ship_tx_pkt_coalesce_wr(adap, txq);
822 txq
839 ship_tx_pkt_coalesce_wr(struct adapter *adap, struct sge_eth_txq *txq) argument
885 should_tx_packet_coalesce(struct sge_eth_txq *txq, struct rte_mbuf *mbuf, unsigned int *nflits, struct adapter *adap) argument
972 tx_do_packet_coalesce(struct sge_eth_txq *txq, struct rte_mbuf *mbuf, int flits, struct adapter *adap, const struct port_info *pi, dma_addr_t *addr) argument
1067 t4_eth_xmit(struct sge_eth_txq *txq, struct rte_mbuf *mbuf) argument
1860 t4_sge_eth_txq_start(struct sge_eth_txq *txq) argument
1875 t4_sge_eth_txq_stop(struct sge_eth_txq *txq) argument
1882 t4_sge_alloc_eth_txq(struct adapter *adap, struct sge_eth_txq *txq, struct rte_eth_dev *eth_dev, uint16_t queue_id, unsigned int iqid, int socket_id) argument
1988 struct sge_eth_txq *txq = &adap->sge.ethtxq[pi->first_qset]; local
2014 t4_sge_eth_txq_release(struct adapter *adap, struct sge_eth_txq *txq) argument
2046 struct sge_eth_txq *txq = &adap->sge.ethtxq[0]; local
[all...]
/deb_dpdk/drivers/net/mlx4/
H A Dmlx4.c218 struct txq { struct
286 struct txq *(*txqs)[]; /* TX queues. */
699 txq_setup(struct rte_eth_dev *dev, struct txq *txq, uint16_t desc,
703 txq_cleanup(struct txq *txq);
902 struct txq *primary_txq = (*sd->primary_priv->txqs)[i];
903 struct txq *txq; local
907 txq
989 txq_alloc_elts(struct txq *txq, unsigned int elts_n) argument
1054 txq_free_elts(struct txq *txq) argument
1104 txq_cleanup(struct txq *txq) argument
1172 txq_complete(struct txq *txq) argument
1354 txq_mp2mr(struct txq *txq, struct rte_mempool *mp) argument
1440 struct txq *txq = arg; local
1511 tx_burst_sg(struct txq *txq, unsigned int segs, struct txq_elt *elt, struct rte_mbuf *buf, unsigned int elts_head, struct ibv_sge (*sges)[MLX4_PMD_SGE_WR_N]) argument
1618 struct txq *txq = (struct txq *)dpdk_txq; local
1805 struct txq *txq = dpdk_txq; local
1843 txq_setup(struct rte_eth_dev *dev, struct txq *txq, uint16_t desc, unsigned int socket, const struct rte_eth_txconf *conf) argument
2039 struct txq *txq = (*priv->txqs)[idx]; local
2095 struct txq *txq = (struct txq *)dpdk_txq; local
4541 struct txq *txq = (*priv->txqs)[i]; local
[all...]

Completed in 30 milliseconds

1234