Lines Matching refs:txrx_peer
621 struct dp_txrx_peer *txrx_peer; in dp_single_rx_tid_setup() local
690 txrx_peer = dp_get_txrx_peer(peer); in dp_single_rx_tid_setup()
696 switch (txrx_peer->security[dp_sec_ucast].sec_type) { in dp_single_rx_tid_setup()
1196 rx_tid_defrag = &peer->txrx_peer->rx_tid[tid]; in dp_peer_rx_tids_init()
1205 rx_tid_defrag->defrag_peer = peer->txrx_peer; in dp_peer_rx_tids_init()
1232 rx_tid_defrag = &peer->txrx_peer->rx_tid[tid]; in dp_peer_rx_tids_init()
1244 rx_tid_defrag->defrag_peer = peer->txrx_peer; in dp_peer_rx_tids_init()
1252 struct dp_txrx_peer *txrx_peer = dp_get_txrx_peer(peer); in dp_peer_rx_tid_setup() local
1268 qdf_unlikely(txrx_peer->nawds_enabled)) in dp_peer_rx_tid_setup()
1287 if (!peer->txrx_peer) in dp_peer_rx_cleanup()
1295 &peer->txrx_peer->rx_tid[tid]; in dp_peer_rx_cleanup()
1300 dp_rx_defrag_waitlist_remove(peer->txrx_peer, tid); in dp_peer_rx_cleanup()
1301 dp_rx_reorder_flush_frag(peer->txrx_peer, tid); in dp_peer_rx_cleanup()