Home
last modified time | relevance | path

Searched refs:txrx_peer (Results 1 – 25 of 27) sorted by relevance

12

/wlan-driver/qca-wifi-host-cmn/dp/wifi3.0/rh/
H A Ddp_rh_rx.c124 struct dp_txrx_peer *txrx_peer, in dp_rx_mec_check_wrapper() argument
128 return dp_rx_mcast_echo_check(soc, txrx_peer, rx_tlv_hdr, nbuf); in dp_rx_mec_check_wrapper()
132 struct dp_txrx_peer *txrx_peer, in dp_rx_mec_check_wrapper() argument
347 struct dp_txrx_peer *txrx_peer, in dp_rx_deliver_to_osif_stack_rh() argument
353 dp_rx_eapol_deliver_to_stack(soc, vdev, txrx_peer, nbuf, NULL); in dp_rx_deliver_to_osif_stack_rh()
355 dp_rx_deliver_to_stack(soc, vdev, txrx_peer, nbuf, NULL); in dp_rx_deliver_to_osif_stack_rh()
361 struct dp_txrx_peer *txrx_peer, in dp_rx_deliver_to_osif_stack_rh() argument
366 dp_rx_deliver_to_stack(soc, vdev, txrx_peer, nbuf, NULL); in dp_rx_deliver_to_osif_stack_rh()
377 struct dp_txrx_peer *txrx_peer = NULL; in dp_rx_decrypt_unecrypt_err_handler_rh() local
408 txrx_peer = dp_tgt_txrx_peer_get_ref_by_id(soc, peer_id, in dp_rx_decrypt_unecrypt_err_handler_rh()
[all …]
/wlan-driver/qca-wifi-host-cmn/dp/wifi3.0/
H A Ddp_rx_defrag.c100 static void dp_rx_clear_saved_desc_info(struct dp_txrx_peer *txrx_peer, in dp_rx_clear_saved_desc_info() argument
103 if (txrx_peer->rx_tid[tid].dst_ring_desc) in dp_rx_clear_saved_desc_info()
104 qdf_mem_free(txrx_peer->rx_tid[tid].dst_ring_desc); in dp_rx_clear_saved_desc_info()
106 txrx_peer->rx_tid[tid].dst_ring_desc = NULL; in dp_rx_clear_saved_desc_info()
107 txrx_peer->rx_tid[tid].head_frag_desc = NULL; in dp_rx_clear_saved_desc_info()
110 static void dp_rx_return_head_frag_desc(struct dp_txrx_peer *txrx_peer, in dp_rx_return_head_frag_desc() argument
121 pdev = txrx_peer->vdev->pdev; in dp_rx_return_head_frag_desc()
124 if (txrx_peer->rx_tid[tid].head_frag_desc) { in dp_rx_return_head_frag_desc()
125 pool_id = txrx_peer->rx_tid[tid].head_frag_desc->pool_id; in dp_rx_return_head_frag_desc()
130 txrx_peer->rx_tid[tid].head_frag_desc); in dp_rx_return_head_frag_desc()
[all …]
H A Ddp_rx_err.c61 struct dp_txrx_peer *txrx_peer, in dp_rx_mcast_echo_check() argument
65 struct dp_vdev *vdev = txrx_peer->vdev; in dp_rx_mcast_echo_check()
129 if (ase && (ase->peer_id != txrx_peer->peer_id)) { in dp_rx_mcast_echo_check()
426 struct dp_txrx_peer *txrx_peer; in dp_rx_pn_error_handle() local
434 txrx_peer = dp_tgt_txrx_peer_get_ref_by_id(soc, peer_id, in dp_rx_pn_error_handle()
438 if (qdf_likely(txrx_peer)) { in dp_rx_pn_error_handle()
443 txrx_peer); in dp_rx_pn_error_handle()
474 struct dp_txrx_peer *txrx_peer, in dp_rx_deliver_oor_frame() argument
500 dp_rx_deliver_to_stack(soc, txrx_peer->vdev, txrx_peer, nbuf, NULL); in dp_rx_deliver_oor_frame()
507 struct dp_txrx_peer *txrx_peer, in dp_rx_deliver_oor_frame() argument
[all …]
H A Ddp_rx.c1125 struct dp_txrx_peer *txrx_peer, uint8_t link_id) in dp_rx_deliver_raw() argument
1138 DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer, rx.raw, 1, in dp_rx_deliver_raw()
1308 struct dp_txrx_peer *txrx_peer) in dp_rx_fill_mesh_stats() argument
1342 peer = dp_peer_get_ref_by_id(soc, txrx_peer->peer_id, DP_MOD_ID_MESH); in dp_rx_fill_mesh_stats()
1359 txrx_peer->peer_id, in dp_rx_fill_mesh_stats()
2036 struct dp_txrx_peer *txrx_peer, qdf_nbuf_t nbuf_head) in dp_rx_deliver_to_stack_ext() argument
2051 if (!txrx_peer->wds_ext.init) in dp_rx_deliver_to_stack_ext()
2054 if (txrx_peer->osif_rx) in dp_rx_deliver_to_stack_ext()
2055 txrx_peer->osif_rx(txrx_peer->wds_ext.osif_peer, nbuf_head); in dp_rx_deliver_to_stack_ext()
2065 struct dp_txrx_peer *txrx_peer, qdf_nbuf_t nbuf_head) in dp_rx_deliver_to_stack_ext() argument
[all …]
H A Ddp_tx.c2503 struct dp_txrx_peer *txrx_peer, in dp_tx_latency_stats_update() argument
2511 struct dp_vdev *vdev = txrx_peer->vdev; in dp_tx_latency_stats_update()
2532 tx_latency = &txrx_peer->stats[link_id].tx_latency; in dp_tx_latency_stats_update()
2534 dp_tx_latency_stats_update_bucket(txrx_peer->vdev, tx_latency, idx, in dp_tx_latency_stats_update()
2537 dp_tx_latency_stats_update_bucket(txrx_peer->vdev, tx_latency, idx, in dp_tx_latency_stats_update()
2540 dp_tx_latency_stats_update_bucket(txrx_peer->vdev, tx_latency, idx, in dp_tx_latency_stats_update()
2608 struct dp_txrx_peer *txrx_peer; in dp_tx_latency_stats_update_cca() local
2636 txrx_peer = dp_get_txrx_peer(peer); in dp_tx_latency_stats_update_cca()
2637 if (qdf_unlikely(!txrx_peer)) { in dp_tx_latency_stats_update_cca()
2644 if (link_id >= txrx_peer->stats_arr_size) in dp_tx_latency_stats_update_cca()
[all …]
H A Ddp_peer.c233 struct dp_txrx_peer *txrx_peer; in dp_peer_check_wds_ext_peer() local
238 txrx_peer = dp_get_txrx_peer(peer); in dp_peer_check_wds_ext_peer()
239 if (!txrx_peer) in dp_peer_check_wds_ext_peer()
243 &txrx_peer->wds_ext.init)) in dp_peer_check_wds_ext_peer()
696 struct dp_txrx_peer *txrx_peer) in dp_txrx_peer_attach_add() argument
700 peer->txrx_peer = txrx_peer; in dp_txrx_peer_attach_add()
701 txrx_peer->bss_peer = peer->bss_peer; in dp_txrx_peer_attach_add()
708 txrx_peer->peer_id = peer->peer_id; in dp_txrx_peer_attach_add()
734 if (peer->txrx_peer) in dp_peer_find_id_to_obj_add()
735 peer->txrx_peer->peer_id = peer_id; in dp_peer_find_id_to_obj_add()
[all …]
H A Ddp_txrx_wds.c409 peer->txrx_peer->wds_ecm.wds_rx_filter = 1; in dp_txrx_set_wds_rx_policy()
410 peer->txrx_peer->wds_ecm.wds_rx_ucast_4addr = in dp_txrx_set_wds_rx_policy()
412 peer->txrx_peer->wds_ecm.wds_rx_mcast_4addr = in dp_txrx_set_wds_rx_policy()
439 if (!peer->txrx_peer) { in dp_txrx_peer_wds_tx_policy_update()
445 peer->txrx_peer->wds_enabled = 1; in dp_txrx_peer_wds_tx_policy_update()
446 peer->txrx_peer->wds_ecm.wds_tx_ucast_4addr = wds_tx_ucast; in dp_txrx_peer_wds_tx_policy_update()
447 peer->txrx_peer->wds_ecm.wds_tx_mcast_4addr = wds_tx_mcast; in dp_txrx_peer_wds_tx_policy_update()
449 peer->txrx_peer->wds_enabled = 0; in dp_txrx_peer_wds_tx_policy_update()
450 peer->txrx_peer->wds_ecm.wds_tx_ucast_4addr = 0; in dp_txrx_peer_wds_tx_policy_update()
451 peer->txrx_peer->wds_ecm.wds_tx_mcast_4addr = 0; in dp_txrx_peer_wds_tx_policy_update()
[all …]
H A Ddp_peer.h350 if (!peer->txrx_peer) { in dp_txrx_peer_get_ref_by_id()
356 return peer->txrx_peer; in dp_txrx_peer_get_ref_by_id()
1205 struct dp_txrx_peer *txrx_peer);
1217 struct dp_txrx_peer *txrx_peer);
1225 void dp_peer_delay_stats_ctx_clr(struct dp_txrx_peer *txrx_peer);
1229 struct dp_txrx_peer *txrx_peer) in dp_peer_delay_stats_ctx_alloc() argument
1236 struct dp_txrx_peer *txrx_peer) in dp_peer_delay_stats_ctx_dealloc() argument
1241 void dp_peer_delay_stats_ctx_clr(struct dp_txrx_peer *txrx_peer) in dp_peer_delay_stats_ctx_clr() argument
1255 struct dp_txrx_peer *txrx_peer);
1265 struct dp_txrx_peer *txrx_peer);
[all …]
H A Ddp_rx_defrag.h175 dp_rx_defrag_fraglist_insert(struct dp_txrx_peer *txrx_peer, unsigned int tid,
188 void dp_rx_defrag_waitlist_add(struct dp_txrx_peer *txrx_peer,
202 QDF_STATUS dp_rx_defrag(struct dp_txrx_peer *txrx_peer, unsigned int tid,
225 void dp_rx_reorder_flush_frag(struct dp_txrx_peer *txrx_peer,
237 void dp_rx_defrag_waitlist_remove(struct dp_txrx_peer *txrx_peer,
247 void dp_rx_defrag_cleanup(struct dp_txrx_peer *txrx_peer, unsigned int tid);
H A Ddp_rx.h1526 struct dp_txrx_peer *txrx_peer, in dp_rx_wds_srcport_learn() argument
1833 struct dp_txrx_peer *txrx_peer);
2089 bool dp_rx_multipass_process(struct dp_txrx_peer *txrx_peer, qdf_nbuf_t nbuf,
2264 struct dp_txrx_peer *txrx_peer,
2469 dp_rx_set_nbuf_band(qdf_nbuf_t nbuf, struct dp_txrx_peer *txrx_peer, in dp_rx_set_nbuf_band() argument
2472 qdf_nbuf_rx_set_band(nbuf, txrx_peer->band[link_id]); in dp_rx_set_nbuf_band()
2482 dp_rx_set_nbuf_band(qdf_nbuf_t nbuf, struct dp_txrx_peer *txrx_peer, in dp_rx_set_nbuf_band() argument
2882 struct dp_txrx_peer *txrx_peer = NULL; in dp_rx_get_txrx_peer_and_vdev() local
2884 txrx_peer = dp_txrx_peer_get_ref_by_id(soc, peer_id, txrx_ref_handle, in dp_rx_get_txrx_peer_and_vdev()
2887 if (qdf_likely(txrx_peer)) { in dp_rx_get_txrx_peer_and_vdev()
[all …]
H A Ddp_main.c1194 struct dp_txrx_peer *txrx_peer = NULL; in dp_print_peer_info() local
1196 txrx_peer = dp_get_txrx_peer(peer); in dp_print_peer_info()
1197 if (!txrx_peer) in dp_print_peer_info()
1209 txrx_peer->nawds_enabled, in dp_print_peer_info()
1210 txrx_peer->bss_peer, in dp_print_peer_info()
1211 txrx_peer->wds_enabled, in dp_print_peer_info()
5446 struct dp_txrx_peer *txrx_peer) in dp_peer_hw_txrx_stats_init() argument
5448 txrx_peer->hw_txrx_stats_en = in dp_peer_hw_txrx_stats_init()
5454 struct dp_txrx_peer *txrx_peer) in dp_peer_hw_txrx_stats_init() argument
5456 txrx_peer->hw_txrx_stats_en = 0; in dp_peer_hw_txrx_stats_init()
[all …]
H A Ddp_stats.c4885 struct dp_txrx_peer *txrx_peer, in dp_vdev_peer_stats_update_protocol_cnt() argument
4913 if (!txrx_peer) { in dp_vdev_peer_stats_update_protocol_cnt()
4920 txrx_peer = peer->txrx_peer; in dp_vdev_peer_stats_update_protocol_cnt()
4921 if (!txrx_peer) in dp_vdev_peer_stats_update_protocol_cnt()
4924 per_pkt_stats = &txrx_peer->stats[0].per_pkt_stats; in dp_vdev_peer_stats_update_protocol_cnt()
6814 if (!peer->txrx_peer || !peer->txrx_peer->jitter_stats) in dp_print_jitter_stats()
6820 &peer->txrx_peer->jitter_stats[tid]; in dp_print_jitter_stats()
7018 if (!peer || !peer->txrx_peer) in dp_peer_print_tx_delay_stats()
7028 delay_stats = peer->txrx_peer->delay_stats; in dp_peer_print_tx_delay_stats()
7065 if (!peer || !peer->txrx_peer) in dp_peer_print_rx_delay_stats()
[all …]
H A Ddp_rx_tid.c621 struct dp_txrx_peer *txrx_peer; in dp_single_rx_tid_setup() local
690 txrx_peer = dp_get_txrx_peer(peer); in dp_single_rx_tid_setup()
696 switch (txrx_peer->security[dp_sec_ucast].sec_type) { in dp_single_rx_tid_setup()
1196 rx_tid_defrag = &peer->txrx_peer->rx_tid[tid]; in dp_peer_rx_tids_init()
1205 rx_tid_defrag->defrag_peer = peer->txrx_peer; in dp_peer_rx_tids_init()
1232 rx_tid_defrag = &peer->txrx_peer->rx_tid[tid]; in dp_peer_rx_tids_init()
1244 rx_tid_defrag->defrag_peer = peer->txrx_peer; in dp_peer_rx_tids_init()
1252 struct dp_txrx_peer *txrx_peer = dp_get_txrx_peer(peer); in dp_peer_rx_tid_setup() local
1268 qdf_unlikely(txrx_peer->nawds_enabled)) in dp_peer_rx_tid_setup()
1287 if (!peer->txrx_peer) in dp_peer_rx_cleanup()
[all …]
H A Ddp_tx.h354 struct dp_txrx_peer *txrx_peer,
369 struct dp_txrx_peer *txrx_peer);
416 void dp_tx_update_peer_basic_stats(struct dp_txrx_peer *txrx_peer,
1309 struct dp_txrx_peer *txrx_peer,
2161 dp_tx_set_nbuf_band(qdf_nbuf_t nbuf, struct dp_txrx_peer *txrx_peer, in dp_tx_set_nbuf_band() argument
2164 qdf_nbuf_tx_set_band(nbuf, txrx_peer->band[link_id]); in dp_tx_set_nbuf_band()
2168 dp_tx_set_nbuf_band(qdf_nbuf_t nbuf, struct dp_txrx_peer *txrx_peer, in dp_tx_set_nbuf_band() argument
H A Ddp_ipa.c3614 struct dp_txrx_peer *txrx_peer; in dp_ipa_rx_wdsext_iface() local
3619 txrx_peer = dp_tgt_txrx_peer_get_ref_by_id(soc_hdl, peer_id, in dp_ipa_rx_wdsext_iface()
3623 if (qdf_likely(txrx_peer)) { in dp_ipa_rx_wdsext_iface()
3624 if (dp_rx_deliver_to_stack_ext(dp_soc, txrx_peer->vdev, in dp_ipa_rx_wdsext_iface()
3625 txrx_peer, skb) in dp_ipa_rx_wdsext_iface()
3652 if (qdf_unlikely(!vdev_peer->txrx_peer)) {
3661 DP_PEER_PER_PKT_STATS_INC_PKT(vdev_peer->txrx_peer,
3668 DP_PEER_PER_PKT_STATS_INC_PKT(vdev_peer->txrx_peer,
4102 struct dp_txrx_peer *txrx_peer;
4109 txrx_peer = dp_get_txrx_peer(peer);
[all …]
H A Ddp_txrx_wds.h186 &ta_peer->txrx_peer->wds_ext.init)) { in dp_wds_ext_peer_learn()
H A Ddp_internal.h1497 static inline void dp_set_peer_isolation(struct dp_txrx_peer *txrx_peer, in dp_set_peer_isolation() argument
1500 txrx_peer->isolation = val; in dp_set_peer_isolation()
1514 static inline void dp_wds_ext_peer_init(struct dp_txrx_peer *txrx_peer) in dp_wds_ext_peer_init() argument
1516 txrx_peer->wds_ext.osif_peer = NULL; in dp_wds_ext_peer_init()
1517 txrx_peer->wds_ext.init = 0; in dp_wds_ext_peer_init()
1520 static inline void dp_wds_ext_peer_init(struct dp_txrx_peer *txrx_peer) in dp_wds_ext_peer_init() argument
2690 struct dp_txrx_peer *txrx_peer);
3929 struct dp_txrx_peer *txrx_peer,
3955 #define dp_vdev_peer_stats_update_protocol_cnt(vdev, nbuf, txrx_peer, \ argument
5794 void dp_update_vdev_be_basic_stats(struct dp_txrx_peer *txrx_peer,
[all …]
H A Ddp_rings_main.c2516 dp_peer_get_local_link_id(struct dp_peer *peer, struct dp_txrx_peer *txrx_peer) in dp_peer_get_local_link_id() argument
2519 &txrx_peer->ll_id_peer_map[0]; in dp_peer_get_local_link_id()
2562 struct dp_txrx_peer *txrx_peer; in dp_peer_set_local_link_id() local
2567 txrx_peer = dp_get_txrx_peer(peer); in dp_peer_set_local_link_id()
2568 if (txrx_peer) in dp_peer_set_local_link_id()
2570 txrx_peer); in dp_peer_set_local_link_id()
2573 QDF_MAC_ADDR_REF(peer->mac_addr.raw), txrx_peer, in dp_peer_set_local_link_id()
/wlan-driver/qca-wifi-host-cmn/dp/wifi3.0/li/
H A Ddp_li_rx.c71 struct dp_txrx_peer *txrx_peer, in dp_rx_mec_check_wrapper() argument
75 return dp_rx_mcast_echo_check(soc, txrx_peer, rx_tlv_hdr, nbuf); in dp_rx_mec_check_wrapper()
79 struct dp_txrx_peer *txrx_peer, in dp_rx_mec_check_wrapper() argument
210 struct dp_txrx_peer *txrx_peer; in dp_rx_process_li() local
269 txrx_peer = NULL; in dp_rx_process_li()
549 if (qdf_likely(txrx_peer)) in dp_rx_process_li()
574 if (dp_rx_is_list_ready(deliver_list_head, vdev, txrx_peer, in dp_rx_process_li()
576 dp_rx_deliver_to_stack(soc, vdev, txrx_peer, in dp_rx_process_li()
594 if (qdf_unlikely(!txrx_peer)) { in dp_rx_process_li()
595 txrx_peer = in dp_rx_process_li()
[all …]
H A Ddp_li_tx.c101 struct dp_txrx_peer *txrx_peer; in dp_tx_process_htt_completion_li() local
222 txrx_peer = dp_txrx_peer_get_ref_by_id(soc, ts.peer_id, in dp_tx_process_htt_completion_li()
225 if (qdf_likely(txrx_peer)) { in dp_tx_process_htt_completion_li()
226 DP_PEER_STATS_FLAT_INC_PKT(txrx_peer, comp_pkt, 1, in dp_tx_process_htt_completion_li()
229 DP_PEER_STATS_FLAT_INC(txrx_peer, tx_failed, 1); in dp_tx_process_htt_completion_li()
232 dp_tx_comp_process_tx_status(soc, tx_desc, &ts, txrx_peer, in dp_tx_process_htt_completion_li()
234 dp_tx_comp_process_desc(soc, tx_desc, &ts, txrx_peer); in dp_tx_process_htt_completion_li()
237 if (qdf_likely(txrx_peer)) in dp_tx_process_htt_completion_li()
H A Ddp_li_rx.h111 struct dp_txrx_peer *txrx_peer, in dp_rx_get_reo_qdesc_addr_li() argument
354 struct dp_txrx_peer *txrx_peer,
/wlan-driver/qca-wifi-host-cmn/dp/wifi3.0/be/
H A Ddp_be_rx.c108 struct dp_txrx_peer *txrx_peer, in dp_rx_wds_learn() argument
118 txrx_peer, in dp_rx_wds_learn()
298 struct dp_txrx_peer *txrx_peer; in dp_rx_process_be() local
358 txrx_peer = NULL; in dp_rx_process_be()
628 if (qdf_likely(txrx_peer)) in dp_rx_process_be()
654 if (dp_rx_is_list_ready(deliver_list_head, vdev, txrx_peer, in dp_rx_process_be()
656 dp_rx_deliver_to_stack(soc, vdev, txrx_peer, in dp_rx_process_be()
673 if (qdf_unlikely(!txrx_peer)) { in dp_rx_process_be()
674 txrx_peer = dp_rx_get_txrx_peer_and_vdev(soc, nbuf, in dp_rx_process_be()
681 if (qdf_unlikely(!txrx_peer) || qdf_unlikely(!vdev)) { in dp_rx_process_be()
[all …]
H A Ddp_be_tx.c272 struct dp_txrx_peer *txrx_peer; in dp_tx_process_htt_completion_be() local
397 txrx_peer = dp_txrx_peer_get_ref_by_id(soc, peer_id, in dp_tx_process_htt_completion_be()
400 if (qdf_likely(txrx_peer)) in dp_tx_process_htt_completion_be()
402 txrx_peer, in dp_tx_process_htt_completion_be()
407 dp_tx_comp_process_tx_status(soc, tx_desc, &ts, txrx_peer, in dp_tx_process_htt_completion_be()
409 dp_tx_comp_process_desc(soc, tx_desc, &ts, txrx_peer); in dp_tx_process_htt_completion_be()
412 if (qdf_likely(txrx_peer)) in dp_tx_process_htt_completion_be()
588 struct dp_txrx_peer *txrx_peer = NULL; in dp_tx_mlo_mcast_multipass_lookup() local
601 TAILQ_FOREACH(txrx_peer, &ptnr_vdev->mpass_peer_list, in dp_tx_mlo_mcast_multipass_lookup()
603 if (vlan_id == txrx_peer->vlan_id) { in dp_tx_mlo_mcast_multipass_lookup()
[all …]
H A Ddp_be_rx.h647 struct dp_txrx_peer *txrx_peer, in dp_rx_get_reo_qdesc_addr_be() argument
654 qdesc_addr = (uint64_t)txrx_peer->peer_id; in dp_rx_get_reo_qdesc_addr_be()
656 peer = dp_peer_get_ref_by_id(txrx_peer->vdev->pdev->soc, in dp_rx_get_reo_qdesc_addr_be()
657 txrx_peer->peer_id, in dp_rx_get_reo_qdesc_addr_be()
672 struct dp_txrx_peer *txrx_peer, in dp_rx_get_reo_qdesc_addr_be() argument
728 struct dp_txrx_peer *txrx_peer,
/wlan-driver/qca-wifi-host-cmn/dp/wifi3.0/monitor/
H A Ddp_mon.c1467 struct dp_txrx_peer *txrx_peer = NULL; in dp_peer_get_tx_rx_stats() local
1473 txrx_peer = tgt_peer->txrx_peer; in dp_peer_get_tx_rx_stats()
1474 peer_stats_intf->rx_packet_count = txrx_peer->to_stack.num; in dp_peer_get_tx_rx_stats()
1475 peer_stats_intf->rx_byte_count = txrx_peer->to_stack.bytes; in dp_peer_get_tx_rx_stats()
1476 stats_arr_size = txrx_peer->stats_arr_size; in dp_peer_get_tx_rx_stats()
1480 txrx_peer->stats[inx].per_pkt_stats.tx.ucast.num; in dp_peer_get_tx_rx_stats()
1482 txrx_peer->stats[inx].per_pkt_stats.tx.tx_success.bytes; in dp_peer_get_tx_rx_stats()
1492 struct dp_txrx_peer *txrx_peer = NULL; in dp_peer_stats_notify() local
1501 txrx_peer = tgt_peer->txrx_peer; in dp_peer_stats_notify()
1502 if (!qdf_unlikely(txrx_peer)) in dp_peer_stats_notify()
[all …]

12