ab
rvb = diff(rvb, ab);
rvb = sum(rvb, ab);
Bfree(ab);
Bigint *ab, *bb, *bb1, *bd, *bd0, *bs, *delta, *rvb, *rvb0;
ab = d2b(dval(&adj), &abe, &abits);
if (ab == NULL)
rshift(ab, -abe);
ab = lshift(ab, abe);
if (ab == NULL)
ab();
ab();
ab();
ab();
ab(void)
BB ab("local");
BB ab("local");
int64_t ab = (int64_t)b ^ (int64_t)a; \
c ^= ab & (c ^ b); \
c &= ab; \
struct pfr_buffer ab;
bzero(&ab, sizeof(ab));
ab.pfrb_type = PFRB_ADDRS;
if (pfr_buf_load(&ab, ti->file, 0, popts)) {
if (append_addr_host(&ab, ti->host, 0, 0)) {
pf->anchor->path, &ab, pf->anchor->ruleset.tticket, ukt)) {
pfr_buf_clear(&ab);
pfr_buf_clear(&ab);
struct pfr_buffer *ab, u_int32_t ticket, struct pfr_uktable *ukt)
if (ab->pfrb_size != 0) {
ukt->pfrukt_addrs = *ab;
ab->pfrb_size = 0;
ab->pfrb_msize = 0;
ab->pfrb_caddr = NULL;
return pfr_ina_define(tbl, ab->pfrb_caddr, ab->pfrb_size, NULL,
ath11k_dp_tx_pending_cleanup, ab);
spin_lock(&ab->base_lock);
if (test_bit(ATH11K_FLAG_RECOVERY, &ab->dev_flags) &&
ab->default_regd[pdev_idx]) {
spin_unlock(&ab->base_lock);
spin_unlock(&ab->base_lock);
if (pdev_idx >= ab->num_radios) {
if (ab->hw_params.single_pdev_only &&
pdev_idx < ab->hw_params.num_rxmda_per_pdev)
if (ab->default_regd[pdev_idx] && !ab->new_regd[pdev_idx] &&
!memcmp((char *)ab->default_regd[pdev_idx]->alpha2,
if (ab->default_regd[pdev_idx] &&
ab->default_regd[pdev_idx]->alpha2) &&
regd = ath11k_reg_build_regd(ab, reg_info, intersect);
ath11k_warn(ab, "failed to build regd from reg_info\n");
spin_lock(&ab->base_lock);
if (ab->default_regd[pdev_idx]) {
ar = ab->pdevs[pdev_idx].ar;
kfree(ab->new_regd[pdev_idx]);
ab->new_regd[pdev_idx] = regd;
queue_work(ab->workqueue, &ar->regd_update_work);
ab->default_regd[pdev_idx] = regd;
ab->dfs_region = reg_info->dfs_region;
spin_unlock(&ab->base_lock);
ath11k_bcn_tx_status_event(ab, skb);
ath11k_peer_sta_kickout_event(ab, skb);
ath11k_pdev_bss_chan_info_event(ab, skb);
ath11k_pdev_ctl_failsafe_check_event(ab, skb);
ath11k_wmi_pdev_csa_switch_count_status_event(ab, skb);
ath11k_tm_wmi_event(ab, id, skb);
ath11k_wmi_pdev_temperature_event(ab, skb);
ath11k_wmi_pdev_dma_ring_buf_release_event(ab, skb);
ath11k_fils_discovery_event(ab, skb);
ath11k_probe_resp_tx_status_event(ab, skb);
ath11k_wmi_obss_color_collision_event(ab, skb);
ath11k_wmi_twt_add_dialog_event(ab, skb);
ath11k_wmi_pdev_dfs_radar_detected_event(ab, skb);
ath11k_vdev_delete_resp_event(ab, skb);
ath11k_wmi_event_wow_wakeup_host(ab, skb);
ath11k_reg_11d_new_cc_event(ab, skb);
ath11k_wmi_event_peer_sta_ps_state_chg(ab, skb);
ath11k_wmi_gtk_offload_status_event(ab, skb);
spin_lock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
spin_lock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
ath11k_htt_pull_ppdu_stats(ab, skb);
ath11k_debugfs_htt_ext_stats_handler(ab, skb);
ath11k_htt_pktlog(ab, skb);
ath11k_htt_backpressure_event_handler(ab, skb);
ret = ath11k_dp_srng_setup(ar->ab,
ath11k_warn(ar->ab,
ret = ath11k_dp_srng_setup(ar->ab, &dp->rxdma_mon_dst_ring,
ath11k_warn(ar->ab,
ret = ath11k_dp_srng_setup(ar->ab, &dp->rxdma_mon_desc_ring,
ath11k_warn(ar->ab,
ret = ath11k_dp_tx_htt_srng_setup(ab, ring_id,
ath11k_warn(ab, "failed to configure rxdma_mon_buf_ring %d\n",
ret = ath11k_dp_tx_htt_srng_setup(ab,
ath11k_warn(ab, "failed to configure rxdma_mon_dst_ring %d\n",
ret = ath11k_dp_tx_htt_srng_setup(ab,
ath11k_warn(ab, "failed to configure rxdma_mon_dst_ring %d\n",
if (unlikely((ath11k_hal_srng_dst_peek(ab, status_ring) != NULL) &&
ath11k_warn(ab, "Unable to process some of the tx_status ring desc because status_fifo is full\n");
ath11k_warn(ab, "tx completion for unknown msdu_id %d\n",
ar = ab->pdevs[mac_id].ar;
ab->soc_stats.reo_error[err_code]++;
uint8_t *hdr_status = ath11k_dp_rx_h_80211_hdr(ar->ab, rx_desc);
ath11k_warn(ar->ab, "invalid msdu leng %u", msdu_len);
ath11k_dbg_dump(ar->ab, ATH11K_DBG_DATA, NULL, "", hdr_status,
ath11k_dbg_dump(ar->ab, ATH11K_DBG_DATA, NULL, "", rx_desc,
ab->soc_stats.err_ring_pkts++;
ab->soc_stats.invalid_rbm++;
ab->soc_stats.invalid_rbm++;
ar->ab->soc_stats.reo_error[rxcb->err_code]++;
ar->ab->soc_stats.rxdma_error[rxcb->err_code]++;
if (ath11k_dp_rx_h_msdu_start_mesh_ctl_present(ar->ab, rxcb->rx_desc))
spin_lock_bh(&ar->ab->base_lock);
spin_unlock_bh(&ar->ab->base_lock);
hdr_status = ath11k_dp_rx_h_80211_hdr(ab, rx_desc);
ath11k_dbg_dump(ab, ATH11K_DBG_DATA, NULL, "", hdr_status,
ath11k_dbg_dump(ab, ATH11K_DBG_DATA, NULL, "", rx_desc,
if (ab->hw_params.full_monitor_mode) {
spin_lock_bh(&ab->base_lock);
peer = ath11k_peer_find_by_id(ab, ppdu_info->peer_id);
ath11k_dbg(ab, ATH11K_DBG_DATA,
spin_unlock_bh(&ab->base_lock);
ab->hw_params.full_monitor_mode)
ret = ath11k_dp_full_mon_process_rx(ab, mac_id, napi, budget);
ab->soc_stats.rxdma_error[rxdma_err_code]++;
ath11k_dbg(ab, ATH11K_DBG_HAL, "Queue stats status:\n");
ath11k_dbg(ab, ATH11K_DBG_HAL, "header: cmd_num %d status %d\n",
ath11k_dbg(ab, ATH11K_DBG_HAL, "ssn %ld cur_idx %ld\n",
ath11k_dbg(ab, ATH11K_DBG_HAL, "pn = [%08x, %08x, %08x, %08x]\n",
ath11k_dbg(ab, ATH11K_DBG_HAL,
ath11k_dbg(ab, ATH11K_DBG_HAL,
ath11k_dbg(ab, ATH11K_DBG_HAL, "count: cur_mpdu %ld cur_msdu %ld\n",
ath11k_dbg(ab, ATH11K_DBG_HAL, "fwd_timeout %ld fwd_bar %ld dup_count %ld\n",
ath11k_dbg(ab, ATH11K_DBG_HAL, "frames_in_order %ld bar_rcvd %ld\n",
ath11k_dbg(ab, ATH11K_DBG_HAL, "num_mpdus %d num_msdus %d total_bytes %d\n",
ath11k_dbg(ab, ATH11K_DBG_HAL, "late_rcvd %ld win_jump_2k %ld hole_cnt %ld\n",
ath11k_dbg(ab, ATH11K_DBG_HAL, "looping count %ld\n",
qwx_hal_reo_flush_timeout_list_status(struct qwx_softc *ab, uint32_t *reo_desc,
qwx_hal_reo_update_rx_reo_queue_status(struct qwx_softc *ab, uint32_t *reo_desc,
ret = ath11k_mac_allocate(ab);
ath11k_err(ab, "failed to create new hw device with mac80211 :%d\n",
ath11k_mac_destroy(ab);
ret = ath11k_thermal_register(ab);
ath11k_err(ab, "could not register thermal device: %d\n",
ret = ath11k_spectral_init(ab);
ath11k_err(ab, "failed to init spectral %d\n", ret);
ath11k_thermal_unregister(ab);
ath11k_mac_unregister(ab);
ath11k_debugfs_pdev_destroy(ab);
mutex_lock(&ab->core_lock);
mutex_unlock(&ab->core_lock);
ath11k_mac_destroy(ab);
ath11k_debugfs_soc_destroy(ab);
ath11k_reg_free(ab);
mutex_lock(&ab->core_lock);
mutex_unlock(&ab->core_lock);
ath11k_mac_destroy(ab);
mutex_unlock(&ab->core_lock);
spin_lock_bh(&ab->ce.ce_lock);
spin_unlock_bh(&ab->ce.ce_lock);
lockdep_assert_held(&ab->ce.ce_lock);
spin_lock_bh(&ab->ce.ce_lock);
spin_unlock_bh(&ab->ce.ce_lock);
mod_timer(&ab->rx_replenish_retry,
spin_lock_bh(&ab->ce.ce_lock);
spin_unlock_bh(&ab->ce.ce_lock);
mod_timer(&ab->rx_replenish_retry,
spin_lock_bh(&ab->ce.ce_lock);
spin_unlock_bh(&ab->ce.ce_lock);
spin_lock_bh(&ab->ce.ce_lock);
spin_unlock_bh(&ab->ce.ce_lock);
spin_lock_bh(&ab->base_lock);
arg.regdomain = ar->ab->dfs_region;
spin_unlock_bh(&ab->base_lock);
ath11k_dbg(ab, ATH11K_DBG_MAC,
ath11k_warn(ab, "failed to create vdev due to insufficient peer entry resource in firmware\n");
ath11k_warn(ab, "failed to vdev %d create peer for AP: %d\n",
ath11k_warn(ar->ab, "failed to set vdev %i kickout parameters: %d\n",
ath11k_mac_11d_scan_stop_all(ar->ab);
ath11k_warn(ar->ab, "failed to create monitor vdev during add interface: %d",
mutex_lock(&ar->ab->vdev_id_11d_lock);
mutex_unlock(&ar->ab->vdev_id_11d_lock);
mutex_lock(&ar->ab->tbl_mtx_lock);
spin_lock_bh(&ar->ab->base_lock);
spin_unlock_bh(&ar->ab->base_lock);
mutex_unlock(&ar->ab->tbl_mtx_lock);
ath11k_peer_rhash_delete(ar->ab, peer);
spin_unlock_bh(&ar->ab->base_lock);
mutex_unlock(&ar->ab->tbl_mtx_lock);
mutex_lock(&ar->ab->tbl_mtx_lock);
spin_lock_bh(&ar->ab->base_lock);
peer = ath11k_peer_find(ar->ab, param->vdev_id, param->peer_addr);
spin_unlock_bh(&ar->ab->base_lock);
mutex_unlock(&ar->ab->tbl_mtx_lock);
ath11k_warn(ar->ab, "failed to find peer %pM on vdev %i after creation\n",
ret = ath11k_peer_rhash_add(ar->ab, peer);
spin_unlock_bh(&ar->ab->base_lock);
mutex_unlock(&ar->ab->tbl_mtx_lock);
spin_unlock_bh(&ar->ab->base_lock);
mutex_unlock(&ar->ab->tbl_mtx_lock);
lockdep_assert_held(&ab->base_lock);
ath11k_dp_rx_link_desc_return(ab, (u32 *)rx_tid->dst_ring_desc,
lockdep_assert_held(&ar->ab->base_lock);
spin_unlock_bh(&ar->ab->base_lock);
spin_lock_bh(&ar->ab->base_lock);
lockdep_assert_held(&ar->ab->base_lock);
spin_unlock_bh(&ar->ab->base_lock);
spin_lock_bh(&ar->ab->base_lock);
spin_lock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
spin_lock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
spin_lock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
rx_tid->ab = ab;
spin_unlock_bh(&ab->base_lock);
spin_lock_bh(&ab->base_lock);
peer = ath11k_peer_find(ab, vdev_id, addr);
ath11k_warn(ab, "failed to find the peer to del rx tid\n");
spin_unlock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
ab->hw_params.hw_ops->tx_mesh_enable(ab, tcl_cmd);
ab->soc_stats.tx_err.desc_na[ti.ring_id]++;
ab->hw_rate_code = hw_rate_code;
ath11k_warn(ar->ab, "failed to recalc he txbf for vdev %i on bss %pM\n",
if (test_bit(ATH11K_FLAG_RAW_MODE, &ar->ab->dev_flags))
bdf_addr = ioremap(ab->hw_params.bdf_addr, ab->hw_params.fw.board_size);
ath11k_warn(ab, "qmi ioremap error for bdf_addr\n");
if (ab->hw_params.fixed_bdf_addr) {
bdf_addr += ab->hw_params.fw.cal_offset;
if (ab->hw_params.fixed_bdf_addr)
if (ab->qmi.target.eeprom_caldata) {
ath11k_bus_str(ab->hif.bus), dev_name(dev));
fw_entry = ath11k_core_firmware_request(ab, filename);
fw_entry = ath11k_core_firmware_request(ab, ATH11K_DEFAULT_CAL_FILE);
ath11k_info(ab,
ath11k_warn(ab,
fw_size = MIN(ab->hw_params.fw.board_size, fw_entry->size);
ret = ath11k_qmi_load_file_target_mem(ab, tmp, fw_size, file_type);
ath11k_warn(ab, "qmi failed to load caldata\n");
ath11k_dbg(ab, ATH11K_DBG_QMI, "caldata type: %u\n", file_type);
if (!ab->qmi.target.eeprom_caldata)
ath11k_hal_srng_prefetch_desc(ab, srng);
struct ath11k_base *ab;
void (*tx_mesh_enable)(struct ath11k_base *ab,
spin_lock(&ab->base_lock);
if (test_bit(ATH12K_FLAG_RECOVERY, &ab->dev_flags) &&
ab->default_regd[pdev_idx]) {
spin_unlock(&ab->base_lock);
spin_unlock(&ab->base_lock);
if (pdev_idx >= ab->num_radios) {
if (ab->hw_params.single_pdev_only &&
pdev_idx < ab->hw_params.num_rxmda_per_pdev)
if (ab->default_regd[pdev_idx] && !ab->new_regd[pdev_idx] &&
!memcmp((char *)ab->default_regd[pdev_idx]->alpha2,
if (ab->default_regd[pdev_idx] &&
ab->default_regd[pdev_idx]->alpha2) &&
regd = ath12k_reg_build_regd(ab, reg_info, intersect);
ath12k_warn(ab, "failed to build regd from reg_info\n");
spin_lock(&ab->base_lock);
if (ab->default_regd[pdev_idx]) {
ar = ab->pdevs[pdev_idx].ar;
kfree(ab->new_regd[pdev_idx]);
ab->new_regd[pdev_idx] = regd;
queue_work(ab->workqueue, &ar->regd_update_work);
ab->default_regd[pdev_idx] = regd;
ab->dfs_region = reg_info->dfs_region;
spin_unlock(&ab->base_lock);
ath12k_bcn_tx_status_event(ab, skb);
ath12k_peer_sta_kickout_event(ab, skb);
ath12k_pdev_bss_chan_info_event(ab, skb);
ath12k_pdev_ctl_failsafe_check_event(ab, skb);
ath12k_wmi_pdev_csa_switch_count_status_event(ab, skb);
ath12k_tm_wmi_event(ab, id, skb);
ath12k_wmi_pdev_temperature_event(ab, skb);
ath12k_wmi_pdev_dma_ring_buf_release_event(ab, skb);
ath12k_fils_discovery_event(ab, skb);
ath12k_probe_resp_tx_status_event(ab, skb);
ath12k_wmi_obss_color_collision_event(ab, skb);
ath12k_wmi_twt_add_dialog_event(ab, skb);
ath12k_wmi_pdev_dfs_radar_detected_event(ab, skb);
ath12k_vdev_delete_resp_event(ab, skb);
ath12k_wmi_event_wow_wakeup_host(ab, skb);
ath12k_reg_11d_new_cc_event(ab, skb);
ath12k_wmi_event_peer_sta_ps_state_chg(ab, skb);
ath12k_wmi_gtk_offload_status_event(ab, skb);
spin_lock_bh(&ab->base_lock);
list_add(&peer->list, &ab->peers);
spin_unlock_bh(&ab->base_lock);
spin_lock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
ath12k_htt_pull_ppdu_stats(ab, skb);
ath12k_debugfs_htt_ext_stats_handler(ab, skb);
ath12k_htt_pktlog(ab, skb);
ath12k_htt_backpressure_event_handler(ab, skb);
ret = ath12k_dp_srng_setup(ar->ab,
ath12k_warn(ar->ab,
ret = ath12k_dp_srng_setup(ar->ab, &dp->rxdma_mon_dst_ring,
ath12k_warn(ar->ab,
ret = ath12k_dp_srng_setup(ar->ab, &dp->rxdma_mon_desc_ring,
ath12k_warn(ar->ab,
if (unlikely((qwz_hal_srng_dst_peek(ab, status_ring) != NULL) &&
ath12k_warn(ab, "Unable to process some of the tx_status ring desc because status_fifo is full\n");
ath12k_warn(ab, "tx completion for unknown msdu_id %d\n",
ar = ab->pdevs[mac_id].ar;
ab->soc_stats.reo_error[err_code]++;
uint8_t *hdr_status = ath12k_dp_rx_h_80211_hdr(ar->ab, rx_desc);
ath12k_warn(ar->ab, "invalid msdu leng %u", msdu_len);
ath12k_dbg_dump(ar->ab, ATH12K_DBG_DATA, NULL, "", hdr_status,
ath12k_dbg_dump(ar->ab, ATH12K_DBG_DATA, NULL, "", rx_desc,
ab->soc_stats.err_ring_pkts++;
ab->soc_stats.invalid_rbm++;
ab->soc_stats.invalid_rbm++;
ar->ab->soc_stats.reo_error[rxcb->err_code]++;
ar->ab->soc_stats.rxdma_error[rxcb->err_code]++;
rxcb->peer_id = ath12k_dp_rx_h_mpdu_start_peer_id(ar->ab, rx_desc);
rxcb->seq_no = ath12k_dp_rx_h_mpdu_start_seq_no(ar->ab, rx_desc);
spin_lock_bh(&ar->ab->base_lock);
peer = ath12k_dp_rx_h_find_peer(ar->ab, msdu);
spin_unlock_bh(&ar->ab->base_lock);
hdr_status = ath12k_dp_rx_h_80211_hdr(ab, rx_desc);
ath12k_dbg_dump(ab, ATH12K_DBG_DATA, NULL, "", hdr_status,
ath12k_dbg_dump(ab, ATH12K_DBG_DATA, NULL, "", rx_desc,
if (ab->hw_params.full_monitor_mode) {
spin_lock_bh(&ab->base_lock);
peer = ath12k_peer_find_by_id(ab, ppdu_info->peer_id);
ath12k_dbg(ab, ATH12K_DBG_DATA,
spin_unlock_bh(&ab->base_lock);
ab->hw_params.full_monitor_mode)
ret = ath12k_dp_full_mon_process_rx(ab, mac_id, napi, budget);
ab->soc_stats.rxdma_error[rxdma_err_code]++;
ath12k_dbg(ab, ATH12K_DBG_HAL, "Queue stats status:\n");
ath12k_dbg(ab, ATH12K_DBG_HAL, "header: cmd_num %d status %d\n",
ath12k_dbg(ab, ATH12K_DBG_HAL, "ssn %ld cur_idx %ld\n",
ath12k_dbg(ab, ATH12K_DBG_HAL, "pn = [%08x, %08x, %08x, %08x]\n",
ath12k_dbg(ab, ATH12K_DBG_HAL,
ath12k_dbg(ab, ATH12K_DBG_HAL,
ath12k_dbg(ab, ATH12K_DBG_HAL, "count: cur_mpdu %ld cur_msdu %ld\n",
ath12k_dbg(ab, ATH12K_DBG_HAL, "fwd_timeout %ld fwd_bar %ld dup_count %ld\n",
ath12k_dbg(ab, ATH12K_DBG_HAL, "frames_in_order %ld bar_rcvd %ld\n",
ath12k_dbg(ab, ATH12K_DBG_HAL, "num_mpdus %d num_msdus %d total_bytes %d\n",
ath12k_dbg(ab, ATH12K_DBG_HAL, "late_rcvd %ld win_jump_2k %ld hole_cnt %ld\n",
ath12k_dbg(ab, ATH12K_DBG_HAL, "looping count %ld\n",
qwz_hal_reo_flush_timeout_list_status(struct qwz_softc *ab, uint32_t *reo_desc,
qwz_hal_reo_update_rx_reo_queue_status(struct qwz_softc *ab, uint32_t *reo_desc,
ret = ath12k_mac_allocate(ab);
ath12k_err(ab, "failed to create new hw device with mac80211 :%d\n",
ath12k_mac_destroy(ab);
ret = ath12k_thermal_register(ab);
ath12k_err(ab, "could not register thermal device: %d\n",
ret = ath12k_spectral_init(ab);
ath12k_err(ab, "failed to init spectral %d\n", ret);
ath12k_thermal_unregister(ab);
ath12k_mac_unregister(ab);
ath12k_debugfs_pdev_destroy(ab);
mutex_lock(&ab->core_lock);
mutex_unlock(&ab->core_lock);
ath12k_mac_destroy(ab);
ath12k_debugfs_soc_destroy(ab);
ath12k_reg_free(ab);
mutex_lock(&ab->core_lock);
mutex_unlock(&ab->core_lock);
ath12k_mac_destroy(ab);
mutex_unlock(&ab->core_lock);
spin_lock_bh(&ab->ce.ce_lock);
spin_unlock_bh(&ab->ce.ce_lock);
lockdep_assert_held(&ab->ce.ce_lock);
spin_lock_bh(&ab->ce.ce_lock);
spin_unlock_bh(&ab->ce.ce_lock);
mod_timer(&ab->rx_replenish_retry,
spin_lock_bh(&ab->ce.ce_lock);
spin_unlock_bh(&ab->ce.ce_lock);
mod_timer(&ab->rx_replenish_retry,
spin_lock_bh(&ab->ce.ce_lock);
spin_unlock_bh(&ab->ce.ce_lock);
spin_lock_bh(&ab->ce.ce_lock);
spin_unlock_bh(&ab->ce.ce_lock);
spin_lock_bh(&ab->base_lock);
arg.regdomain = ar->ab->dfs_region;
spin_unlock_bh(&ab->base_lock);
ath12k_dbg(ab, ATH12K_DBG_MAC,
ath12k_warn(ab, "failed to create vdev due to insufficient peer entry resource in firmware\n");
ath12k_warn(ab, "failed to vdev %d create peer for AP: %d\n",
ath12k_warn(ar->ab, "failed to set vdev %i kickout parameters: %d\n",
ath12k_mac_11d_scan_stop_all(ar->ab);
ath12k_warn(ar->ab, "failed to create monitor vdev during add interface: %d",
mutex_lock(&ar->ab->vdev_id_11d_lock);
mutex_unlock(&ar->ab->vdev_id_11d_lock);
mutex_lock(&ar->ab->tbl_mtx_lock);
spin_lock_bh(&ar->ab->base_lock);
spin_unlock_bh(&ar->ab->base_lock);
mutex_unlock(&ar->ab->tbl_mtx_lock);
ath12k_peer_rhash_delete(ar->ab, peer);
spin_unlock_bh(&ar->ab->base_lock);
mutex_unlock(&ar->ab->tbl_mtx_lock);
mutex_lock(&ar->ab->tbl_mtx_lock);
spin_lock_bh(&ar->ab->base_lock);
peer = ath12k_peer_find(ar->ab, param->vdev_id, param->peer_addr);
spin_unlock_bh(&ar->ab->base_lock);
mutex_unlock(&ar->ab->tbl_mtx_lock);
ath12k_warn(ar->ab, "failed to find peer %pM on vdev %i after creation\n",
ret = ath12k_peer_rhash_add(ar->ab, peer);
spin_unlock_bh(&ar->ab->base_lock);
mutex_unlock(&ar->ab->tbl_mtx_lock);
spin_unlock_bh(&ar->ab->base_lock);
mutex_unlock(&ar->ab->tbl_mtx_lock);
lockdep_assert_held(&ab->base_lock);
ath12k_dp_rx_link_desc_return(ab, (u32 *)rx_tid->dst_ring_desc,
lockdep_assert_held(&ar->ab->base_lock);
spin_unlock_bh(&ar->ab->base_lock);
spin_lock_bh(&ar->ab->base_lock);
lockdep_assert_held(&ar->ab->base_lock);
spin_unlock_bh(&ar->ab->base_lock);
spin_lock_bh(&ar->ab->base_lock);
spin_lock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
spin_lock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
spin_lock_bh(&ab->base_lock);
rx_tid->ab = ab;
spin_unlock_bh(&ab->base_lock);
spin_lock_bh(&ab->base_lock);
peer = ath12k_peer_find(ab, vdev_id, addr);
ath12k_warn(ab, "failed to find the peer to del rx tid\n");
spin_unlock_bh(&ab->base_lock);
spin_unlock_bh(&ab->base_lock);
ab->hw_params.hw_ops->tx_mesh_enable(ab, tcl_cmd);
if (!test_bit(ATH12K_FLAG_RAW_MODE, &ab->dev_flags)) {
atomic_inc(&ab->soc_stats.tx_err.misc_fail);
ab->soc_stats.tx_err.desc_na[ti.ring_id]++;
ab->hw_rate_code = hw_rate_code;
ath12k_warn(ar->ab, "failed to setup peer SMPS for vdev %d: %d\n",
ath12k_warn(ar->ab, "failed to recalc he txbf for vdev %i on bss %pM\n",
if (test_bit(ATH12K_FLAG_RAW_MODE, &ar->ab->dev_flags))
ath12k_dp_tx_pending_cleanup, ab);
struct ath12k_base *ab;
const struct arc_board *ab;
ab = &arc_devices[i];
if (PCI_VENDOR(pa->pa_id) == ab->ab_vendor &&
PCI_PRODUCT(pa->pa_id) == ab->ab_product)
return (ab);
#define amdgpu_ring_emit_gds_switch(r, v, db, ds, wb, ws, ab, as) (r)->funcs->emit_gds_switch((r), (v), (db), (ds), (wb), (ws), (ab), (as))
ath11k_err(ab, "failed to set irq affinity %d\n", ret);
ath11k_err(ab, "failed to config msi_data: %d\n", ret);
ath11k_pcic_sync_ce_irqs(ab);
ath11k_pcic_kill_tasklets(ab);
ath12k_pcic_sync_ce_irqs(ab);
ath12k_pcic_kill_tasklets(ab);
ath12k_err(ab, "failed to set irq affinity %d\n", ret);
ath12k_err(ab, "failed to config msi_data: %d\n", ret);
const union gre_addr *ab,unsigned int if0idxb)
rv = gre_ip_cmp(a->t_af, aa, ab);
char commpid[sizeof(ab.ac_comm) + 13];
if (fread(&ab, sizeof(struct acct), 1, fp) != 1)
if (ab.ac_comm[0] == '\0') {
ab.ac_comm[0] = '?';
ab.ac_comm[1] = '\0';
for (p = &ab.ac_comm[0];
p < &ab.ac_comm[sizeof ab.ac_comm] && *p; ++p)
ab.ac_comm, ab.ac_pid);
if (!*argv || requested(argv, &ab)) {
t = expand(ab.ac_utime) + expand(ab.ac_stime);
commpid, flagbits(ab.ac_flag), UT_NAMESIZE,
UT_NAMESIZE, user_from_uid(ab.ac_uid, 0),
UT_LINESIZE, UT_LINESIZE, getdev(ab.ac_tty),
t / (double)AHZ, ctime(&ab.ac_btime));
delta = expand(ab.ac_etime) / (double)AHZ;
struct acct ab;
const struct pf_addr *aa, *ab;
ab = &s2->key[PF_SK_STACK].addr[ii];
ab = &s2->key[side].addr[io];
ret = compare_addr(af, aa, ab);
const struct pf_addr *aa, *ab;
ab = &s2->key[PF_SK_STACK].addr[ii];
ab = &s2->key[side].addr[io];
ret = compare_addr(af, aa, ab);
struct devbase *ib, *ab;
ab = ht_lookup(devbasetab, cp);
if (ab == NULL) {
for (nv = ab->d_attrs; nv != NULL; nv = nv->nv_next) {
while (i && i->i_atdev != ab)
struct devbase *ab; /* not NULL => at another dev */
ab = NULL;
ab = ht_lookup(devbasetab, cp);
if (ab == NULL) {
for (nv = ab->d_attrs; nv != NULL; nv = nv->nv_next) {
i->i_atdev = ab;
const void *aa, *ab;
ab = &(((const struct sockaddr_in*)b)->sin_addr);
return memcmp(aa, ab, l);
ab = &(((const struct sockaddr_in6*)b)->sin6_addr);
return memcmp(aa, ab, l);
char *startbuf, *ab, *envvar;
ab = emalloc(max_abbr_len + 1);
doabbr(ab, max_abbr_len + 1, zp,
type = addtype(offset, ab, rp->r_stdoff != 0,
free(ab);