__test_and_set_bit
if (!__test_and_set_bit(asid, asid_map))
if (!__test_and_set_bit(vmid2idx(vmid), vmid_map)) {
if (!__test_and_set_bit(ctxid2asid(asid), asid_map))
if (!__test_and_set_bit(asid2idx(info, asid), info->map))
if (!__test_and_set_bit(mmid & mmid_mask, mmid_map)) {
while (__test_and_set_bit(id, map)) {
if (!__test_and_set_bit(irq_nr, ppc_lost_interrupts)) {
if (!__test_and_set_bit(cntx2asid(cntx), context_asid_map))
if (__test_and_set_bit(idx, cpuc->used_mask)) {
if (__test_and_set_bit(entry, pmb_map)) {
if (__test_and_set_bit(die, die_mask))
if (__test_and_set_bit(d->feature, disable))
if (__test_and_set_bit(LSTATE_SSB, &st->local_state))
if (__test_and_set_bit(APIC_VECTOR_TO_BIT_NUMBER(vec),
if (__test_and_set_bit(spte_index(spte), sp->unsync_child_bitmap))
skip_pmi = __test_and_set_bit(GLOBAL_STATUS_BUFFER_OVF_BIT,
if (__test_and_set_bit(idx, ev->used_mask)) {
if (__test_and_set_bit(sinfo_has_content_type, &sinfo->aa_set))
if (__test_and_set_bit(sinfo_has_signing_time, &sinfo->aa_set))
if (__test_and_set_bit(sinfo_has_message_digest, &sinfo->aa_set))
if (__test_and_set_bit(sinfo_has_smime_caps, &sinfo->aa_set))
if (__test_and_set_bit(sinfo_has_ms_opus_info, &sinfo->aa_set))
if (__test_and_set_bit(sinfo_has_ms_statement_type, &sinfo->aa_set))
if (!__test_and_set_bit(__EE_SEND_WRITE_ACK, &peer_req->flags))
if (__test_and_set_bit(*restore_id, qpd->doorbell_bitmap))
if (__test_and_set_bit(qid, pqm->queue_slot_bitmap)) {
if (!__test_and_set_bit(CONTEXT_INIT_BIT, &ce->flags)) {
if (!__test_and_set_bit(CONTEXT_INIT_BIT, &ce->flags))
if (__test_and_set_bit(I915_SW_FENCE_CHECKED_BIT, &fence->flags))
if (__test_and_set_bit(i, effect_handled))
if (__test_and_set_bit(FF_CORE_IS_USED, core_effect->flags)) {
if (!__test_and_set_bit(i, &fdtv->channel_active))
if (!__test_and_set_bit(idx, pf->vlan_ht_filter))
WARN_ON(__test_and_set_bit(ccm->tag_alloc_next, ccm->tag_allocator));
if (__test_and_set_bit(type_idx, found))
if (__test_and_set_bit(req->u.init_rxq.index, vf->rxq_mask))
if (__test_and_set_bit(req->u.init_txq.index, vf->txq_mask))
if (__test_and_set_bit(mem->id, regions)) {
if (__test_and_set_bit(chanctx_idx, map))
if (__test_and_set_bit(hdr->ch, dmux->remote_channels)) {
return __test_and_set_bit(cur, is_kmalloc) ? cur_f : NULL;
if (!__test_and_set_bit(UIO_IRQ_DISABLED, &priv->flags))
if (!__test_and_set_bit(UIO_IRQ_DISABLED, &priv->flags))
if (!__test_and_set_bit(UIO_IRQ_DISABLED, &priv->flags))
if (!__test_and_set_bit(UIO_IRQ_DISABLED, &priv->flags))
if (__test_and_set_bit(_pr->acpi_id, acpi_ids_done)) {
if (!__test_and_set_bit(segno, sit_i->dirty_sentries_bitmap)) {
__test_and_set_bit((nr), (unsigned long *)(addr))
__test_and_set_bit((nr) ^ 16, (unsigned long *)(addr))
if (__test_and_set_bit(NETFS_RREQ_UPLOAD_TO_SERVER, &wreq->flags))
return __test_and_set_bit(nr ^ BITOP_LE_SWIZZLE, addr);
__check_bitop_pr(__test_and_set_bit);
return __test_and_set_bit(entity->internal_idx, ent_enum->bmap);
return __test_and_set_bit(offset, node_marks(node, mark));
KUNIT_EXPECT_KASAN_FAIL(test, __test_and_set_bit(nr, addr));
if (__test_and_set_bit(0, (unsigned long *)
if (!__test_and_set_bit(NAPI_STATE_SCHED, &sd->backlog.state))
if (!__test_and_set_bit(NAPI_STATE_SCHED,
if (__test_and_set_bit(seq_bit, block->seq_nrs[port_type]))
if (!__test_and_set_bit(id ? : msk->mpc_endpoint_id, msk->pm.id_avail_bitmap) &&
WARN_ON_ONCE(__test_and_set_bit(RXRPC_CALL_RELEASED, &call->flags));
if (__test_and_set_bit(ix, &tq->segment_lost)) {
if (__test_and_set_bit(seq_num % GSS_SEQ_WIN, sd->sd_win))
if (__test_and_set_bit(token, &token_mask))
if (!__test_and_set_bit(this_cpu.cpu, sched->map.comp_cpus_mask)) {
if (!__test_and_set_bit(event->id, events_defined))
if (!__test_and_set_bit(event->id, events_defined))
__test_and_set_bit((nr) ^ BITOP_LE_SWIZZLE, addr)
# define __test_and_set_bit_le __test_and_set_bit
#define test_and_set_bit(nr, addr) __test_and_set_bit(nr, addr)