BITS_PER_TYPE
eix = &imsic->swfile->eix[iid / BITS_PER_TYPE(u64)];
set_bit(iid & (BITS_PER_TYPE(u64) - 1), eix->eip);
#define IMSIC_MAX_EIX (IMSIC_MAX_ID / BITS_PER_TYPE(u64))
unsigned long eip[BITS_PER_TYPE(u64) / BITS_PER_LONG];
unsigned long eie[BITS_PER_TYPE(u64) / BITS_PER_LONG];
unsigned long eipend[BITS_PER_TYPE(u64) / BITS_PER_LONG];
imin = ei * BITS_PER_TYPE(u64);
imax = ((imin + BITS_PER_TYPE(u64)) < max_msi) ?
imin + BITS_PER_TYPE(u64) : max_msi;
BITS_PER_TYPE(u64)];
clear_bit(topei & (BITS_PER_TYPE(u64) - 1),
} else if (ret * BITS_PER_TYPE(u8) < num_lanes + 1) {
ret * BITS_PER_TYPE(u8), num_lanes + 1);
BUILD_BUG_ON(BITS_PER_TYPE(unsigned long) <= ACPI_DEVICE_CSI2_DATA_LANES);
mask |= (unsigned long)val[i] << BITS_PER_TYPE(u8) * i;
#define NBITS_MIN BITS_PER_TYPE(unsigned long)
#define CC_TRNG_EHR_IN_BITS (CC_TRNG_EHR_IN_WORDS * BITS_PER_TYPE(u32))
DECLARE_BITMAP(csr, BITS_PER_TYPE(u64));
for_each_set_bit(i, csr, BITS_PER_TYPE(csr)) {
for_each_set_bit(i, csr, BITS_PER_TYPE(csr)) {
SET_CAP_CSR(handle, fcu_dram_hi_csr, bus_addr >> BITS_PER_TYPE(u32));
low_pos = f->lsb / BITS_PER_TYPE(*data);
hi_pos = f->msb / BITS_PER_TYPE(*data);
low_bit = f->lsb % BITS_PER_TYPE(*data);
hi_bit = f->msb % BITS_PER_TYPE(*data);
low_bit = f->lsb % BITS_PER_TYPE(*data);
value = ((u64)data[hi_pos] << BITS_PER_TYPE(*data)) |
hi_bit = f->msb % BITS_PER_TYPE(*data);
low_pos = f->lsb / BITS_PER_TYPE(*data);
hi_pos = f->msb / BITS_PER_TYPE(*data);
low_bit = f->lsb % BITS_PER_TYPE(*data);
array_max = DIV_ROUND_UP(ecc->num_channels, BITS_PER_TYPE(u32));
for (b = 0; b < BITS_PER_TYPE(u64); b++) {
if ((record[i].mem_channel < BITS_PER_TYPE(control->bad_channel_bitmap)) &&
if ((record[i].mem_channel < BITS_PER_TYPE(control->bad_channel_bitmap)) &&
if ((record[i].mem_channel < BITS_PER_TYPE(control->bad_channel_bitmap)) &&
if ((record[i].mem_channel < BITS_PER_TYPE(control->bad_channel_bitmap)) &&
for_each_set_bit(id, &move_mask, BITS_PER_TYPE(move_mask))
stream_id > BITS_PER_TYPE(tunnel_state->stream_mask)))
for_each_set_bit(opt, &method, BITS_PER_TYPE(method)) {
int i, len = min_t(int, info->vics_len, BITS_PER_TYPE(y420cmdb_map));
if (WARN_ON_ONCE(nbits > BITS_PER_TYPE(value)))
nbits = BITS_PER_TYPE(value);
BUILD_BUG_ON(BITS_PER_TYPE(display_runtime->pipe_mask) < I915_MAX_PIPES);
BUILD_BUG_ON(BITS_PER_TYPE(display_runtime->cpu_transcoder_mask) < I915_MAX_TRANSCODERS);
BUILD_BUG_ON(BITS_PER_TYPE(display_runtime->port_mask) < I915_MAX_PORTS);
for_each_set_bit(i, &abox_regs, BITS_PER_TYPE(abox_regs))
for_each_set_bit(i, &abox_mask, BITS_PER_TYPE(abox_mask))
for_each_set_bit(i, &abox_mask, BITS_PER_TYPE(abox_mask)) {
#define INTEL_DP_LINK_RATE_IDX_BITS (BITS_PER_TYPE(u8) - INTEL_DP_LANE_COUNT_EXP_BITS)
(int)BITS_PER_TYPE(sink_slice_count_mask), &sink_slice_count_mask);
BITS_PER_TYPE(atomic_t));
BUILD_BUG_ON(BITS_PER_TYPE(available_pins) < HPD_NUM_PINS);
BUILD_BUG_ON(BITS_PER_TYPE(*pin_mask) < HPD_NUM_PINS);
BITS_PER_TYPE(frame->ring.size) - ilog2(frame->ring.size);
BUILD_BUG_ON(BITS_PER_TYPE(engine->mask) < I915_NUM_ENGINES);
GENMASK(BITS_PER_TYPE(mask) - 1, I915_NUM_ENGINES));
#define VIRTUAL_ENGINES BIT(BITS_PER_TYPE(intel_engine_mask_t) - 1)
ring->wrap = BITS_PER_TYPE(ring->size) - ilog2(size);
#define XEHP_BITMAP_BITS(mask) ((int)BITS_PER_TYPE(typeof(mask.xehp)))
ring->wrap = BITS_PER_TYPE(ring->size) - ilog2(sz);
BITS_PER_TYPE(RUNTIME_INFO(gt->i915)->page_sizes)) {
BITS_PER_TYPE(info->platform_mask[0]) - INTEL_SUBPLATFORM_BITS;
BITS_PER_TYPE(info->platform_mask[0]) - INTEL_SUBPLATFORM_BITS;
const unsigned int msb = BITS_PER_TYPE(info->platform_mask[0]) - 1;
BUILD_BUG_ON(ARRAY_SIZE(names) >= BITS_PER_TYPE(mitigations));
BITS_PER_TYPE(typeof_member(struct i915_pmu,
BITS_PER_TYPE(typeof_member(struct i915_pmu,
BITS_PER_TYPE(typeof_member(struct wait_queue_entry, flags))
BUILD_BUG_ON(KSYNCMAP > BITS_PER_TYPE((*root)->bitmap));
BUILD_BUG_ON(BITS_PER_TYPE(intel_engine_mask_t) < I915_NUM_ENGINES);
BITS_PER_TYPE(typeof_member(struct intel_runtime_pm, wakeref_count))
if (exp >= BITS_PER_TYPE(u32))
if (abs(shift) >= BITS_PER_TYPE(u32))
BITS_PER_TYPE(ec->board_info->sensors)) {
DECLARE_BITMAP(values, BITS_PER_TYPE(val));
BITS_PER_TYPE(struct tsc2046_adc_atom)
DECLARE_BITMAP(values, BITS_PER_TYPE(value));
if (op >= BITS_PER_TYPE(u32))
BITS_PER_TYPE(*bitmap->bitmap);
return (index * BITS_PER_TYPE(*bitmap->bitmap)) << pgshift;
count = (MAX_BIG_FIELD_BITS - ((to_offset + MAX_BIG_FIELD_BITS) % BITS_PER_TYPE(u32)));
offset = from_offset % BITS_PER_TYPE(u32);
from_offset += BITS_PER_TYPE(u32);
to_offset += BITS_PER_TYPE(u32);
size -= BITS_PER_TYPE(u32);
count = (to_offset + size) % BITS_PER_TYPE(u32);
offset = (from_offset + size) % BITS_PER_TYPE(u32);
size -= BITS_PER_TYPE(u32);
if (core->supported_instance_count >= BITS_PER_TYPE(core->instance_mask))
core->supported_instance_count = BITS_PER_TYPE(core->instance_mask);
if (sd->params.exposure.gain >= BITS_PER_TYPE(currentexp))
if (index >= BITS_PER_TYPE(mapping->menu_mask))
DECLARE_BITMAP(values, BITS_PER_TYPE(byte)) = { byte, };
DECLARE_BITMAP(values, BITS_PER_TYPE(byte)) = { byte, };
DECLARE_BITMAP(values, BITS_PER_TYPE(res)) = { 0, };
DECLARE_BITMAP(values, BITS_PER_TYPE(state));
bitmap_from_arr32(values, &value, BITS_PER_TYPE(value));
for_each_set_bit(idx, &supported, BITS_PER_TYPE(u32))
tx_ring->obj_num_shift_to_u8 = BITS_PER_TYPE(tx_ring->obj_num) -
rx_ring->obj_num_shift_to_u8 = BITS_PER_TYPE(rx_ring->obj_num_shift_to_u8) -
const u32 mask = GENMASK(BITS_PER_TYPE(mask) - 1,
BITS_PER_TYPE(echo_msg->packet_idx));
bmap = ppmax / BITS_PER_TYPE(unsigned long);
HCLGE_COMM_QUERY_CAP_LENGTH * BITS_PER_TYPE(u32));
DECLARE_BITMAP(caps, HCLGE_COMM_QUERY_CAP_LENGTH * BITS_PER_TYPE(u32));
for (u32 i = 0; i < BITS_PER_TYPE(typeof(low)); i++) {
for (u32 i = 0; i < BITS_PER_TYPE(typeof(high)); i++) {
for_each_set_bit(bit, &ul_mask, BITS_PER_TYPE(u16)) {
return bitrev16(v) >> (BITS_PER_TYPE(v) - len);
return bitrev32(v) >> (BITS_PER_TYPE(v) - len);
offset = ICE_GPR_HV_IDX + (start / BITS_PER_TYPE(u16));
val >>= start % BITS_PER_TYPE(u16);
((u32)rt->gpr[kb->len_reg1] << BITS_PER_TYPE(u16));
offset = ICE_GPR_HV_IDX + (start / BITS_PER_TYPE(u16));
val >>= start % BITS_PER_TYPE(u16);
DECLARE_BITMAP(smap, BITS_PER_TYPE(sensor_map));
for_each_set_bit(bit_pos, smap, BITS_PER_TYPE(sensor_map)) {
avail_next = i * BITS_PER_TYPE(long);
range = port->portno / BITS_PER_TYPE(u32);
port_mask = ~BIT(port->portno % BITS_PER_TYPE(u32));
range = port->portno / BITS_PER_TYPE(u32);
port_mask = ~BIT(port->portno % BITS_PER_TYPE(u32));
else if (field && field->width == BITS_PER_TYPE(u32))
#define RSWITCH_NUM_IRQ_REGS (RSWITCH_MAX_NUM_QUEUES / BITS_PER_TYPE(u32))
DECLARE_BITMAP(values, BITS_PER_TYPE(desired_child));
if (peer_id >= BITS_PER_TYPE(peer->peer_ids))
if (chiprev >= BITS_PER_TYPE(u32)) {
BUILD_BUG_ON(BITS_PER_TYPE(fwrt->dsm_funcs_valid) < DSM_FUNC_NUM_FUNCS);
static_assert(BITS_PER_TYPE(unsigned long) >= NUM_OF_RTW89_REGD_FUNC);
BITS_PER_TYPE(config1))
BITS_PER_TYPE(config2))
DECLARE_BITMAP(ovs, BITS_PER_TYPE(u64));
for_each_set_clump8(bit, clump, ®, BITS_PER_TYPE(typeof(reg))) {
for_each_set_clump8(pos, clump, ®, BITS_PER_TYPE(typeof(reg))) {
offset / BITS_PER_TYPE(u32) * sizeof(u32);
*shift = offset % BITS_PER_TYPE(u32);
offset / BITS_PER_TYPE(u32) * sizeof(u32);
*shift = offset % BITS_PER_TYPE(u32);
offset / BITS_PER_TYPE(u32) * sizeof(u32);
*shift = offset % BITS_PER_TYPE(u32);
offset / BITS_PER_TYPE(u32) * sizeof(u32);
*shift = offset % BITS_PER_TYPE(u32);
for (unsigned int j = 0; j < BITS_PER_TYPE(u32); j++) {
u32 apic_id = i * BITS_PER_TYPE(u32) + j;
#define NUM_DOMAINS_EACH_REG BITS_PER_TYPE(u32)
val->intval >= BITS_PER_TYPE(typeof(psy->desc->charge_behaviours)) ||
val->intval >= BITS_PER_TYPE(typeof(psy->desc->charge_types)) ||
static const int bit_weights_uAh[BITS_PER_TYPE(u32)] = {
static const int bit_weights_uA[BITS_PER_TYPE(u8)] = {
for_each_set_bit(i, (ulong *) &phys_to_be_added, BITS_PER_TYPE(u64)) {
for_each_set_bit(i, (ulong *) &phys_to_be_removed, BITS_PER_TYPE(u64)) {
for_each_set_bit(i, &irq_status, BITS_PER_TYPE(u32)) {
DIV_ROUND_UP(HIVE_ISP_DDR_WORD_BITS, BITS_PER_TYPE(short));
DIV_ROUND_UP(HIVE_ISP_DDR_WORD_BITS, BITS_PER_TYPE(short));
if (c < BITS_PER_TYPE(CTRL_ALWAYS)) {
u32 valid_pin[LJCA_MAX_GPIO_NUM / BITS_PER_TYPE(u32)];
for (u8 bit_i = 0; bit_i < BITS_PER_TYPE(u64); ++bit_i) {
i * BITS_PER_TYPE(u64) + bit_i;
#define MAX_LOAD_SIZE (BIT_ULL(BITS_PER_TYPE \
#define DIRTY_BITMAP_BYTES(n) (ALIGN(n, BITS_PER_TYPE(u64)) / BITS_PER_BYTE)
unsigned int __shift = BITS_PER_TYPE(__mask) <= 32 ? \
unsigned int __shift = BITS_PER_TYPE(__mask) <= 32 ? \
#define BITS_TO_LONGS(nr) __KERNEL_DIV_ROUND_UP(nr, BITS_PER_TYPE(long))
#define BITS_TO_U64(nr) __KERNEL_DIV_ROUND_UP(nr, BITS_PER_TYPE(u64))
#define BITS_TO_U32(nr) __KERNEL_DIV_ROUND_UP(nr, BITS_PER_TYPE(u32))
#define BITS_TO_BYTES(nr) __KERNEL_DIV_ROUND_UP(nr, BITS_PER_TYPE(char))
type_max(t) >> (BITS_PER_TYPE(t) - 1 - (h)))))
BUILD_BUG_ON_ZERO(const_true((nr) >= BITS_PER_TYPE(type)))
bitmap_set(supp, 0, BITS_PER_TYPE(__be16));
bitmap_write(dst, be16_to_cpu(flags), 0, BITS_PER_TYPE(__be16));
ret = cpu_to_be16(bitmap_read(flags, 0, BITS_PER_TYPE(__be16)));
for_each_set_bit(hwirq, &ipis, BITS_PER_TYPE(int))
if (BITS_PER_TYPE(int) < nr_ipi || !mux_send)
(val >> min_t(typeof(shift), shift, BITS_PER_TYPE(typeof(val)) - 1))
return bitmap_empty((unsigned long *)chunk->data, BITS_PER_TYPE(chunk->data));
u32 npkts = BIT_GAP(BITS_PER_TYPE(u32), end->pkt_ctr, start->pkt_ctr);
u32 nbytes = BIT_GAP(BITS_PER_TYPE(u32), end->byte_ctr,
u32 ncomps = BIT_GAP(BITS_PER_TYPE(u32), end->comp_ctr,
nevents = BIT_GAP(BITS_PER_TYPE(u16),
? next_test_random(BITS_PER_TYPE(T)) \
#define ID_SHIFT (BITS_PER_TYPE(unsigned short))
#define SWAP_USAGE_OFFLIST_BIT (1UL << (BITS_PER_TYPE(atomic_t) - 2))
BUILD_BUG_ON(BLOOM_FILTER_SHIFT * 2 > BITS_PER_TYPE(u32));
key = pgdat->node_id % BITS_PER_TYPE(mm->lru_gen.bitmap);
if (r < BITS_PER_TYPE(sk->sk_gso_max_size))
BUILD_BUG_ON(SEG6_LOCAL_MAX + 1 > BITS_PER_TYPE(unsigned long));
BUILD_BUG_ON(SEG6_LOCAL_FLV_OP_MAX + 1 > BITS_PER_TYPE(__u32));
if (priv->data.data[0] >= BITS_PER_TYPE(u32)) {
carry = src[i - 1] >> (BITS_PER_TYPE(u32) - shift);
carry = src[i] << (BITS_PER_TYPE(u32) - shift);
for (group = 0; group < BITS_PER_TYPE(u32); group++) {
netlink_undo_bind(BITS_PER_TYPE(u32), groups, sk);
BITS_PER_TYPE(int) - 1,
static_assert(BITS_PER_TYPE(access_mask_t) >= LANDLOCK_NUM_ACCESS_FS);
static_assert(BITS_PER_TYPE(access_mask_t) >= LANDLOCK_NUM_ACCESS_NET);
static_assert(BITS_PER_TYPE(access_mask_t) >= LANDLOCK_NUM_SCOPE);
static_assert(BITS_PER_TYPE(deny_masks_t) >=
BITS_PER_TYPE(access_mask_t)) {
for_each_set_bit(access_bit, &access_mask, BITS_PER_TYPE(access)) {
static_assert(BITS_PER_TYPE(typeof_member(struct landlock_cred_security,
for_each_set_bit(slot, &mask, BITS_PER_TYPE(mask)) {
for_each_set_bit(i, &ssp_mask, BITS_PER_TYPE(ssp_mask)) {
BITS_PER_TYPE(control->cn_list)) {
BITS_PER_TYPE(control_list)) {
for_each_set_bit(pin, (unsigned long *)&pin_list, BITS_PER_TYPE(pin_list)) {
BITS_PER_TYPE(control->cn_list)) {
BITS_PER_TYPE(control->cn_list)) {
BITS_PER_TYPE(control->cn_list)) {
#define BITS_TO_LONGS(nr) DIV_ROUND_UP(nr, BITS_PER_TYPE(long))
#define BITS_TO_U64(nr) DIV_ROUND_UP(nr, BITS_PER_TYPE(u64))
#define BITS_TO_U32(nr) DIV_ROUND_UP(nr, BITS_PER_TYPE(u32))
#define BITS_TO_BYTES(nr) DIV_ROUND_UP(nr, BITS_PER_TYPE(char))
type_max(t) >> (BITS_PER_TYPE(t) - 1 - (h)))))
BUILD_BUG_ON_ZERO(const_true((nr) >= BITS_PER_TYPE(type)))
for_each_set_bit(vm_type, &vm_types, BITS_PER_TYPE(vm_types))
unsigned long maxnode = BITS_PER_TYPE(nodemask);
static_assert(BITS_PER_TYPE(int) >= TRANSPORT_NUM);