field_mask
u64 vgic_sanitise_field(u64 reg, u64 field_mask, int field_shift,
u64 field = (reg & field_mask) >> field_shift;
return (reg & ~field_mask) | field;
u64 vgic_sanitise_field(u64 reg, u64 field_mask, int field_shift,
gpa_t addr, u64 value, u64 field_mask)
mask = field_mask << (offset * 8);
data = (value & field_mask) << (offset * 8);
field_get(table[i].field_mask,
.field_mask = _mask_, \
u32 field_mask;
u32 cx231xx_set_field(u32 field_mask, u32 data)
for (temp = field_mask; (temp & 1) == 0; temp >>= 1)
u32 cx231xx_set_field(u32 field_mask, u32 data);
field_mask(MCP251XFD_OBJ_FLAGS_SEQ_MCP2517FD_MASK);
u16 field_mask[ICE_MAX_CHAIN_WORDS];
lkup_exts->field_mask[fv_word_idx] = lkup_mask;
u16 *cr = recp[i].lkup_exts.field_mask;
u16 *de = lkup_exts->field_mask;
lkup_exts->field_mask[word] =
memcpy(rm->word_masks, lkup_exts->field_mask, sizeof(rm->word_masks));
u64 field_mask;
if (req->fields[f] != (reg & req->field_mask))
lso->field_mask = GENMASK(18, 0);
u32 field_mask;
#define OFFLOAD(fw_field, field_bsize, field_mask, field, off, match_field) \
{MLX5_ACTION_IN_FIELD_OUT_ ## fw_field, field_bsize, field_mask, \
return be32_to_cpu(*(__be32 *)mask) & f->field_mask;
return be16_to_cpu(*(__be16 *)mask) & (u16)f->field_mask;
return *(u8 *)mask & (u8)f->field_mask;
*(__be32 *)mask &= ~cpu_to_be32(f->field_mask);
*(__be16 *)mask &= ~cpu_to_be16((u16)f->field_mask);
*(u8 *)mask &= ~(u8)f->field_mask;
unsigned long field_mask = f->field_mask;
start = find_first_bit(&field_mask, f->field_bsize);
u32 field_mask;
field_mask = (1 << bits_per_prio) - 1;
*word &= ~(field_mask << offset);
return (reg_val >> field_offset) & field_mask;
u16 field_mask, u16 field_val)
field_val &= field_mask;
reg_val &= ~(field_mask << field_offset);
reg_val &= ~(field_mask << (field_offset + 8));
u16 field_mask)
u32 field_mask = GENMASK(p->field.msb, p->field.lsb);
val &= ~FIELD_PREP(HSPHY_CFG1_DAT_MASK, field_mask);
data = field_mask & (p->value << p->field.lsb);
u8 field_mask = GENMASK(p->field.msb, p->field.lsb);
val &= ~field_mask;
data = field_mask & (p->value << p->field.lsb);
unsigned long *field, *field_mask;
field_mask = &ctl_mask;
field_mask = &ext_ctl_mask;
field = field_mask = NULL;
__set_bit(bit, field_mask);
#define field_max(field) ((typeof(field))field_mask(field))
if (__builtin_constant_p(v) && (v & ~field_mask(field))) \
return to((v & field_mask(field)) * field_multiplier(field)); \
u32 field_mask;
u32 offset, u32 field_mask);
return rec->field_mask & type;
u32 field_mask, u32 value_size);
#define _IBA_SET(field_struct, field_offset, field_mask, num_bits, ptr, value) \
_iba_set##num_bits((void *)_ptr + (field_offset), field_mask, \
FIELD_PREP(field_mask, value)); \
#define _IBA_GET(field_struct, field_offset, field_mask, num_bits, ptr) \
field_mask, _iba_get##num_bits((const void *)_ptr + \
u32 off, int sz, struct btf_field_info *info, u32 field_mask)
if (!(type & field_mask))
u32 field_mask, u32 *seen_mask, int *align, int *sz)
if (!(field_mask & field_type) || strcmp(name, field_type_name))
if (field_mask & (BPF_KPTR | BPF_UPTR) && !__btf_type_is_struct(var_type)) {
const struct btf_type *t, u32 field_mask,
u32 field_mask, struct btf_field_info *info,
ret = btf_find_struct_field(btf, t, field_mask, info, info_cnt, level);
u32 field_mask, u32 *seen_mask,
field_mask, seen_mask, &align, &sz);
ret = btf_find_nested_struct(btf, var_type, off, nelems, field_mask,
info_cnt ? &info[0] : &tmp, field_mask);
const struct btf_type *t, u32 field_mask,
field_mask, &seen_mask,
u32 field_mask, struct btf_field_info *info,
field_mask, &seen_mask,
u32 field_mask, struct btf_field_info *info,
return btf_find_struct_field(btf, t, field_mask, info, info_cnt, 0);
return btf_find_datasec_var(btf, t, field_mask, info, info_cnt, 0);
u32 field_mask, u32 value_size)
ret = btf_find_field(btf, t, field_mask, info_arr, ARRAY_SIZE(info_arr));
rec->field_mask |= info_arr[i].type;
if (IS_ERR_OR_NULL(rec) || !(rec->field_mask & (BPF_GRAPH_ROOT | BPF_UPTR)))
if (!(rec->field_mask & BPF_GRAPH_NODE))
if (meta->record->field_mask & BPF_GRAPH_ROOT)
for (i = 0; i < sizeof(map->record->field_mask) * 8; i++) {
switch (map->record->field_mask & (1 << i)) {
u32 field_mask)
if (IS_ERR_OR_NULL(rec) || !(rec->field_mask & field_mask))
if (!field || !(field->type & field_mask))
#define field_max(field) ((typeof(field))field_mask(field))
if (__builtin_constant_p(v) && (v & ~field_mask(field))) \
return to((v & field_mask(field)) * field_multiplier(field)); \