BITS_PER_BYTE
#define BITS_PER_TYPE(t) (sizeof(t) * BITS_PER_BYTE)
#define BITS_TO_BYTES(n) howmany((n), BITS_PER_BYTE)
minlen = (nsymbols * nsymbits) / BITS_PER_BYTE;
u8 vsi_list[DIVIDE_AND_ROUND_UP(ICE_MAX_VSI, BITS_PER_BYTE)];
#define BITS_PER_CHUNK (BITS_PER_BYTE * sizeof(ice_bitmap_t))
#define BITS_PER_U32 (sizeof(u32) * BITS_PER_BYTE)
if (ce_info[f].width > (ce_info[f].size_of * BITS_PER_BYTE)) {
BITS_PER_BYTE))
.off = (_offset_bytes) * BITS_PER_BYTE, \
.size = (_size_bytes) * BITS_PER_BYTE, \
ese_bits = ICE_FLOW_FV_EXTRACT_SZ * BITS_PER_BYTE;
sizeof(cluster_mask) * BITS_PER_BYTE) {
bytes_per_sec = DIV_S64((s64)bw * 1000, BITS_PER_BYTE);
bytes_per_sec = DIV_S64((s64)bw * 1000, BITS_PER_BYTE);
#ifndef BITS_PER_BYTE
#define MLX4_EN_MAC_HASH_SIZE (1 << BITS_PER_BYTE)
#define MLX5_L2_ADDR_HASH_SIZE (BIT(BITS_PER_BYTE))
#define MLX5E_ETH_ADDR_HASH_SIZE (1 << BITS_PER_BYTE)
sizeof(accel_dev->au_info->sym_ae_msk) * BITS_PER_BYTE;
#ifndef BITS_PER_BYTE
#define BITS_PER_BYTE 8
byte_index = bit_position / BITS_PER_BYTE;
(1 << (bit_position % BITS_PER_BYTE))) ? true : false;
byte_index = (bit_position / BITS_PER_BYTE) +
(1 << (bit_position % BITS_PER_BYTE))) ? true : false;
byte_index = (bit_position / BITS_PER_BYTE) +
(1 << (bit_position % BITS_PER_BYTE));
#define RVT_QPNMAP_ENTRIES (RVT_QPN_MAX / PAGE_SIZE / BITS_PER_BYTE)
#define RVT_BITS_PER_PAGE (PAGE_SIZE * BITS_PER_BYTE)