KASAN_GRANULE_SIZE
kasan_poison(object, round_up(cache->object_size, KASAN_GRANULE_SIZE),
kasan_poison(object, round_up(cache->object_size, KASAN_GRANULE_SIZE),
KASAN_GRANULE_SIZE);
KASAN_GRANULE_SIZE);
redzone_start = round_up((unsigned long)(ptr + size), KASAN_GRANULE_SIZE);
new_size = round_up(new_size, KASAN_GRANULE_SIZE);
old_size = round_up(old_size, KASAN_GRANULE_SIZE);
old_size = round_down(old_size, KASAN_GRANULE_SIZE);
return shadow_byte >= 0 && shadow_byte < KASAN_GRANULE_SIZE;
size_t aligned_size = round_up(global->size, KASAN_GRANULE_SIZE);
size_t rounded_up_size = round_up(size, KASAN_GRANULE_SIZE);
size_t rounded_down_size = round_down(size, KASAN_GRANULE_SIZE);
if (unlikely(!IS_ALIGNED((unsigned long)addr, KASAN_GRANULE_SIZE)))
KASAN_GRANULE_SIZE);
#define KASAN_GRANULE_MASK (KASAN_GRANULE_SIZE - 1)
#define KASAN_MEMORY_PER_SHADOW_PAGE (KASAN_GRANULE_SIZE << PAGE_SHIFT)
#define META_MEM_BYTES_PER_ROW (META_BYTES_PER_ROW * KASAN_GRANULE_SIZE)
size = round_up(size, KASAN_GRANULE_SIZE);
((volatile char *)&elem[round_up(size, KASAN_GRANULE_SIZE)])[0]);
size_t size = 128 - KASAN_GRANULE_SIZE - 5;
src = kmalloc(KASAN_GRANULE_SIZE, GFP_KERNEL | __GFP_ZERO);
strscpy(src, "f0cacc1a0000000", KASAN_GRANULE_SIZE);
KUNIT_EXPECT_EQ(test, KASAN_GRANULE_SIZE - 2,
strscpy(ptr, src + 1, KASAN_GRANULE_SIZE));
KUNIT_EXPECT_KASAN_FAIL_READ(test, strscpy(ptr, src + KASAN_GRANULE_SIZE,
KASAN_GRANULE_SIZE));
size = round_up(size, KASAN_GRANULE_SIZE);
size_t size = PAGE_SIZE / 2 - KASAN_GRANULE_SIZE - 5;
size -= KASAN_GRANULE_SIZE + 1;
size += 2 * KASAN_GRANULE_SIZE + 2;
size_t size = 128 - KASAN_GRANULE_SIZE - 5;
kasan_poison(ptr, KASAN_GRANULE_SIZE, (u8)tag, false);
kasan_poison(ptr, KASAN_GRANULE_SIZE, get_tag(ptr), false);
ptr = kmalloc(size - KASAN_GRANULE_SIZE, GFP_KERNEL);
size_t size = 128 - KASAN_GRANULE_SIZE;
ptr[size + KASAN_GRANULE_SIZE + 5]);
size_t size = 128 - KASAN_GRANULE_SIZE;
#define OOB_TAG_OFF (IS_ENABLED(CONFIG_KASAN_GENERIC) ? 0 : KASAN_GRANULE_SIZE)
ptr2[round_up(size2, KASAN_GRANULE_SIZE)] = 'x');
ptr2[round_up(size2, KASAN_GRANULE_SIZE)] = 'x');
KUNIT_EXPECT_LE(test, round_up(size2, KASAN_GRANULE_SIZE),
round_down(middle, KASAN_GRANULE_SIZE));
KUNIT_EXPECT_LE(test, round_up(middle, KASAN_GRANULE_SIZE),
round_down(size1, KASAN_GRANULE_SIZE));
size_t size = 128 - KASAN_GRANULE_SIZE;
size_t size = 128 - KASAN_GRANULE_SIZE;
size_t size = 128 - KASAN_GRANULE_SIZE;
size_t size = 128 - KASAN_GRANULE_SIZE;
size_t size = 128 - KASAN_GRANULE_SIZE;
memset(ptr, 0, size + KASAN_GRANULE_SIZE));
size_t size = 128 - KASAN_GRANULE_SIZE - 5;
int size = 128 - KASAN_GRANULE_SIZE;
(addr - row) / KASAN_GRANULE_SIZE * 3 + 1;
orig_addr, orig_addr + KASAN_GRANULE_SIZE - 1);
mem_ptr = round_down(aligned_addr, KASAN_GRANULE_SIZE);
mem_ptr -= KASAN_GRANULE_SIZE;
mem_ptr -= KASAN_GRANULE_SIZE;
frame = (const unsigned long *)(mem_ptr + KASAN_GRANULE_SIZE);
p += KASAN_GRANULE_SIZE;
size += KASAN_GRANULE_SIZE;
else if (*shadow >= 1 && *shadow <= KASAN_GRANULE_SIZE - 1)
if (*shadow_addr > 0 && *shadow_addr <= KASAN_GRANULE_SIZE - 1)
case 0 ... KASAN_GRANULE_SIZE - 1:
memory_tag = hw_get_mem_tag(object + i * KASAN_GRANULE_SIZE);
size += KASAN_GRANULE_SIZE;
buffer[i] = hw_get_mem_tag(row + i * KASAN_GRANULE_SIZE);
p += KASAN_GRANULE_SIZE;
size += KASAN_GRANULE_SIZE;
kasan_poison(addr, round_up(size, KASAN_GRANULE_SIZE), tag, false);
if (WARN_ON(mem_data->nr_pages % KASAN_GRANULE_SIZE) ||
size = round_up(size, KASAN_GRANULE_SIZE);
scaled_size = (size + KASAN_GRANULE_SIZE - 1) >>