BUFFER_SIZE
dma_cache_maintenance( paddr, BUFFER_SIZE, 1 );
dma_wd.fdc_acces_seccount = BUFFER_SIZE/512;
DMABuffer = atari_stram_alloc(BUFFER_SIZE+512, "ataflop");
FILL( TrackBuffer+BUFFER_SIZE-p, 0x4e );
uint8_t BUFFER_SIZE:8;
ps2port->act = (ps2port->act + 1) & BUFFER_SIZE;
} buffer[BUFFER_SIZE+1];
BUFFER_SIZE);
cp->buffer_storage[idx] = kmalloc(BUFFER_SIZE,GFP_KERNEL);
u8 scratch[BUFFER_SIZE];
test->buffer = kzalloc(BUFFER_SIZE, GFP_KERNEL);
for (i = 0; i < BUFFER_SIZE / 512; i++) {
memset(test->scratch, 0, BUFFER_SIZE);
sg_copy_from_buffer(sg, sg_len, test->scratch, BUFFER_SIZE);
if ((sectors * 512) > BUFFER_SIZE)
sg_copy_to_buffer(sg, sg_len, test->scratch, BUFFER_SIZE);
#define MIRROR_SIZE (BUFFER_SIZE * 4)
#define MIRROR_SIZE (BUFFER_SIZE * 4)
unsigned char *buffer = kmalloc (BUFFER_SIZE, GFP_KERNEL);
int i, ret = BUFFER_SIZE;
memset (buffer, byte, BUFFER_SIZE);
ether1_writebuffer (dev, buffer, 0, BUFFER_SIZE);
memset (buffer, byte ^ 0xff, BUFFER_SIZE);
ether1_readbuffer (dev, buffer, 0, BUFFER_SIZE);
for (i = 0; i < BUFFER_SIZE; i++) {
printk (" - 0x%04X\n", BUFFER_SIZE);
char write[BUFFER_SIZE];
u8 buffer[BUFFER_SIZE];
} __aligned(BUFFER_SIZE);
if (len > BUFFER_SIZE * BUFFER_NUM) {
n_blocks = (len + BUFFER_SIZE - 1) / BUFFER_SIZE;
qtd->qtd_buffer_len = BUFFER_SIZE * a_blocks;
/ BUFFER_SIZE;
static unsigned char read_buffers[READ_BUFFERS][BUFFER_SIZE];
if (blk_offset > BUFFER_SIZE ||
blk_offset + len > BUFFER_SIZE)
HW_PARAM_ENTRY(BUFFER_SIZE), \
mte_free_memory_tag_range(ptr, BUFFER_SIZE, mem_type, 0, MT_GRANULE_SIZE);
ptr = mte_allocate_memory(BUFFER_SIZE + MT_GRANULE_SIZE, mem_type, 0, false);
if (check_allocated_memory(ptr, BUFFER_SIZE + MT_GRANULE_SIZE,
ptr = (char *)mte_insert_tags(ptr, BUFFER_SIZE);
mte_free_memory_tag_range(ptr, BUFFER_SIZE, mem_type, 0, MT_GRANULE_SIZE);
ptr = mte_allocate_memory(BUFFER_SIZE, mem_type, 0, false);
if (check_allocated_memory(ptr, BUFFER_SIZE, mem_type, false) != KSFT_PASS)
ptr = (char *)mte_insert_tags(ptr, BUFFER_SIZE);
mte_free_memory((void *)ptr, BUFFER_SIZE, mem_type, true);
mte_initialize_current_context(mode, (uintptr_t)ptr, BUFFER_SIZE);
memset(ptr, '1', BUFFER_SIZE);
mte_free_memory(ptr, BUFFER_SIZE, mem_type, false);
mte_initialize_current_context(mode, (uintptr_t)ptr, BUFFER_SIZE);
memset(ptr, '1', BUFFER_SIZE);
ptr, ptr + BUFFER_SIZE, mode);
mte_initialize_current_context(mode, (uintptr_t)ptr, BUFFER_SIZE + 1);
ptr[BUFFER_SIZE] = '2';
ptr = mte_allocate_memory(BUFFER_SIZE + MT_GRANULE_SIZE, mem_type, 0, false);
if (check_allocated_memory(ptr, BUFFER_SIZE + MT_GRANULE_SIZE,
ptr = mte_insert_tags(ptr, BUFFER_SIZE);
mte_free_memory_tag_range(ptr, BUFFER_SIZE, mem_type, 0, MT_GRANULE_SIZE);
ptr = mte_allocate_memory(BUFFER_SIZE + MT_GRANULE_SIZE, mem_type, 0, false);
if (check_allocated_memory(ptr, BUFFER_SIZE + MT_GRANULE_SIZE,
ptr = mte_insert_tags(ptr, BUFFER_SIZE);
uint8_t buffer[BUFFER_SIZE];
.num_iovas = BUFFER_SIZE / sizeof(*ranges),
memset(ranges, 0, BUFFER_SIZE);
ranges_cmd.num_iovas = BUFFER_SIZE / sizeof(*ranges);
memset(ranges, 0, BUFFER_SIZE);
.length = BUFFER_SIZE,
.length = BUFFER_SIZE,
for (npages = 1; npages < BUFFER_SIZE / PAGE_SIZE; npages++) {
test_ioctl_ioas_map_fixed(buffer, BUFFER_SIZE,
test_ioctl_ioas_unmap(MOCK_APERTURE_START, BUFFER_SIZE);
.length = BUFFER_SIZE,
test_ioctl_ioas_map_fixed(buffer, BUFFER_SIZE, MOCK_APERTURE_START);
test_ioctl_ioas_unmap(MOCK_APERTURE_START, BUFFER_SIZE);
for (i = 0; i != BUFFER_SIZE / sizeof(*buffer16); i++)
tmp2 = malloc(BUFFER_SIZE);
access_cmd.access_rw.length = BUFFER_SIZE;
test_ioctl_ioas_map(buffer, BUFFER_SIZE, &iova);
test_ioctl_ioas_unmap(iova, BUFFER_SIZE);
test_ioctl_ioas_map_fixed(buffer, BUFFER_SIZE, iova);
test_ioctl_ioas_unmap(iova, BUFFER_SIZE);
test_ioctl_ioas_map_fixed(buffer, BUFFER_SIZE,
test_ioctl_ioas_map_fixed(buffer, BUFFER_SIZE,
last_iova = MOCK_APERTURE_START + BUFFER_SIZE - 1;
last_iova = MOCK_APERTURE_START + BUFFER_SIZE - 2;
.access_pages = { .length = BUFFER_SIZE,
.length = BUFFER_SIZE,
.length = BUFFER_SIZE,
test_ioctl_ioas_map_id_file(ioas_id, mfd, 0, BUFFER_SIZE,
test_ioctl_ioas_map_id(ioas_id, buf, BUFFER_SIZE,
check_mock_iova(buf, MOCK_APERTURE_START, BUFFER_SIZE);
test_ioctl_ioas_map_id_file(new_ioas_id, mfd, 0, BUFFER_SIZE,
test_ioctl_ioas_map_id(new_ioas_id, buf, BUFFER_SIZE,
check_mock_iova(buf, MOCK_APERTURE_START, BUFFER_SIZE);
memset(buffer + _cmd_len, 0xAA, BUFFER_SIZE - _cmd_len); \
BUFFER_SIZE - _cmd_len)); \
.size = BUFFER_SIZE,
.size = BUFFER_SIZE,
unsigned long pages_iova[BUFFER_SIZE / PAGE_SIZE];
ASSERT_EQ(BUFFER_SIZE, unmap_cmd.size);
ASSERT_EQ(BUFFER_SIZE, unmap_cmd.size);
BUFFER_SIZE = PAGE_SIZE * 16;
rc = posix_memalign(&buffer, HUGEPAGE_SIZE, BUFFER_SIZE);
vrc = mmap(buffer, BUFFER_SIZE, PROT_READ | PROT_WRITE,
mfd_buffer = memfd_mmap(BUFFER_SIZE, PROT_READ | PROT_WRITE, MAP_SHARED,
ranges[0].last = BUFFER_SIZE + 16 * 1024 * 600 - 1;
if (_test_ioctl_ioas_map(self->fd, ioas_id, buffer, BUFFER_SIZE, &iova,
if (_test_ioctl_ioas_unmap(self->fd, ioas_id, iova, BUFFER_SIZE,
_test_ioctl_ioas_unmap(self->fd, ioas_id, iova, BUFFER_SIZE, NULL);
BUFFER_SIZE = 2*1024*1024;
buffer = mmap(0, BUFFER_SIZE, PROT_READ | PROT_WRITE,
mfd_buffer = memfd_mmap(BUFFER_SIZE, PROT_READ | PROT_WRITE, MAP_SHARED,
if (_test_ioctl_ioas_map(self->fd, ioas_id, buffer, BUFFER_SIZE, &iova,
.length = BUFFER_SIZE,
if (_test_ioctl_ioas_map(self->fd, ioas_id, buffer, BUFFER_SIZE, &iova,
.length = BUFFER_SIZE,
static unsigned long BUFFER_SIZE;
.check_refs = { .length = BUFFER_SIZE,