mmap_size
hsi_if->mmap_size = mmap_size;
cs_char_data.mmap_size = CS_MMAP_SIZE;
unsigned long mmap_size;
} else if ((buf_size_aligned + ctrl_size_aligned) >= hi->mmap_size) {
unsigned long mmap_size;
unsigned long mmap_base, unsigned long mmap_size)
aq->mmap_size = resource_size(res);
resource_size_t mmap_size;
if (op->addr.val + op->data.nbytes > aq->mmap_size)
size_t mmap_size;
if (op->addr.val < qspi->mmap_size) {
if (op->addr.val + op->data.nbytes > qspi->mmap_size) {
max_len = qspi->mmap_size - op->addr.val;
if (from + op->data.nbytes > qspi->mmap_size)
qspi->mmap_size = resource_size(res_mmap);
unsigned long mmap_size;
ctx->mmap_size = nr_pages * PAGE_SIZE;
pr_debug("attempting mmap of %lu bytes\n", ctx->mmap_size);
ctx->mmap_size = 0;
ctx->mmap_base = do_mmap(ctx->aio_ring_file, 0, ctx->mmap_size,
ctx->mmap_size = 0;
if (ctx->mmap_size)
vm_munmap(ctx->mmap_base, ctx->mmap_size);
if (ctx->mmap_size)
vm_munmap(ctx->mmap_base, ctx->mmap_size);
ctx->mmap_size = 0;
size_t mmap_size = 0;
mmap_data = bpf_map__initial_value(map, &mmap_size);
print_hex(mmap_data, mmap_size);
ehdr = base + (data_tail & (mmap_size - 1));
if (((void *)ehdr) + ehdr_size > base + mmap_size) {
size_t len_first = base + mmap_size - copy_start;
munmap(cpu_buf->base, pb->mmap_size + pb->page_size))
cpu_buf->base = mmap(NULL, pb->mmap_size + pb->page_size,
pb->mmap_size = pb->page_size * page_cnt;
ret = perf_event_read_simple(cpu_buf->base, pb->mmap_size,
*buf_size = pb->mmap_size;
buffer->mmap_size = size;
munmap(buffer->mmap_addr, buffer->mmap_size);
buffer->mmap_size = 0;
size_t mmap_size;
prefetch_event(char *buf, u64 head, size_t mmap_size,
if (head + sizeof(event->header) > mmap_size)
if (head + event_size <= mmap_size)
if (event_size <= mmap_size - head % page_size) {
" fuzzed or compressed perf.data?\n", __func__, head, event_size, mmap_size);
fetch_mmaped_event(u64 head, size_t mmap_size, char *buf, bool needs_swap)
return prefetch_event(buf, head, mmap_size, needs_swap, ERR_PTR(-EINVAL));
fetch_decomp_event(u64 head, size_t mmap_size, char *buf, bool needs_swap)
return prefetch_event(buf, head, mmap_size, needs_swap, NULL);
size_t mmap_size;
rd->mmap_size = MMAP_SIZE;
if (rd->mmap_size > data_size) {
rd->mmap_size = data_size;
munmap(mmaps[rd->mmap_idx], rd->mmap_size);
buf = mmap(NULL, rd->mmap_size, mmap_prot, mmap_flags, rd->fd,
event = fetch_mmaped_event(rd->head, rd->mmap_size, rd->mmap_cur,
size_t mmap_size;
mmap_size = ROUND_UP_TO_PAGE_SIZE(size);
mmio = mmap(0, mmap_size, PROT_READ, MAP_SHARED, fd_pmt, 0);
size_t mmap_size;
size_t mmap_mask = pb->mmap_size - 1;
size_t mmap_size;
region->mmap_size = mem_size;
region->mmap_size += alignment;
region->fd = kvm_memfd_alloc(region->mmap_size,
region->mmap_start = kvm_mmap(region->mmap_size, PROT_READ | PROT_WRITE,
region->mmap_alias = kvm_mmap(region->mmap_size,
kvm_munmap(region->mmap_start, region->mmap_size);
kvm_munmap(region->mmap_alias, region->mmap_size);
size_t mmap_size;
mmap_size = 2 * pmdsize;
mmap_mem = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE,
mmap_smem = mmap(NULL, mmap_size, PROT_READ,
munmap(mmap_mem, mmap_size);
munmap(mmap_smem, mmap_size);
size_t size, mmap_size, mremap_size;
mmap_size = 2 * thpsize;
mmap_mem = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE,
munmap(mmap_mem, mmap_size);
munmap(huge_ptr, mmap_size);
size_t mmap_size;
madvise(huge_ptr, mmap_size, MADV_DONTNEED);
ptr = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE,
mmap_size = default_huge_page_size();
huge_ptr = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE,
size_t mmap_size;
mem = mmap_thp_range(PROT_READ, &mmap_mem, &mmap_size);
munmap(mmap_mem, mmap_size);
size_t mmap_size;
mem = mmap_thp_range(PROT_READ|PROT_WRITE, &mmap_mem, &mmap_size);
munmap(mmap_mem, mmap_size);
size_t mmap_size;
mem = mmap_thp_range(PROT_READ|PROT_WRITE, &mmap_mem, &mmap_size);
munmap(mmap_mem, mmap_size);
const size_t mmap_size = 2 * thpsize;
mmap_mem = mmap(NULL, mmap_size, prot, MAP_PRIVATE|MAP_ANON,
munmap(mmap_mem, mmap_size);
*_mmap_size = mmap_size;
size_t mmap_size;
mmap_size = 2 * pmdsize;
mmap_mem = (char *)mmap(NULL, mmap_size, PROT_READ | PROT_WRITE,
munmap(mmap_mem, mmap_size);
size_t mmap_size = size * 2;
mmap_mem = mmap(NULL, mmap_size, prot, flags, -1, 0);
munmap(mem + size, mmap_mem + mmap_size - mem - size);
size_t mmap_size;
mmap_size = page_size * (1 + mmap_pages);
buff = mmap(NULL, mmap_size,
size_t mmap_size;
self->mmap_size = 2 * region_size;
region->vaddr = mmap(NULL, self->mmap_size, PROT_READ | PROT_WRITE,
ASSERT_EQ(munmap(self->region.vaddr, self->mmap_size), 0);
region->size = self->mmap_size;