ne_mem_region
ne_mem_region->memory_size = mem_region.memory_size;
ne_mem_region->userspace_addr = mem_region.userspace_addr;
list_add(&ne_mem_region->mem_region_list_entry, &ne_enclave->mem_regions_list);
for (i = 0; i < ne_mem_region->nr_pages; i++)
put_page(ne_mem_region->pages[i]);
kfree(ne_mem_region->pages);
kfree(ne_mem_region);
struct ne_mem_region *ne_mem_region = NULL;
struct ne_mem_region *ne_mem_region_tmp = NULL;
list_for_each_entry_safe(ne_mem_region, ne_mem_region_tmp,
list_del(&ne_mem_region->mem_region_list_entry);
for (i = 0; i < ne_mem_region->nr_pages; i++)
put_page(ne_mem_region->pages[i]);
kfree(ne_mem_region->pages);
kfree(ne_mem_region);
struct ne_mem_region *ne_mem_region = NULL;
list_for_each_entry(ne_mem_region, &ne_enclave->mem_regions_list,
u64 memory_size = ne_mem_region->memory_size;
u64 userspace_addr = ne_mem_region->userspace_addr;
struct ne_mem_region *ne_mem_region = NULL;
ne_mem_region = kzalloc_obj(*ne_mem_region);
if (!ne_mem_region)
ne_mem_region->pages = kzalloc_objs(*ne_mem_region->pages, max_nr_pages);
if (!ne_mem_region->pages) {
i = ne_mem_region->nr_pages;
ne_mem_region->pages + i, FOLL_GET);
rc = ne_sanity_check_user_mem_region_page(ne_enclave, ne_mem_region->pages[i]);
page_to_phys(ne_mem_region->pages[i]),
page_size(ne_mem_region->pages[i]));
memory_size += page_size(ne_mem_region->pages[i]);
ne_mem_region->nr_pages++;