B_FULL_LOCK
uint32 lock = B_FULL_LOCK);
size, B_FULL_LOCK | ((flags & B_APERTURE_NEED_PHYSICAL) != 0
&scratchAddress, B_ANY_KERNEL_ADDRESS, B_PAGE_SIZE, B_FULL_LOCK,
if (ainfo.lock < B_FULL_LOCK || where.size < low_size) {
B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA);
B_FULL_LOCK,
B_FULL_LOCK,
B_ANY_ADDRESS, size, B_FULL_LOCK, B_READ_AREA | B_WRITE_AREA);
di->sharedArea = create_area(shared_name, (void **)&(di->si), B_ANY_KERNEL_ADDRESS, ((sizeof(ET6000SharedInfo) + (B_PAGE_SIZE - 1)) & ~(B_PAGE_SIZE - 1)), B_FULL_LOCK,
B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA | B_CLONEABLE_AREA);
ROUND_TO_PAGE_SIZE(sharedSize), B_FULL_LOCK,
B_FULL_LOCK,
B_ANY_KERNEL_ADDRESS, gttSize, B_FULL_LOCK | B_CONTIGUOUS,
B_ANY_ADDRESS, si.videoMemSize, B_FULL_LOCK,
B_FULL_LOCK,
((sizeof(shared_info) + (B_PAGE_SIZE - 1)) & ~(B_PAGE_SIZE - 1)), B_FULL_LOCK,
di->shared_area = create_area(shared_name, (void **)&(di->si), B_ANY_KERNEL_ADDRESS, ((sizeof(shared_info) + (B_PAGE_SIZE - 1)) & ~(B_PAGE_SIZE - 1)), B_FULL_LOCK,
((sizeof(shared_info) + (B_PAGE_SIZE - 1)) & ~(B_PAGE_SIZE - 1)), B_FULL_LOCK,
map_area_size, B_FULL_LOCK, B_READ_AREA | B_WRITE_AREA);
size, B_FULL_LOCK,
B_FULL_LOCK,
B_FULL_LOCK,
ROUND_TO_PAGE_SIZE(sizeof(radeon_shared_info)), B_FULL_LOCK,
B_FULL_LOCK,
di->shared_area = create_area(shared_name, (void **)&(di->si), B_ANY_KERNEL_ADDRESS, ((sizeof(shared_info) + (B_PAGE_SIZE - 1)) & ~(B_PAGE_SIZE - 1)), B_FULL_LOCK,
ROUND_TO_PAGE_SIZE(sharedSize + modesSize), B_FULL_LOCK,
di->shared_area = create_area(shared_name, (void **)&(di->si), B_ANY_KERNEL_ADDRESS, ((sizeof(shared_info) + (B_PAGE_SIZE - 1)) & ~(B_PAGE_SIZE - 1)), B_FULL_LOCK,
B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA);
ROUND_TO_PAGE_SIZE(sharedSize), B_FULL_LOCK,
B_FULL_LOCK | B_CONTIGUOUS, B_READ_AREA | B_WRITE_AREA);
B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA);
B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA);
3, B_ANY_ADDRESS, B_FULL_LOCK );
case B_FULL_LOCK:
B_FULL_LOCK, B_READ_AREA | B_WRITE_AREA | B_CLONEABLE_AREA);
PAGE_ALIGN(areaSize), B_FULL_LOCK,
PAGE_ALIGN(areaSize), B_FULL_LOCK,
B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA);
PAGE_ALIGN(areaSize), B_FULL_LOCK,
PAGE_ALIGN(areaSize), B_FULL_LOCK,
B_ANY_ADDRESS, COMMPAGE_SIZE, B_FULL_LOCK,
areaInfo->fLock = B_FULL_LOCK;
B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA,
create_area("guarded heap meta", &meta, B_ANY_KERNEL_ADDRESS, growSize, B_FULL_LOCK,
B_ANY_KERNEL_BLOCK_ADDRESS, size, B_FULL_LOCK,
B_ANY_KERNEL_BLOCK_ADDRESS, HEAP_DEDICATED_GROW_SIZE, B_FULL_LOCK,
B_ANY_KERNEL_ADDRESS, VIP_HEAP_SIZE, B_FULL_LOCK,
(void **)&base, B_ANY_KERNEL_ADDRESS, size, B_FULL_LOCK,
B_ANY_KERNEL_BLOCK_ADDRESS, areaSize, B_FULL_LOCK,
B_ANY_KERNEL_ADDRESS, areaSize, B_FULL_LOCK,
B_PAGE_SIZE, B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA,
B_EXACT_ADDRESS, region->size, B_FULL_LOCK,
B_ANY_KERNEL_ADDRESS, kMessagingAreaSize, B_FULL_LOCK,
sizeof(struct sem_entry) * sMaxSems, B_FULL_LOCK,
"slab large raw allocation", size, B_FULL_LOCK,
size, B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA);
kTeamUserDataInitialSize, B_FULL_LOCK,
B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA
case B_FULL_LOCK:
wiring = B_FULL_LOCK;
wiring = B_FULL_LOCK;
if (wiring == B_FULL_LOCK || wiring == B_CONTIGUOUS) {
if (wiring == B_FULL_LOCK)
case B_FULL_LOCK:
B_FULL_LOCK, protection, 0, REGION_NO_PRIVATE_MAP, CREATE_AREA_DONT_COMMIT_MEMORY,
B_FULL_LOCK, protection, 0, REGION_NO_PRIVATE_MAP, CREATE_AREA_DONT_COMMIT_MEMORY,
if (newArea->wiring == B_FULL_LOCK) {
|| area->wiring == B_FULL_LOCK
|| area->wiring == B_FULL_LOCK
case B_FULL_LOCK:
B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA,
B_ANY_KERNEL_ADDRESS, size, B_FULL_LOCK,
B_ANY_KERNEL_ADDRESS, size, B_FULL_LOCK,