SHMLBA
(addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))
#define CACHE_COLOUR(vaddr) ((vaddr & (SHMLBA - 1)) >> PAGE_SHIFT)
#if SHMLBA > 16384
(addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))
info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0;
((((addr)+SHMLBA-1)&~(SHMLBA-1)) + \
(((pgoff)<<PAGE_SHIFT) & (SHMLBA-1)))
(addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))
info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0;
return (addr1 ^ addr2) & (SHMLBA-1);
((((addr)+SHMLBA-1)&~(SHMLBA-1)) + \
(((pgoff)<<PAGE_SHIFT) & (SHMLBA-1)))
(addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))
info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0;
#define SHM_ALIGN_MASK (SHMLBA - 1)
((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)))
(PAGE_MASK & (SHMLBA - 1)) : 0;
((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)))
((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)))
err = do_shmat(first, ptr, (int)second, &raddr, SHMLBA);
unsigned long base = (addr+SHMLBA-1)&~(SHMLBA-1);
unsigned long off = (pgoff<<PAGE_SHIFT) & (SHMLBA-1);
return PAGE_MASK & (SHMLBA - 1);
((((addr) + SHMLBA - 1) & ~(SHMLBA - 1)) + \
(((pgoff) << PAGE_SHIFT) & (SHMLBA - 1)))
err = do_shmat(shmid, shmaddr, shmflg, &ret, SHMLBA);
((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)))
rdi->mmap_offset = ALIGN(PAGE_SIZE, SHMLBA);
rdi->mmap_offset += ALIGN(size, SHMLBA);
rxe->mmap_offset = ALIGN(PAGE_SIZE, SHMLBA);
rxe->mmap_offset += ALIGN(size, SHMLBA);
err = do_shmat(shmid, shmaddr, shmflg, &ret, SHMLBA);
#define COMPAT_SHMLBA SHMLBA
#define COMPAT_SHMLBA SHMLBA
second, &raddr, SHMLBA);
align = SHMLBA;
return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,