totalram_pages
if ((size >> PAGE_SHIFT) > totalram_pages())
const long total_pages = totalram_pages() + __loaned_pages;
oom_freed_pages, totalram_pages());
memory = totalram_pages() >> (20 - PAGE_SHIFT);
memory = totalram_pages() << (PAGE_SHIFT - 20);
totalram_pages() << PAGE_SHIFT,
totalram_pages() << PAGE_SHIFT,
const u64 limit = totalram_pages() << PAGE_SHIFT;
limit = totalram_pages() << PAGE_SHIFT;
const unsigned long max_graphics_pages = totalram_pages() / 2;
unsigned long nr_pages = totalram_pages();
mem = (__u64)mult_frac(totalram_pages() - totalhigh_pages(),
unsigned long pages = (totalram_pages() - totalhigh_pages()) * DM_CRYPT_MEMORY_PERCENT / 100;
if (journal_pages >= totalram_pages() - totalhigh_pages() || journal_desc_size > ULONG_MAX) {
if (a >> PAGE_SHIFT > totalram_pages() / DM_STATS_MEMORY_FACTOR)
#define MAX_WRITEBACK_JOBS min(0x10000000 / PAGE_SIZE, totalram_pages() / 16)
vpu->enable_4GB = !!(totalram_pages() > (SZ_2G >> PAGE_SHIFT));
limit = totalram_pages();
table_size = (totalram_pages() > (1U << 30) / PAGE_SIZE) ? 8192 :
(totalram_pages() << PAGE_SHIFT) /
iova_space_size = (u32) (totalram_pages() / count_parisc_driver(&ccio_driver));
(unsigned long) totalram_pages() >> (20 - PAGE_SHIFT),
iova_space_size = (u32) (totalram_pages()/global_ioc_cnt);
(unsigned long) totalram_pages() >> (20 - PAGE_SHIFT),
congestion_kb = (16*int_sqrt(totalram_pages())) << (PAGE_SHIFT-10);
unsigned long nr_pages = totalram_pages();
*limit = ((totalram_pages() << PAGE_SHIFT) >> 13) / 392;
nfs_congestion_kb = (16*int_sqrt(totalram_pages())) << (PAGE_SHIFT-10);
unsigned long low_pages = totalram_pages() - totalhigh_pages();
unsigned long totalpages = totalram_pages() + total_swap_pages;
unsigned long pages = totalram_pages() / (SZ_1G / SZ_128K);
unsigned long nr_pages = totalram_pages();
image_size = ((totalram_pages() * 2) / 5) * PAGE_SIZE;
totalram_pages();
if (totalram_pages() < MB_TO_PAGES(512)) {
total_size = (totalram_pages() << PAGE_SHIFT) /
unsigned long ram_pages = totalram_pages();
K(physpages - totalram_pages() - totalcma_pages),
oc->totalpages = totalram_pages() + total_swap_pages;
return totalram_pages() / 2;
unsigned long nr_pages = totalram_pages();
size *= totalram_pages();
val->totalram = totalram_pages();
unsigned long megs = PAGES_TO_MB(totalram_pages());
allowed = ((totalram_pages() - hugetlb_total_pages())
if (pages > totalram_pages() + total_swap_pages)
if (count > totalram_pages())
if ((size >> PAGE_SHIFT) > totalram_pages()) {
if ((end - start) >> PAGE_SHIFT > totalram_pages())
max_order = fls_long(totalram_pages() - 1);
return totalram_pages() * zswap_max_pool_percent / 100;
nr_entries = div64_ul((u64)totalram_pages() << PAGE_SHIFT,
if (totalram_pages() >= 128 * 1024)
max_avail = order_base_2(totalram_pages()) + PAGE_SHIFT;
unsigned long nr_pages = totalram_pages();
unsigned long nr_pages = totalram_pages();
unsigned long nr_pages = totalram_pages();
((kexec_segment_size >> PAGE_SHIFT) > totalram_pages() / 2)) {