compound_head
pageshift = page_shift(compound_head(page));
struct page *head_page = compound_head(page);
return page_shift(compound_head(page)) >= it_page_shift;
return compound_head(page);
page = compound_head(page);
unsigned long compound_head; /* Bit zero is set */
FOLIO_MATCH(compound_head, lru);
FOLIO_MATCH(compound_head, _head_1);
FOLIO_MATCH(compound_head, _head_2);
FOLIO_MATCH(compound_head, _head_3);
TABLE_MATCH(compound_head, pt_list);
TABLE_MATCH(compound_head, _pt_pad_1);
page = compound_head(page);
unsigned long head = READ_ONCE(page[1].compound_head);
unsigned long head = READ_ONCE(page->compound_head);
return READ_ONCE(page->compound_head) & 1 || page_is_fake_head(page);
READ_ONCE(page->compound_head) & 1;
VM_BUG_ON_PGFLAGS(page->compound_head & 1, page);
VM_BUG_ON_PGFLAGS(page->compound_head & 1, page);
#define PF_HEAD(page, enforce) PF_POISONED_CHECK(compound_head(page))
PF_POISONED_CHECK(compound_head(page)); })
WRITE_ONCE(page->compound_head, (unsigned long)head + 1);
WRITE_ONCE(page->compound_head, 0);
page = compound_head(page);
return page_to_netmem(compound_head(netmem_to_page(netmem)));
if (compound_head(pages[i]) == hpage)
if (compound_head(imu->bvec[j].bv_page) == hpage)
hpage = compound_head(pages[i]);
struct page *p = compound_head(page_array[j]);
VMCOREINFO_OFFSET(page, compound_head);
head = compound_head(page);
page = compound_head(page);
&& compound_head(page) == compound_head(tree_page);
size = page_size(compound_head(page));
if (unlikely(compound_head(page) != head_page)) {
page = compound_head(page);
if (WARN_ON_ONCE(first_page != compound_head(first_page)))
struct page *head = compound_head(page);
head = READ_ONCE(page->compound_head);
SLAB_MATCH(compound_head, slab_cache); /* Ensure bit 0 is clear */
page = compound_head(page);
head = ps->page_snapshot.compound_head;
head = compound_head(pages[n]);
head = compound_head(pages[n]);