arch/arm64/include/asm/hugetlb.h
19
extern bool arch_hugetlb_migration_supported(struct hstate *h);
arch/arm64/mm/hugetlbpage.c
320
unsigned long hugetlb_mask_last_page(struct hstate *h)
arch/arm64/mm/hugetlbpage.c
65
bool arch_hugetlb_migration_supported(struct hstate *h)
arch/powerpc/include/asm/book3s/64/hugetlb.h
19
static inline int hstate_get_psize(struct hstate *hstate)
arch/powerpc/include/asm/book3s/64/hugetlb.h
23
shift = huge_page_shift(hstate);
arch/powerpc/include/asm/kvm_book3s_asm.h
140
struct kvmppc_host_state hstate;
arch/powerpc/kernel/asm-offsets.c
524
# define HSTATE_FIELD(x, f) DEFINE(x, offsetof(struct kvmppc_book3s_shadow_vcpu, hstate.f))
arch/powerpc/mm/book3s64/radix_hugetlbpage.c
13
struct hstate *hstate = hstate_file(vma->vm_file);
arch/powerpc/mm/book3s64/radix_hugetlbpage.c
15
psize = hstate_get_psize(hstate);
arch/powerpc/mm/book3s64/radix_hugetlbpage.c
22
struct hstate *hstate = hstate_file(vma->vm_file);
arch/powerpc/mm/book3s64/radix_hugetlbpage.c
24
psize = hstate_get_psize(hstate);
arch/powerpc/mm/book3s64/radix_hugetlbpage.c
32
struct hstate *hstate = hstate_file(vma->vm_file);
arch/powerpc/mm/book3s64/radix_hugetlbpage.c
34
psize = hstate_get_psize(hstate);
arch/powerpc/mm/book3s64/slice.c
641
struct hstate *hstate = hstate_file(file);
arch/powerpc/mm/book3s64/slice.c
643
return shift_to_mmu_psize(huge_page_shift(hstate));
arch/powerpc/mm/hugetlbpage.c
107
static int __init pseries_alloc_bootmem_huge_page(struct hstate *hstate)
arch/powerpc/mm/hugetlbpage.c
115
m->hstate = hstate;
arch/powerpc/mm/hugetlbpage.c
127
int __init alloc_bootmem_huge_page(struct hstate *h, int nid)
arch/powerpc/mm/nohash/e500_hugetlbpage.c
189
struct hstate *hstate = hstate_file(vma->vm_file);
arch/powerpc/mm/nohash/e500_hugetlbpage.c
190
unsigned long tsize = huge_page_shift(hstate) - 10;
arch/powerpc/mm/pgtable.c
290
struct hstate *h = hstate_vma(vma);
arch/riscv/include/asm/hugetlb.h
15
bool arch_hugetlb_migration_supported(struct hstate *h);
arch/riscv/mm/hugetlbpage.c
128
unsigned long hugetlb_mask_last_page(struct hstate *h)
arch/riscv/mm/hugetlbpage.c
434
bool arch_hugetlb_migration_supported(struct hstate *h)
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_crtc.c
419
struct atmel_hlcdc_crtc_state *hstate;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_crtc.c
446
hstate = drm_crtc_state_to_atmel_hlcdc_crtc_state(state);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_crtc.c
447
hstate->output_mode = fls(output_fmts) - 1;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_crtc.c
451
hstate->output_mode -= 4;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_crtc.c
452
hstate->dpi = 1;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_crtc.c
454
hstate->dpi = 0;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
725
struct atmel_hlcdc_plane_state *hstate =
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
728
struct drm_framebuffer *fb = hstate->base.fb;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
734
if (!hstate->base.crtc || WARN_ON(!fb))
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
746
hstate->src_x = s->src.x1 >> 16;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
747
hstate->src_y = s->src.y1 >> 16;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
748
hstate->src_w = drm_rect_width(&s->src) >> 16;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
749
hstate->src_h = drm_rect_height(&s->src) >> 16;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
750
hstate->crtc_x = s->dst.x1;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
751
hstate->crtc_y = s->dst.y1;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
752
hstate->crtc_w = drm_rect_width(&s->dst);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
753
hstate->crtc_h = drm_rect_height(&s->dst);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
755
hstate->nplanes = fb->format->num_planes;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
756
if (hstate->nplanes > ATMEL_HLCDC_LAYER_MAX_PLANES)
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
759
for (i = 0; i < hstate->nplanes; i++) {
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
764
hstate->bpp[i] = fb->format->cpp[i];
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
765
if (!hstate->bpp[i])
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
768
switch (hstate->base.rotation & DRM_MODE_ROTATE_MASK) {
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
770
offset = (hstate->src_y / ydiv) *
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
772
offset += ((hstate->src_x + hstate->src_w - 1) /
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
773
xdiv) * hstate->bpp[i];
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
774
hstate->xstride[i] = -(((hstate->src_h - 1) / ydiv) *
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
776
(2 * hstate->bpp[i]);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
777
hstate->pstride[i] = fb->pitches[i] - hstate->bpp[i];
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
780
offset = ((hstate->src_y + hstate->src_h - 1) /
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
782
offset += ((hstate->src_x + hstate->src_w - 1) /
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
783
xdiv) * hstate->bpp[i];
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
784
hstate->xstride[i] = ((((hstate->src_w - 1) / xdiv) - 1) *
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
785
hstate->bpp[i]) - fb->pitches[i];
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
786
hstate->pstride[i] = -2 * hstate->bpp[i];
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
789
offset = ((hstate->src_y + hstate->src_h - 1) /
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
791
offset += (hstate->src_x / xdiv) * hstate->bpp[i];
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
792
hstate->xstride[i] = ((hstate->src_h - 1) / ydiv) *
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
794
hstate->pstride[i] = -fb->pitches[i] - hstate->bpp[i];
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
798
offset = (hstate->src_y / ydiv) * fb->pitches[i];
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
799
offset += (hstate->src_x / xdiv) * hstate->bpp[i];
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
800
hstate->xstride[i] = fb->pitches[i] -
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
801
((hstate->src_w / xdiv) *
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
802
hstate->bpp[i]);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
803
hstate->pstride[i] = 0;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
807
hstate->offsets[i] = offset + fb->offsets[i];
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
813
if (drm_rotation_90_or_270(hstate->base.rotation)) {
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
814
swap(hstate->src_w, hstate->src_h);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
818
(mode->hdisplay != hstate->crtc_w ||
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
819
mode->vdisplay != hstate->crtc_h))
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
822
if ((hstate->crtc_h != hstate->src_h || hstate->crtc_w != hstate->src_w) &&
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
824
hstate->base.fb->format->has_alpha))
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
927
struct atmel_hlcdc_plane_state *hstate =
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
934
if (!hstate->base.visible) {
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
939
atmel_hlcdc_plane_update_pos_and_size(plane, hstate);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
940
dc->desc->ops->lcdc_update_general_settings(plane, hstate);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
941
atmel_hlcdc_plane_update_format(plane, hstate);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
942
atmel_hlcdc_plane_update_clut(plane, hstate);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
943
atmel_hlcdc_plane_update_buffers(plane, hstate);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
944
atmel_hlcdc_plane_update_disc_area(plane, hstate);
drivers/scsi/lpfc/lpfc_nvmet.c
1316
u32 hstate;
drivers/scsi/lpfc/lpfc_nvmet.c
1325
hstate = atomic_read(&lpfc_nvmet->state);
drivers/scsi/lpfc/lpfc_nvmet.c
1326
if (hstate == LPFC_NVMET_INV_HOST_ACTIVE)
fs/hugetlbfs/inode.c
1092
unsigned long hpage_size = huge_page_size(sbinfo->hstate);
fs/hugetlbfs/inode.c
1093
unsigned hpage_shift = huge_page_shift(sbinfo->hstate);
fs/hugetlbfs/inode.c
111
struct hstate *h = hstate_file(file);
fs/hugetlbfs/inode.c
1128
struct hstate *h = hstate_inode(d_inode(dentry));
fs/hugetlbfs/inode.c
1283
hugetlbfs_size_to_hpages(struct hstate *h, unsigned long long size_opt,
fs/hugetlbfs/inode.c
1306
struct hstate *h;
fs/hugetlbfs/inode.c
1352
ctx->hstate = h;
fs/hugetlbfs/inode.c
1385
ctx->max_hpages = hugetlbfs_size_to_hpages(ctx->hstate,
fs/hugetlbfs/inode.c
1388
ctx->min_hpages = hugetlbfs_size_to_hpages(ctx->hstate,
fs/hugetlbfs/inode.c
1415
sbinfo->hstate = ctx->hstate;
fs/hugetlbfs/inode.c
1429
sbinfo->spool = hugepage_new_subpool(ctx->hstate,
fs/hugetlbfs/inode.c
1436
sb->s_blocksize = huge_page_size(ctx->hstate);
fs/hugetlbfs/inode.c
1437
sb->s_blocksize_bits = huge_page_shift(ctx->hstate);
fs/hugetlbfs/inode.c
1490
ctx->hstate = &default_hstate;
fs/hugetlbfs/inode.c
1518
struct hstate *h = hstate_sizelog(page_size_log);
fs/hugetlbfs/inode.c
1584
static struct vfsmount *__init mount_one_hugetlbfs(struct hstate *h)
fs/hugetlbfs/inode.c
1594
ctx->hstate = h;
fs/hugetlbfs/inode.c
1607
struct hstate *h;
fs/hugetlbfs/inode.c
198
struct hstate *h = hstate_file(file);
fs/hugetlbfs/inode.c
242
struct hstate *h = hstate_file(file);
fs/hugetlbfs/inode.c
399
static void hugetlb_unmap_file_folio(struct hstate *h,
fs/hugetlbfs/inode.c
518
static bool remove_inode_single_folio(struct hstate *h, struct inode *inode,
fs/hugetlbfs/inode.c
53
struct hstate *hstate;
fs/hugetlbfs/inode.c
580
struct hstate *h = hstate_inode(inode);
fs/hugetlbfs/inode.c
642
struct hstate *h = hstate_inode(inode);
fs/hugetlbfs/inode.c
656
static void hugetlbfs_zero_partial_page(struct hstate *h,
fs/hugetlbfs/inode.c
683
struct hstate *h = hstate_inode(inode);
fs/hugetlbfs/inode.c
738
struct hstate *h = hstate_inode(inode);
fs/hugetlbfs/inode.c
867
struct hstate *h = hstate_inode(inode);
include/linux/hugetlb.h
1096
static inline struct folio *filemap_lock_hugetlb_folio(struct hstate *h,
include/linux/hugetlb.h
1126
alloc_hugetlb_folio_reserve(struct hstate *h, int preferred_nid,
include/linux/hugetlb.h
1133
alloc_hugetlb_folio_nodemask(struct hstate *h, int preferred_nid,
include/linux/hugetlb.h
1140
static inline int __alloc_bootmem_huge_page(struct hstate *h)
include/linux/hugetlb.h
1145
static inline struct hstate *hstate_file(struct file *f)
include/linux/hugetlb.h
1150
static inline struct hstate *hstate_sizelog(int page_size_log)
include/linux/hugetlb.h
1155
static inline struct hstate *hstate_vma(struct vm_area_struct *vma)
include/linux/hugetlb.h
1160
static inline struct hstate *folio_hstate(struct folio *folio)
include/linux/hugetlb.h
1165
static inline struct hstate *size_to_hstate(unsigned long size)
include/linux/hugetlb.h
117
struct hugepage_subpool *hugepage_new_subpool(struct hstate *h, long max_hpages,
include/linux/hugetlb.h
1170
static inline unsigned long huge_page_size(struct hstate *h)
include/linux/hugetlb.h
1175
static inline unsigned long huge_page_mask(struct hstate *h)
include/linux/hugetlb.h
1190
static inline unsigned int huge_page_order(struct hstate *h)
include/linux/hugetlb.h
1195
static inline unsigned int huge_page_shift(struct hstate *h)
include/linux/hugetlb.h
1200
static inline bool hstate_is_gigantic(struct hstate *h)
include/linux/hugetlb.h
1205
static inline unsigned int pages_per_huge_page(struct hstate *h)
include/linux/hugetlb.h
1215
static inline int hstate_index(struct hstate *h)
include/linux/hugetlb.h
1231
static inline bool hugepage_migration_supported(struct hstate *h)
include/linux/hugetlb.h
1236
static inline bool hugepage_movable_supported(struct hstate *h)
include/linux/hugetlb.h
1241
static inline gfp_t htlb_alloc_mask(struct hstate *h)
include/linux/hugetlb.h
1246
static inline gfp_t htlb_modify_alloc_mask(struct hstate *h, gfp_t gfp_mask)
include/linux/hugetlb.h
1256
static inline spinlock_t *huge_pte_lockptr(struct hstate *h,
include/linux/hugetlb.h
1298
struct hstate *h, struct vm_area_struct *vma, unsigned long address)
include/linux/hugetlb.h
1308
static inline spinlock_t *huge_pte_lock(struct hstate *h,
include/linux/hugetlb.h
166
bool hugetlbfs_pagecache_present(struct hstate *h,
include/linux/hugetlb.h
240
unsigned long hugetlb_mask_last_page(struct hstate *h);
include/linux/hugetlb.h
43
struct hstate *hstate;
include/linux/hugetlb.h
507
struct hstate *hstate;
include/linux/hugetlb.h
538
static inline struct hstate *hstate_inode(struct inode *i)
include/linux/hugetlb.h
540
return HUGETLBFS_SB(i->i_sb)->hstate;
include/linux/hugetlb.h
552
static inline struct hstate *hstate_inode(struct inode *i)
include/linux/hugetlb.h
693
struct hstate *hstate;
include/linux/hugetlb.h
709
struct folio *alloc_hugetlb_folio_nodemask(struct hstate *h, int preferred_nid,
include/linux/hugetlb.h
712
struct folio *alloc_hugetlb_folio_reserve(struct hstate *h, int preferred_nid,
include/linux/hugetlb.h
717
void restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma,
include/linux/hugetlb.h
721
int __init __alloc_bootmem_huge_page(struct hstate *h, int nid);
include/linux/hugetlb.h
722
int __init alloc_bootmem_huge_page(struct hstate *h, int nid);
include/linux/hugetlb.h
727
struct hstate *size_to_hstate(unsigned long size);
include/linux/hugetlb.h
733
extern struct hstate hstates[HUGE_MAX_HSTATE];
include/linux/hugetlb.h
754
static inline struct hstate *hstate_file(struct file *f)
include/linux/hugetlb.h
759
static inline struct hstate *hstate_sizelog(int page_size_log)
include/linux/hugetlb.h
770
static inline struct hstate *hstate_vma(struct vm_area_struct *vma)
include/linux/hugetlb.h
775
static inline unsigned long huge_page_size(const struct hstate *h)
include/linux/hugetlb.h
784
static inline unsigned long huge_page_mask(struct hstate *h)
include/linux/hugetlb.h
789
static inline unsigned int huge_page_order(struct hstate *h)
include/linux/hugetlb.h
794
static inline unsigned huge_page_shift(struct hstate *h)
include/linux/hugetlb.h
804
static inline bool hstate_is_gigantic(struct hstate *h)
include/linux/hugetlb.h
809
static inline unsigned int pages_per_huge_page(const struct hstate *h)
include/linux/hugetlb.h
814
static inline unsigned int blocks_per_huge_page(struct hstate *h)
include/linux/hugetlb.h
819
static inline struct folio *filemap_lock_hugetlb_folio(struct hstate *h,
include/linux/hugetlb.h
860
static inline struct hstate *folio_hstate(struct folio *folio)
include/linux/hugetlb.h
871
static inline int hstate_index(struct hstate *h)
include/linux/hugetlb.h
890
static inline bool arch_hugetlb_migration_supported(struct hstate *h)
include/linux/hugetlb.h
901
static inline bool arch_hugetlb_migration_supported(struct hstate *h)
include/linux/hugetlb.h
907
static inline bool hugepage_migration_supported(struct hstate *h)
include/linux/hugetlb.h
927
static inline bool hugepage_movable_supported(struct hstate *h)
include/linux/hugetlb.h
938
static inline gfp_t htlb_alloc_mask(struct hstate *h)
include/linux/hugetlb.h
947
static inline gfp_t htlb_modify_alloc_mask(struct hstate *h, gfp_t gfp_mask)
include/linux/hugetlb.h
983
static inline spinlock_t *huge_pte_lockptr(struct hstate *h,
ipc/shm.c
742
struct hstate *hs;
ipc/shm.c
943
struct hstate *h = hstate_file(shp->shm_file);
kernel/events/uprobes.c
1975
enum hprobe_state hstate;
kernel/events/uprobes.c
1977
(void)hprobe_consume(&ri->hprobe, &hstate);
kernel/events/uprobes.c
1978
hprobe_finalize(&ri->hprobe, hstate);
kernel/events/uprobes.c
2640
enum hprobe_state hstate;
kernel/events/uprobes.c
2675
uprobe = hprobe_consume(&ri->hprobe, &hstate);
kernel/events/uprobes.c
2678
hprobe_finalize(&ri->hprobe, hstate);
kernel/events/uprobes.c
735
static inline struct uprobe *hprobe_consume(struct hprobe *hprobe, enum hprobe_state *hstate)
kernel/events/uprobes.c
737
*hstate = xchg(&hprobe->state, HPROBE_CONSUMED);
kernel/events/uprobes.c
738
switch (*hstate) {
kernel/events/uprobes.c
746
WARN(1, "hprobe invalid state %d", *hstate);
kernel/events/uprobes.c
756
static void hprobe_finalize(struct hprobe *hprobe, enum hprobe_state hstate)
kernel/events/uprobes.c
758
switch (hstate) {
kernel/events/uprobes.c
769
WARN(1, "hprobe invalid state %d", hstate);
kernel/events/uprobes.c
792
enum hprobe_state hstate;
kernel/events/uprobes.c
804
hstate = READ_ONCE(hprobe->state);
kernel/events/uprobes.c
805
switch (hstate) {
kernel/events/uprobes.c
830
if (try_cmpxchg(&hprobe->state, &hstate, uprobe ? HPROBE_STABLE : HPROBE_GONE)) {
kernel/events/uprobes.c
853
WARN(1, "unknown hprobe state %d", hstate);
mm/damon/vaddr.c
365
struct hstate *h = hstate_vma(walk->vma);
mm/damon/vaddr.c
485
struct hstate *h = hstate_vma(walk->vma);
mm/gup.c
3445
struct hstate *h;
mm/gup.c
625
struct hstate *h = hstate_vma(vma);
mm/hugetlb.c
1013
static pgoff_t vma_hugecache_offset(struct hstate *h,
mm/hugetlb.c
1090
struct hstate *h)
mm/hugetlb.c
117
static int hugetlb_acct_memory(struct hstate *h, long delta);
mm/hugetlb.c
1287
static void enqueue_hugetlb_folio(struct hstate *h, struct folio *folio)
mm/hugetlb.c
1300
static struct folio *dequeue_hugetlb_folio_node_exact(struct hstate *h,
mm/hugetlb.c
1328
static struct folio *dequeue_hugetlb_folio_nodemask(struct hstate *h, gfp_t gfp_mask,
mm/hugetlb.c
1368
static unsigned long available_huge_pages(struct hstate *h)
mm/hugetlb.c
1373
static struct folio *dequeue_hugetlb_folio_vma(struct hstate *h,
mm/hugetlb.c
1445
void remove_hugetlb_folio(struct hstate *h, struct folio *folio,
mm/hugetlb.c
146
hugetlb_acct_memory(spool->hstate,
mm/hugetlb.c
1481
void add_hugetlb_folio(struct hstate *h, struct folio *folio,
mm/hugetlb.c
1511
static void __update_and_free_hugetlb_folio(struct hstate *h,
mm/hugetlb.c
152
struct hugepage_subpool *hugepage_new_subpool(struct hstate *h, long max_hpages,
mm/hugetlb.c
1588
struct hstate *h;
mm/hugetlb.c
1609
static inline void flush_free_hpage_work(struct hstate *h)
mm/hugetlb.c
1615
static void update_and_free_hugetlb_folio(struct hstate *h, struct folio *folio,
mm/hugetlb.c
1634
static void bulk_vmemmap_restore_error(struct hstate *h,
mm/hugetlb.c
164
spool->hstate = h;
mm/hugetlb.c
1685
static void update_and_free_pages_bulk(struct hstate *h,
mm/hugetlb.c
1726
struct hstate *size_to_hstate(unsigned long size)
mm/hugetlb.c
1728
struct hstate *h;
mm/hugetlb.c
1743
struct hstate *h = folio_hstate(folio);
mm/hugetlb.c
1808
static void account_new_hugetlb_folio(struct hstate *h, struct folio *folio)
mm/hugetlb.c
1889
static struct folio *only_alloc_fresh_hugetlb_folio(struct hstate *h,
mm/hugetlb.c
1916
static struct folio *alloc_fresh_hugetlb_folio(struct hstate *h,
mm/hugetlb.c
1927
void prep_and_add_allocated_folios(struct hstate *h,
mm/hugetlb.c
1949
static struct folio *alloc_pool_huge_folio(struct hstate *h,
mm/hugetlb.c
1976
static struct folio *remove_pool_hugetlb_folio(struct hstate *h,
mm/hugetlb.c
2030
struct hstate *h = folio_hstate(folio);
mm/hugetlb.c
2107
struct hstate *h;
mm/hugetlb.c
2129
static struct folio *alloc_surplus_hugetlb_folio(struct hstate *h,
mm/hugetlb.c
2177
static struct folio *alloc_migrate_hugetlb_folio(struct hstate *h, gfp_t gfp_mask,
mm/hugetlb.c
2208
struct folio *alloc_buddy_hugetlb_folio_with_mpol(struct hstate *h,
mm/hugetlb.c
2233
struct folio *alloc_hugetlb_folio_reserve(struct hstate *h, int preferred_nid,
mm/hugetlb.c
2254
struct folio *alloc_hugetlb_folio_nodemask(struct hstate *h, int preferred_nid,
mm/hugetlb.c
2298
static int gather_surplus_pages(struct hstate *h, long delta)
mm/hugetlb.c
2403
static void return_unused_surplus_pages(struct hstate *h,
mm/hugetlb.c
2483
static long __vma_reservation_common(struct hstate *h,
mm/hugetlb.c
2563
static long vma_needs_reservation(struct hstate *h,
mm/hugetlb.c
2569
static long vma_commit_reservation(struct hstate *h,
mm/hugetlb.c
2575
static void vma_end_reservation(struct hstate *h,
mm/hugetlb.c
2581
static long vma_add_reservation(struct hstate *h,
mm/hugetlb.c
2587
static long vma_del_reservation(struct hstate *h,
mm/hugetlb.c
2613
void restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma,
mm/hugetlb.c
2698
struct hstate *h;
mm/hugetlb.c
2815
struct hstate *h;
mm/hugetlb.c
2912
struct hstate *h = hstate_vma(vma);
mm/hugetlb.c
3080
static __init void *alloc_bootmem(struct hstate *h, int nid, bool node_exact)
mm/hugetlb.c
3124
m->hstate = h;
mm/hugetlb.c
3130
int alloc_bootmem_huge_page(struct hstate *h, int nid)
mm/hugetlb.c
3132
int __alloc_bootmem_huge_page(struct hstate *h, int nid)
mm/hugetlb.c
3192
struct hstate *h,
mm/hugetlb.c
3231
struct hstate *h)
mm/hugetlb.c
3246
static void __init prep_and_add_bootmem_folios(struct hstate *h,
mm/hugetlb.c
3297
pages_per_huge_page(m->hstate));
mm/hugetlb.c
3300
hstate_boot_nrinvalid[hstate_index(m->hstate)]++;
mm/hugetlb.c
3314
struct hstate *h)
mm/hugetlb.c
3335
struct hstate *h = NULL, *prev_h = NULL;
mm/hugetlb.c
3341
h = m->hstate;
mm/hugetlb.c
3422
static void __init hugetlb_hstate_alloc_pages_onenode(struct hstate *h, int nid)
mm/hugetlb.c
3465
static bool __init hugetlb_hstate_alloc_pages_specific_nodes(struct hstate *h)
mm/hugetlb.c
3480
static void __init hugetlb_hstate_alloc_pages_errcheck(unsigned long allocated, struct hstate *h)
mm/hugetlb.c
3494
struct hstate *h = (struct hstate *)arg;
mm/hugetlb.c
3523
static unsigned long __init hugetlb_gigantic_pages_alloc_boot(struct hstate *h)
mm/hugetlb.c
3536
static unsigned long __init hugetlb_pages_alloc_boot(struct hstate *h)
mm/hugetlb.c
3615
static void __init hugetlb_hstate_alloc_pages(struct hstate *h)
mm/hugetlb.c
3647
struct hstate *h, *h2;
mm/hugetlb.c
3687
struct hstate *h;
mm/hugetlb.c
3708
static void try_to_free_low(struct hstate *h, unsigned long count,
mm/hugetlb.c
3740
static inline void try_to_free_low(struct hstate *h, unsigned long count,
mm/hugetlb.c
3751
static int adjust_pool_surplus(struct hstate *h, nodemask_t *nodes_allowed,
mm/hugetlb.c
3780
static int set_max_huge_pages(struct hstate *h, unsigned long count, int nid,
mm/hugetlb.c
3969
static long demote_free_hugetlb_folios(struct hstate *src, struct hstate *dst,
mm/hugetlb.c
4028
long demote_pool_huge_page(struct hstate *src, nodemask_t *nodes_allowed,
mm/hugetlb.c
4033
struct hstate *dst;
mm/hugetlb.c
4098
struct hstate *h, int nid,
mm/hugetlb.c
4210
struct hstate *h;
mm/hugetlb.c
4387
struct hstate *h;
mm/hugetlb.c
4507
struct hstate *h;
mm/hugetlb.c
4547
static unsigned int allowed_mems_nr(struct hstate *h)
mm/hugetlb.c
4566
struct hstate *h;
mm/hugetlb.c
4596
struct hstate *h = &default_hstate;
mm/hugetlb.c
4612
struct hstate *h;
mm/hugetlb.c
4635
struct hstate *h;
mm/hugetlb.c
4643
static int hugetlb_acct_memory(struct hstate *h, long delta)
mm/hugetlb.c
4735
struct hstate *h = hstate_vma(vma);
mm/hugetlb.c
489
struct hstate *h,
mm/hugetlb.c
4893
struct hstate *h = hstate_vma(src_vma);
mm/hugetlb.c
5073
struct hstate *h = hstate_vma(vma);
mm/hugetlb.c
5112
struct hstate *h = hstate_vma(vma);
mm/hugetlb.c
5182
struct hstate *h = hstate_vma(vma);
mm/hugetlb.c
5396
struct hstate *h = hstate_vma(vma);
mm/hugetlb.c
5456
struct hstate *h = hstate_vma(vma);
mm/hugetlb.c
5645
bool hugetlbfs_pagecache_present(struct hstate *h,
mm/hugetlb.c
5663
struct hstate *h = hstate_inode(inode);
mm/hugetlb.c
57
struct hstate hstates[HUGE_MAX_HSTATE];
mm/hugetlb.c
5709
static bool hugetlb_pte_stable(struct hstate *h, struct mm_struct *mm, unsigned long addr,
mm/hugetlb.c
571
long to, struct hstate *h, struct hugetlb_cgroup *cg,
mm/hugetlb.c
5729
struct hstate *h = hstate_vma(vma);
mm/hugetlb.c
5978
struct hstate *h = hstate_vma(vma);
mm/hugetlb.c
598
struct hstate *h, long *regions_needed)
mm/hugetlb.c
6167
static struct folio *alloc_hugetlb_folio_vma(struct hstate *h,
mm/hugetlb.c
6204
struct hstate *h = hstate_vma(dst_vma);
mm/hugetlb.c
6422
struct hstate *h = hstate_vma(vma);
mm/hugetlb.c
6576
struct hstate *h = hstate_inode(inode);
mm/hugetlb.c
6754
struct hstate *h = hstate_inode(inode);
mm/hugetlb.c
7093
unsigned long hugetlb_mask_last_page(struct hstate *h)
mm/hugetlb.c
7108
__weak unsigned long hugetlb_mask_last_page(struct hstate *h)
mm/hugetlb.c
7206
struct hstate *h = folio_hstate(old_folio);
mm/hugetlb.c
7265
struct hstate *h = hstate_vma(vma);
mm/hugetlb.c
730
long in_regions_needed, struct hstate *h,
mm/hugetlb.c
82
static struct hstate * __initdata parsed_hstate;
mm/hugetlb.c
966
struct hstate *h = hstate_inode(inode);
mm/hugetlb_cgroup.c
225
struct hstate *h;
mm/hugetlb_cgroup.c
805
hugetlb_cgroup_cfttypes_init(struct hstate *h, struct cftype *cft,
mm/hugetlb_cgroup.c
81
struct hstate *h;
mm/hugetlb_cgroup.c
832
static void __init __hugetlb_cgroup_file_dfl_init(struct hstate *h)
mm/hugetlb_cgroup.c
840
static void __init __hugetlb_cgroup_file_legacy_init(struct hstate *h)
mm/hugetlb_cgroup.c
848
static void __init __hugetlb_cgroup_file_init(struct hstate *h)
mm/hugetlb_cgroup.c
876
struct hstate *h;
mm/hugetlb_cgroup.c
892
struct hstate *h = folio_hstate(old_folio);
mm/hugetlb_cma.c
274
bool __init hugetlb_early_cma(struct hstate *h)
mm/hugetlb_cma.c
60
hugetlb_cma_alloc_bootmem(struct hstate *h, int *nid, bool node_exact)
mm/hugetlb_cma.h
14
bool hugetlb_early_cma(struct hstate *h);
mm/hugetlb_cma.h
27
struct huge_bootmem_page *hugetlb_cma_alloc_bootmem(struct hstate *h, int *nid,
mm/hugetlb_cma.h
47
static inline bool hugetlb_early_cma(struct hstate *h)
mm/hugetlb_cma.h
9
struct huge_bootmem_page *hugetlb_cma_alloc_bootmem(struct hstate *h, int *nid,
mm/hugetlb_internal.h
100
extern void prep_and_add_allocated_folios(struct hstate *h,
mm/hugetlb_internal.h
102
extern long demote_pool_huge_page(struct hstate *src,
mm/hugetlb_internal.h
106
struct hstate *h, int nid,
mm/hugetlb_internal.h
19
static inline bool hstate_is_gigantic_no_runtime(struct hstate *h)
mm/hugetlb_internal.h
71
static inline int hstate_next_node_to_free(struct hstate *h, nodemask_t *nodes_allowed)
mm/hugetlb_internal.h
95
extern void remove_hugetlb_folio(struct hstate *h, struct folio *folio,
mm/hugetlb_internal.h
97
extern void add_hugetlb_folio(struct hstate *h, struct folio *folio,
mm/hugetlb_sysctl.c
34
struct hstate *h = &default_hstate;
mm/hugetlb_sysctl.c
73
struct hstate *h = &default_hstate;
mm/hugetlb_sysfs.c
113
struct hstate *h = kobj_to_hstate(kobj, NULL);
mm/hugetlb_sysfs.c
122
struct hstate *h = kobj_to_hstate(kobj, NULL);
mm/hugetlb_sysfs.c
142
struct hstate *h;
mm/hugetlb_sysfs.c
159
struct hstate *h = kobj_to_hstate(kobj, NULL);
mm/hugetlb_sysfs.c
167
struct hstate *h;
mm/hugetlb_sysfs.c
187
struct hstate *h;
mm/hugetlb_sysfs.c
243
struct hstate *h = kobj_to_hstate(kobj, NULL);
mm/hugetlb_sysfs.c
253
struct hstate *h, *demote_hstate;
mm/hugetlb_sysfs.c
26
static struct hstate *kobj_to_node_hstate(struct kobject *kobj, int *nidp);
mm/hugetlb_sysfs.c
28
static struct hstate *kobj_to_hstate(struct kobject *kobj, int *nidp)
mm/hugetlb_sysfs.c
306
static int hugetlb_sysfs_add_hstate(struct hstate *h, struct kobject *parent,
mm/hugetlb_sysfs.c
373
static struct hstate *kobj_to_node_hstate(struct kobject *kobj, int *nidp)
mm/hugetlb_sysfs.c
398
struct hstate *h;
mm/hugetlb_sysfs.c
428
struct hstate *h;
mm/hugetlb_sysfs.c
45
struct hstate *h;
mm/hugetlb_sysfs.c
470
static struct hstate *kobj_to_node_hstate(struct kobject *kobj, int *nidp)
mm/hugetlb_sysfs.c
484
struct hstate *h;
mm/hugetlb_sysfs.c
62
struct hstate *h;
mm/hugetlb_vmemmap.c
454
static int __hugetlb_vmemmap_restore_folio(const struct hstate *h,
mm/hugetlb_vmemmap.c
500
int hugetlb_vmemmap_restore_folio(const struct hstate *h, struct folio *folio)
mm/hugetlb_vmemmap.c
518
long hugetlb_vmemmap_restore_folios(const struct hstate *h,
mm/hugetlb_vmemmap.c
550
static bool vmemmap_should_optimize_folio(const struct hstate *h, struct folio *folio)
mm/hugetlb_vmemmap.c
564
static int __hugetlb_vmemmap_optimize_folio(const struct hstate *h,
mm/hugetlb_vmemmap.c
626
void hugetlb_vmemmap_optimize_folio(const struct hstate *h, struct folio *folio)
mm/hugetlb_vmemmap.c
634
static int hugetlb_vmemmap_split_folio(const struct hstate *h, struct folio *folio)
mm/hugetlb_vmemmap.c
653
static void __hugetlb_vmemmap_optimize_folios(struct hstate *h,
mm/hugetlb_vmemmap.c
737
void hugetlb_vmemmap_optimize_folios(struct hstate *h, struct list_head *folio_list)
mm/hugetlb_vmemmap.c
742
void hugetlb_vmemmap_optimize_bootmem_folios(struct hstate *h, struct list_head *folio_list)
mm/hugetlb_vmemmap.c
758
if (!hugetlb_vmemmap_optimizable(m->hstate))
mm/hugetlb_vmemmap.c
761
psize = huge_page_size(m->hstate);
mm/hugetlb_vmemmap.c
806
nr_pages = pages_per_huge_page(m->hstate);
mm/hugetlb_vmemmap.c
839
struct hstate *h;
mm/hugetlb_vmemmap.c
850
h = m->hstate;
mm/hugetlb_vmemmap.c
891
const struct hstate *h;
mm/hugetlb_vmemmap.h
23
int hugetlb_vmemmap_restore_folio(const struct hstate *h, struct folio *folio);
mm/hugetlb_vmemmap.h
24
long hugetlb_vmemmap_restore_folios(const struct hstate *h,
mm/hugetlb_vmemmap.h
27
void hugetlb_vmemmap_optimize_folio(const struct hstate *h, struct folio *folio);
mm/hugetlb_vmemmap.h
28
void hugetlb_vmemmap_optimize_folios(struct hstate *h, struct list_head *folio_list);
mm/hugetlb_vmemmap.h
29
void hugetlb_vmemmap_optimize_bootmem_folios(struct hstate *h, struct list_head *folio_list);
mm/hugetlb_vmemmap.h
36
static inline unsigned int hugetlb_vmemmap_size(const struct hstate *h)
mm/hugetlb_vmemmap.h
45
static inline unsigned int hugetlb_vmemmap_optimizable_size(const struct hstate *h)
mm/hugetlb_vmemmap.h
54
static inline int hugetlb_vmemmap_restore_folio(const struct hstate *h, struct folio *folio)
mm/hugetlb_vmemmap.h
59
static inline long hugetlb_vmemmap_restore_folios(const struct hstate *h,
mm/hugetlb_vmemmap.h
67
static inline void hugetlb_vmemmap_optimize_folio(const struct hstate *h, struct folio *folio)
mm/hugetlb_vmemmap.h
71
static inline void hugetlb_vmemmap_optimize_folios(struct hstate *h, struct list_head *folio_list)
mm/hugetlb_vmemmap.h
75
static inline void hugetlb_vmemmap_optimize_bootmem_folios(struct hstate *h,
mm/hugetlb_vmemmap.h
88
static inline unsigned int hugetlb_vmemmap_optimizable_size(const struct hstate *h)
mm/hugetlb_vmemmap.h
94
static inline bool hugetlb_vmemmap_optimizable(const struct hstate *h)
mm/memfd.c
81
struct hstate *h = hstate_file(memfd);
mm/memory-failure.c
780
struct hstate *h = hstate_vma(walk->vma);
mm/mempolicy.c
1449
struct hstate *h;
mm/migrate.c
2184
struct hstate *h = folio_hstate(src);
mm/migrate.c
416
struct hstate *h = hstate_vma(vma);
mm/mmap.c
586
struct hstate *hs;
mm/mremap.c
1488
struct hstate *h __maybe_unused = hstate_vma(vrm->vma);
mm/page_isolation.c
48
struct hstate *h;
mm/page_vma_mapped.c
196
struct hstate *hstate = hstate_vma(vma);
mm/page_vma_mapped.c
197
unsigned long size = huge_page_size(hstate);
mm/page_vma_mapped.c
210
pvmw->ptl = huge_pte_lock(hstate, mm, pvmw->pte);
mm/page_vma_mapped.c
211
if (!check_pte(pvmw, pages_per_huge_page(hstate)))
mm/pagewalk.c
331
static unsigned long hugetlb_entry_end(struct hstate *h, unsigned long addr,
mm/pagewalk.c
343
struct hstate *h = hstate_vma(vma);