mm/filemap.c
959
ret = mem_cgroup_charge(folio, NULL, gfp);
mm/huge_memory.c
1272
if (mem_cgroup_charge(folio, vma->vm_mm, gfp)) {
mm/khugepaged.c
1066
if (unlikely(mem_cgroup_charge(folio, mm, gfp))) {
mm/ksm.c
3131
mem_cgroup_charge(new_folio, vma->vm_mm, GFP_KERNEL)) {
mm/memfd_luo.c
424
err = mem_cgroup_charge(folio, NULL, mapping_gfp_mask(mapping));
mm/memory.c
1211
if (mem_cgroup_charge(new_folio, src_mm, GFP_KERNEL)) {
mm/memory.c
5185
if (mem_cgroup_charge(folio, vma->vm_mm, gfp)) {
mm/migrate_device.c
1023
if (mem_cgroup_charge(folio, vma->vm_mm, GFP_KERNEL))
mm/migrate_device.c
823
if (mem_cgroup_charge(folio, migrate->vma->vm_mm, gfp)) {
mm/shmem.c
1982
error = mem_cgroup_charge(folio, fault_mm, gfp);
mm/shmem.c
3273
ret = mem_cgroup_charge(folio, dst_vma->vm_mm, gfp);
mm/userfaultfd.c
303
if (mem_cgroup_charge(folio, dst_vma->vm_mm, GFP_KERNEL))
mm/userfaultfd.c
328
if (mem_cgroup_charge(folio, dst_vma->vm_mm, GFP_KERNEL))