Symbol: zone_idx
drivers/block/virtio_blk.c
665
unsigned int zone_idx = 0;
drivers/block/virtio_blk.c
682
while (zone_idx < nr_zones && sector < get_capacity(vblk->disk)) {
drivers/block/virtio_blk.c
695
for (i = 0; i < nz && zone_idx < nr_zones; i++) {
drivers/block/virtio_blk.c
697
zone_idx, args);
drivers/block/virtio_blk.c
704
zone_idx++;
drivers/block/virtio_blk.c
708
if (zone_idx > 0)
drivers/block/virtio_blk.c
709
ret = zone_idx;
drivers/md/dm-zone.c
123
ret = args->cb(zone, args->zone_idx, args->data);
drivers/md/dm-zone.c
128
return disk_report_zone(args->disk, zone, args->zone_idx++,
drivers/md/dm-zone.c
33
nr_zones - args->zone_idx);
drivers/md/dm-zone.c
36
} while (args->zone_idx < nr_zones &&
drivers/md/dm-zone.c
39
return args->zone_idx;
drivers/net/ethernet/sfc/tc.h
253
u8 zone_idx; /* for TABLE_FIELD_ID_DOMAIN */
drivers/nvme/host/zns.c
180
int ret, zone_idx = 0;
drivers/nvme/host/zns.c
199
while (zone_idx < nr_zones && sector < get_capacity(ns->disk)) {
drivers/nvme/host/zns.c
214
for (i = 0; i < nz && zone_idx < nr_zones; i++) {
drivers/nvme/host/zns.c
216
zone_idx, args);
drivers/nvme/host/zns.c
219
zone_idx++;
drivers/nvme/host/zns.c
225
if (zone_idx > 0)
drivers/nvme/host/zns.c
226
ret = zone_idx;
drivers/scsi/sd_zbc.c
229
int zone_idx = 0;
drivers/scsi/sd_zbc.c
244
while (zone_idx < nr_zones && lba < sdkp->capacity) {
drivers/scsi/sd_zbc.c
254
for (i = 0; i < nr && zone_idx < nr_zones; i++) {
drivers/scsi/sd_zbc.c
258
if ((zone_idx == 0 &&
drivers/scsi/sd_zbc.c
261
(zone_idx > 0 && start_lba != lba) ||
drivers/scsi/sd_zbc.c
265
zone_idx, lba, start_lba, zone_length);
drivers/scsi/sd_zbc.c
279
ret = sd_zbc_parse_report(sdkp, buf + offset, zone_idx,
drivers/scsi/sd_zbc.c
284
zone_idx++;
drivers/scsi/sd_zbc.c
288
ret = zone_idx;
fs/btrfs/zoned.c
1308
static int btrfs_load_zone_info(struct btrfs_fs_info *fs_info, int zone_idx,
fs/btrfs/zoned.c
1319
info->physical = map->stripes[zone_idx].physical;
fs/btrfs/zoned.c
1322
device = map->stripes[zone_idx].dev;
fs/btrfs/zoned.c
1332
__set_bit(zone_idx, active);
fs/btrfs/zoned.c
1407
__set_bit(zone_idx, active);
include/linux/device-mapper.h
544
unsigned int zone_idx;
include/linux/memcontrol.h
1323
enum lru_list lru, int zone_idx)
include/linux/memcontrol.h
903
enum lru_list lru, int zone_idx)
include/linux/memcontrol.h
908
return READ_ONCE(mz->lru_zone_size[zone_idx][lru]);
include/linux/mmzone.h
1312
int zone_idx; /* zone_idx(zoneref->zone) */
include/linux/mmzone.h
1595
return zone_idx(zone) == ZONE_DEVICE;
include/linux/mmzone.h
1661
return is_highmem_idx(zone_idx(zone));
include/linux/mmzone.h
1728
return zoneref->zone_idx;
include/trace/events/compaction.h
209
__entry->idx = zone_idx(zone);
include/trace/events/compaction.h
256
__entry->idx = zone_idx(zone);
include/trace/events/oom.h
48
__field( int, zone_idx)
include/trace/events/oom.h
59
__entry->zone_idx = zonelist_zone_idx(zoneref);
include/trace/events/oom.h
69
__entry->node, __print_symbolic(__entry->zone_idx, ZONE_TYPE),
mm/khugepaged.c
2649
if (zone_idx(zone) > gfp_zone(GFP_USER))
mm/memory_hotplug.c
1073
const bool movable = zone_idx(zone) == ZONE_MOVABLE;
mm/memory_hotplug.c
1216
if (!node_state(nid, N_NORMAL_MEMORY) && zone_idx(zone) <= ZONE_NORMAL)
mm/memory_hotplug.c
2072
if (zone_idx(zone) <= ZONE_NORMAL) {
mm/memory_hotplug.c
782
memmap_init_range(nr_pages, nid, zone_idx(zone), start_pfn, 0,
mm/memory_hotplug.c
797
if (zone_idx(zone) == ZONE_MOVABLE) {
mm/mm_init.c
1005
unsigned long zone_idx, int nid,
mm/mm_init.c
1009
__init_single_page(page, pfn, zone_idx, nid);
mm/mm_init.c
1084
unsigned long zone_idx, int nid,
mm/mm_init.c
1101
__init_zone_device_page(page, pfn, zone_idx, nid, pgmap);
mm/mm_init.c
1117
unsigned long zone_idx = zone_idx(zone);
mm/mm_init.c
1121
if (WARN_ON_ONCE(!pgmap || zone_idx != ZONE_DEVICE))
mm/mm_init.c
1137
__init_zone_device_page(page, pfn, zone_idx, nid, pgmap);
mm/mm_init.c
1142
memmap_init_compound(page, pfn, zone_idx, nid, pgmap,
mm/mm_init.c
1451
int zone_idx = zone_idx(zone) + 1;
mm/mm_init.c
1453
if (zone_idx > pgdat->nr_zones)
mm/mm_init.c
1454
pgdat->nr_zones = zone_idx;
mm/mm_init.c
1461
(unsigned long)zone_idx(zone),
mm/mm_init.c
2037
int zid = zone_idx(zone);
mm/mm_init.c
948
int nid = zone_to_nid(zone), zone_id = zone_idx(zone);
mm/page_alloc.c
3432
wakeup_kswapd(zone, 0, 0, zone_idx(zone));
mm/page_alloc.c
3775
if (zone_idx(zone) != ZONE_NORMAL)
mm/page_alloc.c
5197
__count_zid_vm_events(PGALLOC, zone_idx(zone), nr_account);
mm/page_alloc.c
5543
zoneref->zone_idx = zone_idx(zone);
mm/page_alloc.c
5686
zonerefs->zone_idx = 0;
mm/page_alloc.c
5701
zonerefs->zone_idx = 0;
mm/page_alloc.c
5770
zonerefs->zone_idx = 0;
mm/page_alloc.c
6442
if (!is_highmem(zone) && zone_idx(zone) != ZONE_MOVABLE)
mm/page_alloc.c
6452
if (is_highmem(zone) || zone_idx(zone) == ZONE_MOVABLE) {
mm/page_isolation.c
35
if (zone_idx(zone) == ZONE_MOVABLE)
mm/show_mem.c
164
int zone_idx;
mm/show_mem.c
165
for (zone_idx = 0; zone_idx <= max_zone_idx; zone_idx++)
mm/show_mem.c
166
if (zone_managed_pages(pgdat->node_zones + zone_idx))
mm/show_mem.c
188
if (zone_idx(zone) > max_zone_idx)
mm/show_mem.c
288
if (zone_idx(zone) > max_zone_idx)
mm/show_mem.c
359
if (zone_idx(zone) > max_zone_idx)
mm/vmscan.c
413
int zone_idx)
mm/vmscan.c
419
for_each_managed_zone_pgdat(zone, lruvec_pgdat(lruvec), zid, zone_idx) {
mm/vmscan.c
6526
if (zone_idx(zone) > ZONE_NORMAL)