/kernel/linux/linux-5.10/drivers/nvme/host/ |
D | zns.c | 174 int ret, zone_idx = 0; in nvme_ns_report_zones() local 190 while (zone_idx < nr_zones && sector < get_capacity(ns->disk)) { in nvme_ns_report_zones() 205 for (i = 0; i < nz && zone_idx < nr_zones; i++) { in nvme_ns_report_zones() 207 zone_idx, cb, data); in nvme_ns_report_zones() 210 zone_idx++; in nvme_ns_report_zones() 216 if (zone_idx > 0) in nvme_ns_report_zones() 217 ret = zone_idx; in nvme_ns_report_zones()
|
/kernel/linux/linux-5.10/include/trace/events/ |
D | oom.h | 46 __field( int, zone_idx) 57 __entry->zone_idx = zoneref->zone_idx; 67 __entry->node, __print_symbolic(__entry->zone_idx, ZONE_TYPE),
|
D | compaction.h | 214 __entry->idx = zone_idx(zone); 261 __entry->idx = zone_idx(zone);
|
/kernel/linux/linux-5.10/drivers/scsi/ |
D | sd_zbc.c | 201 int zone_idx = 0; in sd_zbc_report_zones() local 216 while (zone_idx < nr_zones && sector < capacity) { in sd_zbc_report_zones() 227 for (i = 0; i < nr && zone_idx < nr_zones; i++) { in sd_zbc_report_zones() 229 ret = sd_zbc_parse_report(sdkp, buf + offset, zone_idx, in sd_zbc_report_zones() 233 zone_idx++; in sd_zbc_report_zones() 239 ret = zone_idx; in sd_zbc_report_zones()
|
/kernel/linux/linux-5.10/include/linux/ |
D | mmzone.h | 695 int zone_idx; /* zone_idx(zoneref->zone) */ member 918 extern unsigned long lruvec_lru_size(struct lruvec *lruvec, enum lru_list lru, int zone_idx); 929 #define zone_idx(zone) ((zone) - (zone)->zone_pgdat->node_zones) macro 1008 return is_highmem_idx(zone_idx(zone)); in is_highmem() 1087 return zoneref->zone_idx; in zonelist_zone_idx()
|
D | memcontrol.h | 653 enum lru_list lru, int zone_idx) in mem_cgroup_get_zone_lru_size() argument 658 return READ_ONCE(mz->lru_zone_size[zone_idx][lru]); in mem_cgroup_get_zone_lru_size() 1145 enum lru_list lru, int zone_idx) in mem_cgroup_get_zone_lru_size() argument
|
D | device-mapper.h | 458 unsigned int zone_idx; member
|
/kernel/linux/linux-5.10/mm/ |
D | memory_hotplug.c | 496 if (zone_idx(zone) == ZONE_DEVICE) in remove_pfn_range_from_zone() 641 if (zone_idx(zone) <= ZONE_NORMAL && !node_state(nid, N_NORMAL_MEMORY)) in node_states_check_changes_online() 644 if (zone_idx(zone) <= ZONE_HIGHMEM && !node_state(nid, N_HIGH_MEMORY)) in node_states_check_changes_online() 718 memmap_init_zone(nr_pages, nid, zone_idx(zone), start_pfn, 0, in move_pfn_range_to_zone() 1402 if (zone_idx(zone) <= ZONE_NORMAL && nr_pages >= present_pages) in node_states_check_changes_offline() 1415 if (zone_idx(zone) <= ZONE_HIGHMEM && nr_pages >= present_pages) in node_states_check_changes_offline()
|
D | page_alloc.c | 1785 int zid = zone_idx(zone); in deferred_init_pages() 3516 wakeup_kswapd(zone, 0, 0, zone_idx(zone)); in rmqueue() 3789 if (zone_idx(zone) != ZONE_NORMAL) in alloc_flags_nofragment() 5724 zoneref->zone_idx = zone_idx(zone); in zoneref_set_zone() 5866 zonerefs->zone_idx = 0; in build_zonelists_in_node_order() 5881 zonerefs->zone_idx = 0; in build_thisnode_zonelists() 5979 zonerefs->zone_idx = 0; in build_zonelists() 6211 unsigned long zone_idx = zone_idx(zone); in memmap_init_zone_device() local 6215 if (WARN_ON_ONCE(!pgmap || zone_idx(zone) != ZONE_DEVICE)) in memmap_init_zone_device() 6231 __init_single_page(page, pfn, zone_idx, nid); in memmap_init_zone_device() [all …]
|
D | vmscan.c | 226 unsigned long lruvec_lru_size(struct lruvec *lruvec, enum lru_list lru, int zone_idx) in lruvec_lru_size() argument 233 for (zid = 0; zid <= zone_idx && zid < MAX_NR_ZONES; zid++) { in lruvec_lru_size() 245 for (zid = 0; zid <= zone_idx && zid < MAX_NR_ZONES; zid++) { in lruvec_lru_size() 3192 if (zone_idx(zone) > ZONE_NORMAL) in throttle_direct_reclaim()
|
D | khugepaged.c | 2291 if (zone_idx(zone) > gfp_zone(GFP_USER)) in set_recommended_min_free_kbytes()
|
D | migrate.c | 1578 zidx = zone_idx(page_zone(page)); in alloc_migration_target()
|
/kernel/linux/linux-5.10/fs/f2fs/ |
D | segment.c | 5090 static bool is_conv_zone(struct f2fs_sb_info *sbi, unsigned int zone_idx, in is_conv_zone() argument 5095 return !test_bit(zone_idx, FDEV(dev_idx).blkz_seq); in is_conv_zone() 5115 unsigned int dev_idx, zone_idx, unusable_segs_in_sec; in f2fs_usable_zone_segs_in_sec() local 5118 zone_idx = get_zone_idx(sbi, GET_SEC_FROM_SEG(sbi, segno), dev_idx); in f2fs_usable_zone_segs_in_sec() 5121 if (is_conv_zone(sbi, zone_idx, dev_idx)) in f2fs_usable_zone_segs_in_sec() 5133 FDEV(dev_idx).zone_capacity_blocks[zone_idx]) >> in f2fs_usable_zone_segs_in_sec() 5150 unsigned int zone_idx, dev_idx, secno; in f2fs_usable_zone_blks_in_seg() local 5155 zone_idx = get_zone_idx(sbi, secno, dev_idx); in f2fs_usable_zone_blks_in_seg() 5161 if (is_conv_zone(sbi, zone_idx, dev_idx)) in f2fs_usable_zone_blks_in_seg() 5169 FDEV(dev_idx).zone_capacity_blocks[zone_idx]; in f2fs_usable_zone_blks_in_seg()
|
/kernel/linux/linux-5.10/drivers/md/ |
D | dm.c | 469 return args->orig_cb(zone, args->zone_idx++, args->orig_data); in dm_report_zones_cb() 505 nr_zones - args.zone_idx); in dm_blk_report_zones() 508 } while (args.zone_idx < nr_zones && in dm_blk_report_zones() 511 ret = args.zone_idx; in dm_blk_report_zones()
|
/kernel/linux/patches/linux-5.10/yangfan_patch/ |
D | others.patch | 688 - __init_single_page(page, pfn, zone_idx, nid); 689 + __init_single_page(page, pfn, zone_idx, nid, true);
|