/kernel/linux/linux-4.19/drivers/gpu/drm/ttm/ |
D | ttm_memory.c | 78 struct ttm_mem_zone *zone = in ttm_mem_zone_kobj_release() local 82 zone->name, (unsigned long long)zone->used_mem >> 10); in ttm_mem_zone_kobj_release() 83 kfree(zone); in ttm_mem_zone_kobj_release() 90 struct ttm_mem_zone *zone = in ttm_mem_zone_show() local 94 spin_lock(&zone->glob->lock); in ttm_mem_zone_show() 96 val = zone->zone_mem; in ttm_mem_zone_show() 98 val = zone->emer_mem; in ttm_mem_zone_show() 100 val = zone->max_mem; in ttm_mem_zone_show() 102 val = zone->swap_limit; in ttm_mem_zone_show() 104 val = zone->used_mem; in ttm_mem_zone_show() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/ttm/ |
D | ttm_memory.c | 81 struct ttm_mem_zone *zone = in ttm_mem_zone_kobj_release() local 85 zone->name, (unsigned long long)zone->used_mem >> 10); in ttm_mem_zone_kobj_release() 86 kfree(zone); in ttm_mem_zone_kobj_release() 93 struct ttm_mem_zone *zone = in ttm_mem_zone_show() local 97 spin_lock(&zone->glob->lock); in ttm_mem_zone_show() 99 val = zone->zone_mem; in ttm_mem_zone_show() 101 val = zone->emer_mem; in ttm_mem_zone_show() 103 val = zone->max_mem; in ttm_mem_zone_show() 105 val = zone->swap_limit; in ttm_mem_zone_show() 107 val = zone->used_mem; in ttm_mem_zone_show() [all …]
|
/kernel/linux/linux-5.10/fs/pstore/ |
D | zone.c | 164 static inline int buffer_datalen(struct pstore_zone *zone) in buffer_datalen() argument 166 return atomic_read(&zone->buffer->datalen); in buffer_datalen() 169 static inline int buffer_start(struct pstore_zone *zone) in buffer_start() argument 171 return atomic_read(&zone->buffer->start); in buffer_start() 179 static ssize_t psz_zone_read_buffer(struct pstore_zone *zone, char *buf, in psz_zone_read_buffer() argument 182 if (!buf || !zone || !zone->buffer) in psz_zone_read_buffer() 184 if (off > zone->buffer_size) in psz_zone_read_buffer() 186 len = min_t(size_t, len, zone->buffer_size - off); in psz_zone_read_buffer() 187 memcpy(buf, zone->buffer->data + off, len); in psz_zone_read_buffer() 191 static int psz_zone_read_oldbuf(struct pstore_zone *zone, char *buf, in psz_zone_read_oldbuf() argument [all …]
|
/kernel/linux/linux-4.19/mm/ |
D | compaction.c | 141 void defer_compaction(struct zone *zone, int order) in defer_compaction() argument 143 zone->compact_considered = 0; in defer_compaction() 144 zone->compact_defer_shift++; in defer_compaction() 146 if (order < zone->compact_order_failed) in defer_compaction() 147 zone->compact_order_failed = order; in defer_compaction() 149 if (zone->compact_defer_shift > COMPACT_MAX_DEFER_SHIFT) in defer_compaction() 150 zone->compact_defer_shift = COMPACT_MAX_DEFER_SHIFT; in defer_compaction() 152 trace_mm_compaction_defer_compaction(zone, order); in defer_compaction() 156 bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument 158 unsigned long defer_limit = 1UL << zone->compact_defer_shift; in compaction_deferred() [all …]
|
D | vmstat.c | 39 static void zero_zone_numa_counters(struct zone *zone) in zero_zone_numa_counters() argument 44 atomic_long_set(&zone->vm_numa_stat[item], 0); in zero_zone_numa_counters() 46 per_cpu_ptr(zone->pageset, cpu)->vm_numa_stat_diff[item] in zero_zone_numa_counters() 54 struct zone *zone; in zero_zones_numa_counters() local 56 for_each_populated_zone(zone) in zero_zones_numa_counters() 57 zero_zone_numa_counters(zone); in zero_zones_numa_counters() 171 int calculate_pressure_threshold(struct zone *zone) in calculate_pressure_threshold() argument 184 watermark_distance = low_wmark_pages(zone) - min_wmark_pages(zone); in calculate_pressure_threshold() 195 int calculate_normal_threshold(struct zone *zone) in calculate_normal_threshold() argument 230 mem = zone->managed_pages >> (27 - PAGE_SHIFT); in calculate_normal_threshold() [all …]
|
D | page_alloc.c | 483 static int page_outside_zone_boundaries(struct zone *zone, struct page *page) in page_outside_zone_boundaries() argument 491 seq = zone_span_seqbegin(zone); in page_outside_zone_boundaries() 492 start_pfn = zone->zone_start_pfn; in page_outside_zone_boundaries() 493 sp = zone->spanned_pages; in page_outside_zone_boundaries() 494 if (!zone_spans_pfn(zone, pfn)) in page_outside_zone_boundaries() 496 } while (zone_span_seqretry(zone, seq)); in page_outside_zone_boundaries() 500 pfn, zone_to_nid(zone), zone->name, in page_outside_zone_boundaries() 506 static int page_is_consistent(struct zone *zone, struct page *page) in page_is_consistent() argument 510 if (zone != page_zone(page)) in page_is_consistent() 518 static int __maybe_unused bad_range(struct zone *zone, struct page *page) in bad_range() argument [all …]
|
D | memory_hotplug.c | 313 static unsigned long find_smallest_section_pfn(int nid, struct zone *zone, in find_smallest_section_pfn() argument 324 if (zone && zone != page_zone(pfn_to_page(start_pfn))) in find_smallest_section_pfn() 334 static unsigned long find_biggest_section_pfn(int nid, struct zone *zone, in find_biggest_section_pfn() argument 349 if (zone && zone != page_zone(pfn_to_page(pfn))) in find_biggest_section_pfn() 358 static void shrink_zone_span(struct zone *zone, unsigned long start_pfn, in shrink_zone_span() argument 361 unsigned long zone_start_pfn = zone->zone_start_pfn; in shrink_zone_span() 362 unsigned long z = zone_end_pfn(zone); /* zone_end_pfn namespace clash */ in shrink_zone_span() 365 int nid = zone_to_nid(zone); in shrink_zone_span() 367 zone_span_writelock(zone); in shrink_zone_span() 375 pfn = find_smallest_section_pfn(nid, zone, end_pfn, in shrink_zone_span() [all …]
|
/kernel/linux/linux-5.10/drivers/block/ |
D | null_blk_zoned.c | 91 struct blk_zone *zone = &dev->zones[i]; in null_init_zoned_dev() local 93 zone->start = sector; in null_init_zoned_dev() 94 zone->len = dev->zone_size_sects; in null_init_zoned_dev() 95 zone->capacity = zone->len; in null_init_zoned_dev() 96 zone->wp = zone->start + zone->len; in null_init_zoned_dev() 97 zone->type = BLK_ZONE_TYPE_CONVENTIONAL; in null_init_zoned_dev() 98 zone->cond = BLK_ZONE_COND_NOT_WP; in null_init_zoned_dev() 104 struct blk_zone *zone = &dev->zones[i]; in null_init_zoned_dev() local 106 zone->start = zone->wp = sector; in null_init_zoned_dev() 107 if (zone->start + dev->zone_size_sects > dev_capacity_sects) in null_init_zoned_dev() [all …]
|
/kernel/linux/linux-4.19/include/linux/ |
D | mmzone.h | 359 struct zone { struct 525 static inline unsigned long zone_end_pfn(const struct zone *zone) in zone_end_pfn() argument 527 return zone->zone_start_pfn + zone->spanned_pages; in zone_end_pfn() 530 static inline bool zone_spans_pfn(const struct zone *zone, unsigned long pfn) in zone_spans_pfn() argument 532 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn() 535 static inline bool zone_is_initialized(struct zone *zone) in zone_is_initialized() argument 537 return zone->initialized; in zone_is_initialized() 540 static inline bool zone_is_empty(struct zone *zone) in zone_is_empty() argument 542 return zone->spanned_pages == 0; in zone_is_empty() 549 static inline bool zone_intersects(struct zone *zone, in zone_intersects() argument [all …]
|
D | vmstat.h | 136 static inline void zone_numa_state_add(long x, struct zone *zone, in zone_numa_state_add() argument 139 atomic_long_add(x, &zone->vm_numa_stat[item]); in zone_numa_state_add() 150 static inline unsigned long zone_numa_state_snapshot(struct zone *zone, in zone_numa_state_snapshot() argument 153 long x = atomic_long_read(&zone->vm_numa_stat[item]); in zone_numa_state_snapshot() 157 x += per_cpu_ptr(zone->pageset, cpu)->vm_numa_stat_diff[item]; in zone_numa_state_snapshot() 163 static inline void zone_page_state_add(long x, struct zone *zone, in zone_page_state_add() argument 166 atomic_long_add(x, &zone->vm_stat[item]); in zone_page_state_add() 197 static inline unsigned long zone_page_state(struct zone *zone, in zone_page_state() argument 200 long x = atomic_long_read(&zone->vm_stat[item]); in zone_page_state() 214 static inline unsigned long zone_page_state_snapshot(struct zone *zone, in zone_page_state_snapshot() argument [all …]
|
D | memory_hotplug.h | 11 struct zone; 63 static inline unsigned zone_span_seqbegin(struct zone *zone) in zone_span_seqbegin() argument 65 return read_seqbegin(&zone->span_seqlock); in zone_span_seqbegin() 67 static inline int zone_span_seqretry(struct zone *zone, unsigned iv) in zone_span_seqretry() argument 69 return read_seqretry(&zone->span_seqlock, iv); in zone_span_seqretry() 71 static inline void zone_span_writelock(struct zone *zone) in zone_span_writelock() argument 73 write_seqlock(&zone->span_seqlock); in zone_span_writelock() 75 static inline void zone_span_writeunlock(struct zone *zone) in zone_span_writeunlock() argument 77 write_sequnlock(&zone->span_seqlock); in zone_span_writeunlock() 79 static inline void zone_seqlock_init(struct zone *zone) in zone_seqlock_init() argument [all …]
|
D | compaction.h | 95 extern int fragmentation_index(struct zone *zone, unsigned int order); 100 extern enum compact_result compaction_suitable(struct zone *zone, int order, 103 extern void defer_compaction(struct zone *zone, int order); 104 extern bool compaction_deferred(struct zone *zone, int order); 105 extern void compaction_defer_reset(struct zone *zone, int order, 107 extern bool compaction_restarting(struct zone *zone, int order); 186 static inline enum compact_result compaction_suitable(struct zone *zone, int order, in compaction_suitable() argument 192 static inline void defer_compaction(struct zone *zone, int order) in defer_compaction() argument 196 static inline bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument
|
/kernel/linux/linux-5.10/mm/ |
D | page_alloc.c | 134 struct zone *zone; member 589 static int page_outside_zone_boundaries(struct zone *zone, struct page *page) in page_outside_zone_boundaries() argument 597 seq = zone_span_seqbegin(zone); in page_outside_zone_boundaries() 598 start_pfn = zone->zone_start_pfn; in page_outside_zone_boundaries() 599 sp = zone->spanned_pages; in page_outside_zone_boundaries() 600 if (!zone_spans_pfn(zone, pfn)) in page_outside_zone_boundaries() 602 } while (zone_span_seqretry(zone, seq)); in page_outside_zone_boundaries() 606 pfn, zone_to_nid(zone), zone->name, in page_outside_zone_boundaries() 612 static int page_is_consistent(struct zone *zone, struct page *page) in page_is_consistent() argument 616 if (zone != page_zone(page)) in page_is_consistent() [all …]
|
D | compaction.c | 160 void defer_compaction(struct zone *zone, int order) in defer_compaction() argument 162 zone->compact_considered = 0; in defer_compaction() 163 zone->compact_defer_shift++; in defer_compaction() 165 if (order < zone->compact_order_failed) in defer_compaction() 166 zone->compact_order_failed = order; in defer_compaction() 168 if (zone->compact_defer_shift > COMPACT_MAX_DEFER_SHIFT) in defer_compaction() 169 zone->compact_defer_shift = COMPACT_MAX_DEFER_SHIFT; in defer_compaction() 171 trace_mm_compaction_defer_compaction(zone, order); in defer_compaction() 175 bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument 177 unsigned long defer_limit = 1UL << zone->compact_defer_shift; in compaction_deferred() [all …]
|
D | memory_hotplug.c | 355 static unsigned long find_smallest_section_pfn(int nid, struct zone *zone, in find_smallest_section_pfn() argument 366 if (zone != page_zone(pfn_to_page(start_pfn))) in find_smallest_section_pfn() 376 static unsigned long find_biggest_section_pfn(int nid, struct zone *zone, in find_biggest_section_pfn() argument 391 if (zone != page_zone(pfn_to_page(pfn))) in find_biggest_section_pfn() 400 static void shrink_zone_span(struct zone *zone, unsigned long start_pfn, in shrink_zone_span() argument 404 int nid = zone_to_nid(zone); in shrink_zone_span() 406 zone_span_writelock(zone); in shrink_zone_span() 407 if (zone->zone_start_pfn == start_pfn) { in shrink_zone_span() 414 pfn = find_smallest_section_pfn(nid, zone, end_pfn, in shrink_zone_span() 415 zone_end_pfn(zone)); in shrink_zone_span() [all …]
|
D | vmstat.c | 40 static void zero_zone_numa_counters(struct zone *zone) in zero_zone_numa_counters() argument 45 atomic_long_set(&zone->vm_numa_stat[item], 0); in zero_zone_numa_counters() 47 per_cpu_ptr(zone->pageset, cpu)->vm_numa_stat_diff[item] in zero_zone_numa_counters() 55 struct zone *zone; in zero_zones_numa_counters() local 57 for_each_populated_zone(zone) in zero_zones_numa_counters() 58 zero_zone_numa_counters(zone); in zero_zones_numa_counters() 172 int calculate_pressure_threshold(struct zone *zone) in calculate_pressure_threshold() argument 185 watermark_distance = low_wmark_pages(zone) - min_wmark_pages(zone); in calculate_pressure_threshold() 196 int calculate_normal_threshold(struct zone *zone) in calculate_normal_threshold() argument 231 mem = zone_managed_pages(zone) >> (27 - PAGE_SHIFT); in calculate_normal_threshold() [all …]
|
/kernel/linux/linux-5.10/include/linux/ |
D | mmzone.h | 437 struct zone { struct 604 static inline unsigned long zone_managed_pages(struct zone *zone) in zone_managed_pages() argument 606 return (unsigned long)atomic_long_read(&zone->managed_pages); in zone_managed_pages() 609 static inline unsigned long zone_end_pfn(const struct zone *zone) in zone_end_pfn() argument 611 return zone->zone_start_pfn + zone->spanned_pages; in zone_end_pfn() 614 static inline bool zone_spans_pfn(const struct zone *zone, unsigned long pfn) in zone_spans_pfn() argument 616 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn() 619 static inline bool zone_is_initialized(struct zone *zone) in zone_is_initialized() argument 621 return zone->initialized; in zone_is_initialized() 624 static inline bool zone_is_empty(struct zone *zone) in zone_is_empty() argument [all …]
|
D | memory_hotplug.h | 11 struct zone; 91 static inline unsigned zone_span_seqbegin(struct zone *zone) in zone_span_seqbegin() argument 93 return read_seqbegin(&zone->span_seqlock); in zone_span_seqbegin() 95 static inline int zone_span_seqretry(struct zone *zone, unsigned iv) in zone_span_seqretry() argument 97 return read_seqretry(&zone->span_seqlock, iv); in zone_span_seqretry() 99 static inline void zone_span_writelock(struct zone *zone) in zone_span_writelock() argument 101 write_seqlock(&zone->span_seqlock); in zone_span_writelock() 103 static inline void zone_span_writeunlock(struct zone *zone) in zone_span_writeunlock() argument 105 write_sequnlock(&zone->span_seqlock); in zone_span_writeunlock() 107 static inline void zone_seqlock_init(struct zone *zone) in zone_seqlock_init() argument [all …]
|
D | vmstat.h | 145 static inline void zone_numa_state_add(long x, struct zone *zone, in zone_numa_state_add() argument 148 atomic_long_add(x, &zone->vm_numa_stat[item]); in zone_numa_state_add() 159 static inline unsigned long zone_numa_state_snapshot(struct zone *zone, in zone_numa_state_snapshot() argument 162 long x = atomic_long_read(&zone->vm_numa_stat[item]); in zone_numa_state_snapshot() 166 x += per_cpu_ptr(zone->pageset, cpu)->vm_numa_stat_diff[item]; in zone_numa_state_snapshot() 172 static inline void zone_page_state_add(long x, struct zone *zone, in zone_page_state_add() argument 175 atomic_long_add(x, &zone->vm_stat[item]); in zone_page_state_add() 214 static inline unsigned long zone_page_state(struct zone *zone, in zone_page_state() argument 217 long x = atomic_long_read(&zone->vm_stat[item]); in zone_page_state() 231 static inline unsigned long zone_page_state_snapshot(struct zone *zone, in zone_page_state_snapshot() argument [all …]
|
D | compaction.h | 91 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order); 92 extern int fragmentation_index(struct zone *zone, unsigned int order); 98 extern enum compact_result compaction_suitable(struct zone *zone, int order, 101 extern void defer_compaction(struct zone *zone, int order); 102 extern bool compaction_deferred(struct zone *zone, int order); 103 extern void compaction_defer_reset(struct zone *zone, int order, 105 extern bool compaction_restarting(struct zone *zone, int order); 191 static inline enum compact_result compaction_suitable(struct zone *zone, int order, in compaction_suitable() argument 197 static inline void defer_compaction(struct zone *zone, int order) in defer_compaction() argument 201 static inline bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument
|
/kernel/linux/linux-5.10/include/net/netfilter/ |
D | nf_conntrack_zones.h | 12 return &ct->zone; in nf_ct_zone() 19 nf_ct_zone_init(struct nf_conntrack_zone *zone, u16 id, u8 dir, u8 flags) in nf_ct_zone_init() argument 21 zone->id = id; in nf_ct_zone_init() 22 zone->flags = flags; in nf_ct_zone_init() 23 zone->dir = dir; in nf_ct_zone_init() 25 return zone; in nf_ct_zone_init() 36 if (tmpl->zone.flags & NF_CT_FLAG_MARK) in nf_ct_zone_tmpl() 37 return nf_ct_zone_init(tmp, skb->mark, tmpl->zone.dir, 0); in nf_ct_zone_tmpl() 43 const struct nf_conntrack_zone *zone) in nf_ct_zone_add() argument 46 ct->zone = *zone; in nf_ct_zone_add() [all …]
|
/kernel/linux/linux-4.19/fs/btrfs/ |
D | reada.c | 228 struct reada_zone *zone; in reada_find_zone() local 234 zone = NULL; in reada_find_zone() 236 ret = radix_tree_gang_lookup(&dev->reada_zones, (void **)&zone, in reada_find_zone() 238 if (ret == 1 && logical >= zone->start && logical <= zone->end) { in reada_find_zone() 239 kref_get(&zone->refcnt); in reada_find_zone() 241 return zone; in reada_find_zone() 254 zone = kzalloc(sizeof(*zone), GFP_KERNEL); in reada_find_zone() 255 if (!zone) in reada_find_zone() 260 kfree(zone); in reada_find_zone() 264 zone->start = start; in reada_find_zone() [all …]
|
/kernel/linux/linux-4.19/include/net/netfilter/ |
D | nf_conntrack_zones.h | 14 return &ct->zone; in nf_ct_zone() 21 nf_ct_zone_init(struct nf_conntrack_zone *zone, u16 id, u8 dir, u8 flags) in nf_ct_zone_init() argument 23 zone->id = id; in nf_ct_zone_init() 24 zone->flags = flags; in nf_ct_zone_init() 25 zone->dir = dir; in nf_ct_zone_init() 27 return zone; in nf_ct_zone_init() 38 if (tmpl->zone.flags & NF_CT_FLAG_MARK) in nf_ct_zone_tmpl() 39 return nf_ct_zone_init(tmp, skb->mark, tmpl->zone.dir, 0); in nf_ct_zone_tmpl() 45 const struct nf_conntrack_zone *zone) in nf_ct_zone_add() argument 48 ct->zone = *zone; in nf_ct_zone_add() [all …]
|
/kernel/linux/linux-5.10/drivers/md/ |
D | dm-zoned-metadata.c | 137 struct dm_zone *zone; member 221 static unsigned int dmz_dev_zone_id(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_dev_zone_id() argument 223 if (WARN_ON(!zone)) in dmz_dev_zone_id() 226 return zone->id - zone->dev->zone_offset; in dmz_dev_zone_id() 229 sector_t dmz_start_sect(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_start_sect() argument 231 unsigned int zone_id = dmz_dev_zone_id(zmd, zone); in dmz_start_sect() 236 sector_t dmz_start_block(struct dmz_metadata *zmd, struct dm_zone *zone) in dmz_start_block() argument 238 unsigned int zone_id = dmz_dev_zone_id(zmd, zone); in dmz_start_block() 311 struct dm_zone *zone = kzalloc(sizeof(struct dm_zone), GFP_KERNEL); in dmz_insert() local 313 if (!zone) in dmz_insert() [all …]
|
/kernel/linux/linux-5.10/fs/btrfs/ |
D | reada.c | 229 struct reada_zone *zone; in reada_find_zone() local 235 zone = NULL; in reada_find_zone() 237 ret = radix_tree_gang_lookup(&dev->reada_zones, (void **)&zone, in reada_find_zone() 239 if (ret == 1 && logical >= zone->start && logical <= zone->end) { in reada_find_zone() 240 kref_get(&zone->refcnt); in reada_find_zone() 242 return zone; in reada_find_zone() 255 zone = kzalloc(sizeof(*zone), GFP_KERNEL); in reada_find_zone() 256 if (!zone) in reada_find_zone() 261 kfree(zone); in reada_find_zone() 265 zone->start = start; in reada_find_zone() [all …]
|