Home
last modified time | relevance | path

Searched refs:zone (Results 1 – 25 of 32) sorted by relevance

12

/include/net/netfilter/
Dnf_conntrack_zones.h12 return &ct->zone; in nf_ct_zone()
19 nf_ct_zone_init(struct nf_conntrack_zone *zone, u16 id, u8 dir, u8 flags) in nf_ct_zone_init() argument
21 zone->id = id; in nf_ct_zone_init()
22 zone->flags = flags; in nf_ct_zone_init()
23 zone->dir = dir; in nf_ct_zone_init()
25 return zone; in nf_ct_zone_init()
36 if (tmpl->zone.flags & NF_CT_FLAG_MARK) in nf_ct_zone_tmpl()
37 return nf_ct_zone_init(tmp, skb->mark, tmpl->zone.dir, 0); in nf_ct_zone_tmpl()
43 const struct nf_conntrack_zone *zone) in nf_ct_zone_add() argument
46 ct->zone = *zone; in nf_ct_zone_add()
[all …]
Dnf_conntrack_count.h26 const struct nf_conntrack_zone *zone);
30 const struct nf_conntrack_zone *zone);
Dnf_conntrack_expect.h92 const struct nf_conntrack_zone *zone,
97 const struct nf_conntrack_zone *zone,
102 const struct nf_conntrack_zone *zone,
Dnf_conntrack.h76 struct nf_conntrack_zone zone; member
240 const struct nf_conntrack_zone *zone,
329 const struct nf_conntrack_zone *zone,
Dnf_conntrack_core.h50 const struct nf_conntrack_zone *zone,
/include/linux/
Dmemory_hotplug.h11 struct zone;
91 static inline unsigned zone_span_seqbegin(struct zone *zone) in zone_span_seqbegin() argument
93 return read_seqbegin(&zone->span_seqlock); in zone_span_seqbegin()
95 static inline int zone_span_seqretry(struct zone *zone, unsigned iv) in zone_span_seqretry() argument
97 return read_seqretry(&zone->span_seqlock, iv); in zone_span_seqretry()
99 static inline void zone_span_writelock(struct zone *zone) in zone_span_writelock() argument
101 write_seqlock(&zone->span_seqlock); in zone_span_writelock()
103 static inline void zone_span_writeunlock(struct zone *zone) in zone_span_writeunlock() argument
105 write_sequnlock(&zone->span_seqlock); in zone_span_writeunlock()
107 static inline void zone_seqlock_init(struct zone *zone) in zone_seqlock_init() argument
[all …]
Dcompaction.h93 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order);
94 extern int fragmentation_index(struct zone *zone, unsigned int order);
100 extern enum compact_result compaction_suitable(struct zone *zone, int order,
103 extern void defer_compaction(struct zone *zone, int order);
104 extern bool compaction_deferred(struct zone *zone, int order);
105 extern void compaction_defer_reset(struct zone *zone, int order,
107 extern bool compaction_restarting(struct zone *zone, int order);
195 static inline enum compact_result compaction_suitable(struct zone *zone, int order, in compaction_suitable() argument
201 static inline void defer_compaction(struct zone *zone, int order) in defer_compaction() argument
205 static inline bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument
Dpage-isolation.h6 static inline bool has_isolate_pageblock(struct zone *zone) in has_isolate_pageblock() argument
8 return zone->nr_isolate_pageblock; in has_isolate_pageblock()
19 static inline bool has_isolate_pageblock(struct zone *zone) in has_isolate_pageblock() argument
36 struct page *has_unmovable_pages(struct zone *zone, struct page *page,
39 int move_freepages_block(struct zone *zone, struct page *page,
Dmmzone.h431 struct zone { struct
609 static inline unsigned long zone_managed_pages(struct zone *zone) in zone_managed_pages() argument
611 return (unsigned long)atomic_long_read(&zone->managed_pages); in zone_managed_pages()
614 static inline unsigned long zone_cma_pages(struct zone *zone) in zone_cma_pages() argument
617 return zone->cma_pages; in zone_cma_pages()
623 static inline unsigned long zone_end_pfn(const struct zone *zone) in zone_end_pfn() argument
625 return zone->zone_start_pfn + zone->spanned_pages; in zone_end_pfn()
628 static inline bool zone_spans_pfn(const struct zone *zone, unsigned long pfn) in zone_spans_pfn() argument
630 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn()
633 static inline bool zone_is_initialized(struct zone *zone) in zone_is_initialized() argument
[all …]
Dvmstat.h145 static inline void zone_numa_state_add(long x, struct zone *zone, in zone_numa_state_add() argument
148 atomic_long_add(x, &zone->vm_numa_stat[item]); in zone_numa_state_add()
159 static inline unsigned long zone_numa_state_snapshot(struct zone *zone, in zone_numa_state_snapshot() argument
162 long x = atomic_long_read(&zone->vm_numa_stat[item]); in zone_numa_state_snapshot()
166 x += per_cpu_ptr(zone->pageset, cpu)->vm_numa_stat_diff[item]; in zone_numa_state_snapshot()
172 static inline void zone_page_state_add(long x, struct zone *zone, in zone_page_state_add() argument
175 atomic_long_add(x, &zone->vm_stat[item]); in zone_page_state_add()
214 static inline unsigned long zone_page_state(struct zone *zone, in zone_page_state() argument
217 long x = atomic_long_read(&zone->vm_stat[item]); in zone_page_state()
231 static inline unsigned long zone_page_state_snapshot(struct zone *zone, in zone_page_state_snapshot() argument
[all …]
Ddtpm.h16 struct powercap_zone zone; member
60 static inline struct dtpm *to_dtpm(struct powercap_zone *zone) in to_dtpm() argument
62 return container_of(zone, struct dtpm, zone); in to_dtpm()
Dcpuset.h80 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed()
85 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed()
222 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed()
227 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed()
Dpage_owner.h23 pg_data_t *pgdat, struct zone *zone);
Dgfp.h614 void drain_zone_pages(struct zone *zone, struct per_cpu_pages *pcp);
615 void drain_all_pages(struct zone *zone);
616 void drain_local_pages(struct zone *zone);
Dmemblock.h271 void __next_mem_pfn_range_in_zone(u64 *idx, struct zone *zone,
288 #define for_each_free_mem_pfn_range_in_zone(i, zone, p_start, p_end) \ argument
290 __next_mem_pfn_range_in_zone(&i, zone, p_start, p_end); \
292 __next_mem_pfn_range_in_zone(&i, zone, p_start, p_end))
306 #define for_each_free_mem_pfn_range_in_zone_from(i, zone, p_start, p_end) \ argument
308 __next_mem_pfn_range_in_zone(&i, zone, p_start, p_end))
Dswap.h140 struct zone;
350 extern void lru_add_drain_cpu_zone(struct zone *zone);
369 extern unsigned long zone_reclaimable_pages(struct zone *zone);
Dsuspend.h369 extern void mark_free_pages(struct zone *zone);
/include/trace/events/
Dcompaction.h199 TP_PROTO(struct zone *zone,
203 TP_ARGS(zone, order, ret),
213 __entry->nid = zone_to_nid(zone);
214 __entry->idx = zone_idx(zone);
228 TP_PROTO(struct zone *zone,
232 TP_ARGS(zone, order, ret)
237 TP_PROTO(struct zone *zone,
241 TP_ARGS(zone, order, ret)
246 TP_PROTO(struct zone *zone, int order),
248 TP_ARGS(zone, order),
[all …]
/include/uapi/linux/netfilter/
Dxt_CT.h21 __u16 zone; member
32 __u16 zone; member
/include/net/tc_act/
Dtc_connmark.h10 u16 zone; member
Dtc_ct.h14 u16 zone; member
46 return to_ct_params(a)->zone; in tcf_ct_zone()
Dtc_ctinfo.h13 u16 zone; member
/include/uapi/linux/tc_act/
Dtc_connmark.h10 __u16 zone; member
/include/trace/hooks/
Dmm.h66 TP_PROTO(struct zone *preferred_zone, struct zone *zone,
69 TP_ARGS(preferred_zone, zone, order,
203 int migratetype, struct zone *preferred_zone, struct page **page, bool *should_alloc_retry),
207 TP_PROTO(bool force, struct zone *zone, bool *skip_unreserve_highatomic),
208 TP_ARGS(force, zone, skip_unreserve_highatomic));
/include/uapi/linux/
Dtipc.h291 static inline __u32 tipc_addr(unsigned int zone, in tipc_addr() argument
295 return (zone << TIPC_ZONE_OFFSET) | in tipc_addr()

12