Home
last modified time | relevance | path

Searched refs:entry (Results 1 – 25 of 116) sorted by relevance

12345

/include/linux/
Dswapops.h69 static inline bool is_pfn_swap_entry(swp_entry_t entry);
98 static inline unsigned swp_type(swp_entry_t entry) in swp_type() argument
100 return (entry.val >> SWP_TYPE_SHIFT); in swp_type()
107 static inline pgoff_t swp_offset(swp_entry_t entry) in swp_offset() argument
109 return entry.val & SWP_OFFSET_MASK; in swp_offset()
117 static inline unsigned long swp_offset_pfn(swp_entry_t entry) in swp_offset_pfn() argument
119 VM_BUG_ON(!is_pfn_swap_entry(entry)); in swp_offset_pfn()
120 return swp_offset(entry) & SWP_PFN_MASK; in swp_offset_pfn()
146 static inline pte_t swp_entry_to_pte(swp_entry_t entry) in swp_entry_to_pte() argument
150 arch_entry = __swp_entry(swp_type(entry), swp_offset(entry)); in swp_entry_to_pte()
[all …]
Dresource_ext.h34 static inline void resource_list_add(struct resource_entry *entry, in resource_list_add() argument
37 list_add(&entry->node, head); in resource_list_add()
40 static inline void resource_list_add_tail(struct resource_entry *entry, in resource_list_add_tail() argument
43 list_add_tail(&entry->node, head); in resource_list_add_tail()
46 static inline void resource_list_del(struct resource_entry *entry) in resource_list_del() argument
48 list_del(&entry->node); in resource_list_del()
51 static inline void resource_list_free_entry(struct resource_entry *entry) in resource_list_free_entry() argument
53 kfree(entry); in resource_list_free_entry()
57 resource_list_destroy_entry(struct resource_entry *entry) in resource_list_destroy_entry() argument
59 resource_list_del(entry); in resource_list_destroy_entry()
[all …]
Dbpf_mprog.h115 #define bpf_mprog_foreach_tuple(entry, fp, cp, t) \ argument
116 for (fp = &entry->fp_items[0], cp = &entry->parent->cp_items[0];\
124 #define bpf_mprog_foreach_prog(entry, fp, p) \ argument
125 for (fp = &entry->fp_items[0]; \
159 bpf_mprog_peer(const struct bpf_mprog_entry *entry) in bpf_mprog_peer() argument
161 if (entry == &entry->parent->a) in bpf_mprog_peer()
162 return &entry->parent->b; in bpf_mprog_peer()
164 return &entry->parent->a; in bpf_mprog_peer()
179 static inline void bpf_mprog_inc(struct bpf_mprog_entry *entry) in bpf_mprog_inc() argument
181 entry->parent->count++; in bpf_mprog_inc()
[all …]
Dxarray.h71 static inline unsigned long xa_to_value(const void *entry) in xa_to_value() argument
73 return (unsigned long)entry >> 1; in xa_to_value()
83 static inline bool xa_is_value(const void *entry) in xa_is_value() argument
85 return (unsigned long)entry & 1; in xa_is_value()
116 static inline void *xa_untag_pointer(void *entry) in xa_untag_pointer() argument
118 return (void *)((unsigned long)entry & ~3UL); in xa_untag_pointer()
131 static inline unsigned int xa_pointer_tag(void *entry) in xa_pointer_tag() argument
133 return (unsigned long)entry & 3UL; in xa_pointer_tag()
161 static inline unsigned long xa_to_internal(const void *entry) in xa_to_internal() argument
163 return (unsigned long)entry >> 2; in xa_to_internal()
[all …]
Dmbcache.h46 struct mb_cache_entry *entry);
47 void mb_cache_entry_wait_unused(struct mb_cache_entry *entry);
49 struct mb_cache_entry *entry) in mb_cache_entry_put() argument
51 unsigned int cnt = atomic_dec_return(&entry->e_refcnt); in mb_cache_entry_put()
55 wake_up_var(&entry->e_refcnt); in mb_cache_entry_put()
58 __mb_cache_entry_free(cache, entry); in mb_cache_entry_put()
68 struct mb_cache_entry *entry);
70 struct mb_cache_entry *entry);
Dswap.h357 swp_entry_t entry = folio->swap; in page_swap_entry() local
359 entry.val += folio_page_idx(folio, page); in page_swap_entry()
360 return entry; in page_swap_entry()
498 void put_swap_folio(struct folio *folio, swp_entry_t entry);
504 extern int swapcache_prepare(swp_entry_t entry, int nr);
505 extern void swap_free_nr(swp_entry_t entry, int nr_pages);
507 extern void free_swap_and_cache_nr(swp_entry_t entry, int nr);
512 extern int __swap_count(swp_entry_t entry);
513 extern int swap_swapcount(struct swap_info_struct *si, swp_entry_t entry);
514 extern int swp_swapcount(swp_entry_t entry);
[all …]
Djump_label.h123 static inline unsigned long jump_entry_code(const struct jump_entry *entry) in jump_entry_code() argument
125 return (unsigned long)&entry->code + entry->code; in jump_entry_code()
128 static inline unsigned long jump_entry_target(const struct jump_entry *entry) in jump_entry_target() argument
130 return (unsigned long)&entry->target + entry->target; in jump_entry_target()
133 static inline struct static_key *jump_entry_key(const struct jump_entry *entry) in jump_entry_key() argument
135 long offset = entry->key & ~3L; in jump_entry_key()
137 return (struct static_key *)((unsigned long)&entry->key + offset); in jump_entry_key()
142 static inline unsigned long jump_entry_code(const struct jump_entry *entry) in jump_entry_code() argument
144 return entry->code; in jump_entry_code()
147 static inline unsigned long jump_entry_target(const struct jump_entry *entry) in jump_entry_target() argument
[all …]
Dlist.h96 extern bool __list_valid_slowpath __list_del_entry_valid_or_report(struct list_head *entry);
106 static __always_inline bool __list_del_entry_valid(struct list_head *entry) in __list_del_entry_valid() argument
111 struct list_head *prev = entry->prev; in __list_del_entry_valid()
112 struct list_head *next = entry->next; in __list_del_entry_valid()
119 if (likely(prev->next == entry && next->prev == entry)) in __list_del_entry_valid()
124 ret &= __list_del_entry_valid_or_report(entry); in __list_del_entry_valid()
134 static inline bool __list_del_entry_valid(struct list_head *entry) in __list_del_entry_valid() argument
207 static inline void __list_del_clearprev(struct list_head *entry) in __list_del_clearprev() argument
209 __list_del(entry->prev, entry->next); in __list_del_clearprev()
210 entry->prev = NULL; in __list_del_clearprev()
[all …]
Didr.h187 #define idr_for_each_entry(idr, entry, id) \ argument
188 for (id = 0; ((entry) = idr_get_next(idr, &(id))) != NULL; id += 1U)
201 #define idr_for_each_entry_ul(idr, entry, tmp, id) \ argument
203 ((entry) = tmp <= id ? idr_get_next_ul(idr, &(id)) : NULL) != NULL; \
214 #define idr_for_each_entry_continue(idr, entry, id) \ argument
215 for ((entry) = idr_get_next((idr), &(id)); \
216 entry; \
217 ++id, (entry) = idr_get_next((idr), &(id)))
230 #define idr_for_each_entry_continue_ul(idr, entry, tmp, id) \ argument
232 ((entry) = tmp <= id ? idr_get_next_ul(idr, &(id)) : NULL) != NULL; \
Dmaple_tree.h329 void *entry, gfp_t gfp);
331 unsigned long last, void *entry, gfp_t gfp);
333 void *entry, unsigned long size, unsigned long min,
336 void *entry, unsigned long range_lo, unsigned long range_hi,
339 void *entry, unsigned long size, unsigned long min,
343 unsigned long last, void *entry, gfp_t gfp);
345 void *entry, gfp_t gfp);
464 void *entry; /* The entry to write */ member
500 .entry = wr_entry, \
511 void *mas_store(struct ma_state *mas, void *entry);
[all …]
Dpldmfw.h20 struct list_head entry; member
59 struct list_head entry; member
92 struct list_head entry; member
/include/net/
Dtcx.h34 static inline struct tcx_entry *tcx_entry(struct bpf_mprog_entry *entry) in tcx_entry() argument
36 struct bpf_mprog_bundle *bundle = entry->parent; in tcx_entry()
58 tcx_entry_update(struct net_device *dev, struct bpf_mprog_entry *entry, in tcx_entry_update() argument
63 rcu_assign_pointer(dev->tcx_ingress, entry); in tcx_entry_update()
65 rcu_assign_pointer(dev->tcx_egress, entry); in tcx_entry_update()
90 static inline void tcx_entry_free(struct bpf_mprog_entry *entry) in tcx_entry_free() argument
92 kfree_rcu(tcx_entry(entry), rcu); in tcx_entry_free()
98 struct bpf_mprog_entry *entry = tcx_entry_fetch(dev, ingress); in tcx_entry_fetch_or_create() local
101 if (!entry) { in tcx_entry_fetch_or_create()
102 entry = tcx_entry_create(); in tcx_entry_fetch_or_create()
[all …]
/include/sound/
Dinfo.h31 void (*read)(struct snd_info_entry *entry,
33 void (*write)(struct snd_info_entry *entry,
38 int (*open)(struct snd_info_entry *entry,
40 int (*release)(struct snd_info_entry *entry,
42 ssize_t (*read)(struct snd_info_entry *entry, void *file_private_data,
45 ssize_t (*write)(struct snd_info_entry *entry, void *file_private_data,
48 loff_t (*llseek)(struct snd_info_entry *entry,
51 __poll_t (*poll)(struct snd_info_entry *entry,
54 int (*ioctl)(struct snd_info_entry *entry, void *file_private_data,
56 int (*mmap)(struct snd_info_entry *entry, void *file_private_data,
[all …]
/include/trace/
Dperf.h22 struct trace_event_raw_##call *entry; \
39 __entry_size = ALIGN(__data_size + sizeof(*entry) + sizeof(u32),\
43 entry = perf_trace_buf_alloc(__entry_size, &__regs, &rctx); \
44 if (!entry) \
53 perf_trace_run_bpf_submit(entry, __entry_size, rctx, \
Dtrace_events.h218 struct trace_entry *entry; \
221 entry = iter->ent; \
223 if (entry->type != event_##call.event.type) { \
228 field = (typeof(field))entry; \
264 struct trace_event_raw_##call __maybe_unused *entry; \
389 struct trace_event_raw_##call *entry; \
397 entry = trace_event_buffer_reserve(&fbuffer, trace_file, \
398 sizeof(*entry) + __data_size); \
400 if (!entry) \
Dtrace_custom_events.h126 struct trace_custom_event_raw_##call __maybe_unused *entry; \
148 struct trace_custom_event_raw_##call *entry; \
156 entry = trace_event_buffer_reserve(&fbuffer, trace_file, \
157 sizeof(*entry) + __data_size); \
159 if (!entry) \
/include/linux/netfilter_arp/
Darp_tables.h20 struct arpt_entry entry; member
25 struct arpt_entry entry; member
37 .entry = ARPT_ENTRY_INIT(sizeof(struct arpt_standard)), \
45 .entry = ARPT_ENTRY_INIT(sizeof(struct arpt_error)), \
/include/linux/netfilter_ipv4/
Dip_tables.h34 struct ipt_entry entry; member
39 struct ipt_entry entry; member
51 .entry = IPT_ENTRY_INIT(sizeof(struct ipt_standard)), \
59 .entry = IPT_ENTRY_INIT(sizeof(struct ipt_error)), \
/include/net/tc_act/
Dtc_gate.h113 struct tcfg_gate_entry *entry; in tcf_gate_get_list() local
120 list_for_each_entry(entry, &p->entries, list) in tcf_gate_get_list()
131 list_for_each_entry(entry, &p->entries, list) { in tcf_gate_get_list()
132 oe[i].gate_state = entry->gate_state; in tcf_gate_get_list()
133 oe[i].interval = entry->interval; in tcf_gate_get_list()
134 oe[i].ipv = entry->ipv; in tcf_gate_get_list()
135 oe[i].maxoctets = entry->maxoctets; in tcf_gate_get_list()
/include/drm/
Ddrm_gpuvm.h122 struct list_head entry; member
141 struct list_head entry; member
430 if (va && !list_is_last(&va->rb.entry, &va->vm->rb.list)) in __drm_gpuva_next()
431 return list_next_entry(va, rb.entry); in __drm_gpuva_next()
490 list_for_each_entry(va__, &(gpuvm__)->rb.list, rb.entry)
503 list_for_each_entry_safe(va__, next__, &(gpuvm__)->rb.list, rb.entry)
703 } entry; member
772 list_for_each_entry(va__, &(vm_bo)->list.gpuva, gem.entry)
788 list_for_each_entry_safe(va__, next__, &(vm_bo)->list.gpuva, gem.entry)
948 struct list_head entry; member
[all …]
/include/uapi/linux/
Dnexthop.h23 static inline __u16 nexthop_grp_weight(const struct nexthop_grp *entry) in nexthop_grp_weight() argument
25 return ((entry->weight_high << 8) | entry->weight) + 1; in nexthop_grp_weight()
/include/uapi/linux/netfilter_ipv6/
Dip6_tables.h133 struct ip6t_entry entry; member
138 struct ip6t_entry entry; member
150 .entry = IP6T_ENTRY_INIT(sizeof(struct ip6t_standard)), \
158 .entry = IP6T_ENTRY_INIT(sizeof(struct ip6t_error)), \
/include/trace/stages/
Dstage5_get_offsets.h10 #define __entry entry
54 offsetof(typeof(*entry), __data); \
75 offsetof(typeof(*entry), __data) - \
76 offsetof(typeof(*entry), __rel_loc_##item) - \
/include/uapi/linux/netfilter/
Dx_tables.h180 #define xt_ematch_foreach(pos, entry) \ argument
181 for ((pos) = (struct xt_entry_match *)entry->elems; \
182 (pos) < (struct xt_entry_match *)((char *)(entry) + \
183 (entry)->target_offset); \
/include/linux/platform_data/x86/
Dsimatic-ipc.h49 u8 entry; in simatic_ipc_get_station_id() member
58 data_entry->entry == SIMATIC_IPC_DMI_ENTRY && in simatic_ipc_get_station_id()

12345