| /include/linux/ |
| D | swapops.h | 69 static inline bool is_pfn_swap_entry(swp_entry_t entry); 98 static inline unsigned swp_type(swp_entry_t entry) in swp_type() argument 100 return (entry.val >> SWP_TYPE_SHIFT); in swp_type() 107 static inline pgoff_t swp_offset(swp_entry_t entry) in swp_offset() argument 109 return entry.val & SWP_OFFSET_MASK; in swp_offset() 117 static inline unsigned long swp_offset_pfn(swp_entry_t entry) in swp_offset_pfn() argument 119 VM_BUG_ON(!is_pfn_swap_entry(entry)); in swp_offset_pfn() 120 return swp_offset(entry) & SWP_PFN_MASK; in swp_offset_pfn() 146 static inline pte_t swp_entry_to_pte(swp_entry_t entry) in swp_entry_to_pte() argument 150 arch_entry = __swp_entry(swp_type(entry), swp_offset(entry)); in swp_entry_to_pte() [all …]
|
| D | resource_ext.h | 34 static inline void resource_list_add(struct resource_entry *entry, in resource_list_add() argument 37 list_add(&entry->node, head); in resource_list_add() 40 static inline void resource_list_add_tail(struct resource_entry *entry, in resource_list_add_tail() argument 43 list_add_tail(&entry->node, head); in resource_list_add_tail() 46 static inline void resource_list_del(struct resource_entry *entry) in resource_list_del() argument 48 list_del(&entry->node); in resource_list_del() 51 static inline void resource_list_free_entry(struct resource_entry *entry) in resource_list_free_entry() argument 53 kfree(entry); in resource_list_free_entry() 57 resource_list_destroy_entry(struct resource_entry *entry) in resource_list_destroy_entry() argument 59 resource_list_del(entry); in resource_list_destroy_entry() [all …]
|
| D | bpf_mprog.h | 115 #define bpf_mprog_foreach_tuple(entry, fp, cp, t) \ argument 116 for (fp = &entry->fp_items[0], cp = &entry->parent->cp_items[0];\ 124 #define bpf_mprog_foreach_prog(entry, fp, p) \ argument 125 for (fp = &entry->fp_items[0]; \ 159 bpf_mprog_peer(const struct bpf_mprog_entry *entry) in bpf_mprog_peer() argument 161 if (entry == &entry->parent->a) in bpf_mprog_peer() 162 return &entry->parent->b; in bpf_mprog_peer() 164 return &entry->parent->a; in bpf_mprog_peer() 179 static inline void bpf_mprog_inc(struct bpf_mprog_entry *entry) in bpf_mprog_inc() argument 181 entry->parent->count++; in bpf_mprog_inc() [all …]
|
| D | xarray.h | 71 static inline unsigned long xa_to_value(const void *entry) in xa_to_value() argument 73 return (unsigned long)entry >> 1; in xa_to_value() 83 static inline bool xa_is_value(const void *entry) in xa_is_value() argument 85 return (unsigned long)entry & 1; in xa_is_value() 116 static inline void *xa_untag_pointer(void *entry) in xa_untag_pointer() argument 118 return (void *)((unsigned long)entry & ~3UL); in xa_untag_pointer() 131 static inline unsigned int xa_pointer_tag(void *entry) in xa_pointer_tag() argument 133 return (unsigned long)entry & 3UL; in xa_pointer_tag() 161 static inline unsigned long xa_to_internal(const void *entry) in xa_to_internal() argument 163 return (unsigned long)entry >> 2; in xa_to_internal() [all …]
|
| D | mbcache.h | 46 struct mb_cache_entry *entry); 47 void mb_cache_entry_wait_unused(struct mb_cache_entry *entry); 49 struct mb_cache_entry *entry) in mb_cache_entry_put() argument 51 unsigned int cnt = atomic_dec_return(&entry->e_refcnt); in mb_cache_entry_put() 55 wake_up_var(&entry->e_refcnt); in mb_cache_entry_put() 58 __mb_cache_entry_free(cache, entry); in mb_cache_entry_put() 68 struct mb_cache_entry *entry); 70 struct mb_cache_entry *entry);
|
| D | swap.h | 357 swp_entry_t entry = folio->swap; in page_swap_entry() local 359 entry.val += folio_page_idx(folio, page); in page_swap_entry() 360 return entry; in page_swap_entry() 498 void put_swap_folio(struct folio *folio, swp_entry_t entry); 504 extern int swapcache_prepare(swp_entry_t entry, int nr); 505 extern void swap_free_nr(swp_entry_t entry, int nr_pages); 507 extern void free_swap_and_cache_nr(swp_entry_t entry, int nr); 512 extern int __swap_count(swp_entry_t entry); 513 extern int swap_swapcount(struct swap_info_struct *si, swp_entry_t entry); 514 extern int swp_swapcount(swp_entry_t entry); [all …]
|
| D | jump_label.h | 123 static inline unsigned long jump_entry_code(const struct jump_entry *entry) in jump_entry_code() argument 125 return (unsigned long)&entry->code + entry->code; in jump_entry_code() 128 static inline unsigned long jump_entry_target(const struct jump_entry *entry) in jump_entry_target() argument 130 return (unsigned long)&entry->target + entry->target; in jump_entry_target() 133 static inline struct static_key *jump_entry_key(const struct jump_entry *entry) in jump_entry_key() argument 135 long offset = entry->key & ~3L; in jump_entry_key() 137 return (struct static_key *)((unsigned long)&entry->key + offset); in jump_entry_key() 142 static inline unsigned long jump_entry_code(const struct jump_entry *entry) in jump_entry_code() argument 144 return entry->code; in jump_entry_code() 147 static inline unsigned long jump_entry_target(const struct jump_entry *entry) in jump_entry_target() argument [all …]
|
| D | list.h | 96 extern bool __list_valid_slowpath __list_del_entry_valid_or_report(struct list_head *entry); 106 static __always_inline bool __list_del_entry_valid(struct list_head *entry) in __list_del_entry_valid() argument 111 struct list_head *prev = entry->prev; in __list_del_entry_valid() 112 struct list_head *next = entry->next; in __list_del_entry_valid() 119 if (likely(prev->next == entry && next->prev == entry)) in __list_del_entry_valid() 124 ret &= __list_del_entry_valid_or_report(entry); in __list_del_entry_valid() 134 static inline bool __list_del_entry_valid(struct list_head *entry) in __list_del_entry_valid() argument 207 static inline void __list_del_clearprev(struct list_head *entry) in __list_del_clearprev() argument 209 __list_del(entry->prev, entry->next); in __list_del_clearprev() 210 entry->prev = NULL; in __list_del_clearprev() [all …]
|
| D | idr.h | 187 #define idr_for_each_entry(idr, entry, id) \ argument 188 for (id = 0; ((entry) = idr_get_next(idr, &(id))) != NULL; id += 1U) 201 #define idr_for_each_entry_ul(idr, entry, tmp, id) \ argument 203 ((entry) = tmp <= id ? idr_get_next_ul(idr, &(id)) : NULL) != NULL; \ 214 #define idr_for_each_entry_continue(idr, entry, id) \ argument 215 for ((entry) = idr_get_next((idr), &(id)); \ 216 entry; \ 217 ++id, (entry) = idr_get_next((idr), &(id))) 230 #define idr_for_each_entry_continue_ul(idr, entry, tmp, id) \ argument 232 ((entry) = tmp <= id ? idr_get_next_ul(idr, &(id)) : NULL) != NULL; \
|
| D | maple_tree.h | 329 void *entry, gfp_t gfp); 331 unsigned long last, void *entry, gfp_t gfp); 333 void *entry, unsigned long size, unsigned long min, 336 void *entry, unsigned long range_lo, unsigned long range_hi, 339 void *entry, unsigned long size, unsigned long min, 343 unsigned long last, void *entry, gfp_t gfp); 345 void *entry, gfp_t gfp); 464 void *entry; /* The entry to write */ member 500 .entry = wr_entry, \ 511 void *mas_store(struct ma_state *mas, void *entry); [all …]
|
| D | pldmfw.h | 20 struct list_head entry; member 59 struct list_head entry; member 92 struct list_head entry; member
|
| /include/net/ |
| D | tcx.h | 34 static inline struct tcx_entry *tcx_entry(struct bpf_mprog_entry *entry) in tcx_entry() argument 36 struct bpf_mprog_bundle *bundle = entry->parent; in tcx_entry() 58 tcx_entry_update(struct net_device *dev, struct bpf_mprog_entry *entry, in tcx_entry_update() argument 63 rcu_assign_pointer(dev->tcx_ingress, entry); in tcx_entry_update() 65 rcu_assign_pointer(dev->tcx_egress, entry); in tcx_entry_update() 90 static inline void tcx_entry_free(struct bpf_mprog_entry *entry) in tcx_entry_free() argument 92 kfree_rcu(tcx_entry(entry), rcu); in tcx_entry_free() 98 struct bpf_mprog_entry *entry = tcx_entry_fetch(dev, ingress); in tcx_entry_fetch_or_create() local 101 if (!entry) { in tcx_entry_fetch_or_create() 102 entry = tcx_entry_create(); in tcx_entry_fetch_or_create() [all …]
|
| /include/sound/ |
| D | info.h | 31 void (*read)(struct snd_info_entry *entry, 33 void (*write)(struct snd_info_entry *entry, 38 int (*open)(struct snd_info_entry *entry, 40 int (*release)(struct snd_info_entry *entry, 42 ssize_t (*read)(struct snd_info_entry *entry, void *file_private_data, 45 ssize_t (*write)(struct snd_info_entry *entry, void *file_private_data, 48 loff_t (*llseek)(struct snd_info_entry *entry, 51 __poll_t (*poll)(struct snd_info_entry *entry, 54 int (*ioctl)(struct snd_info_entry *entry, void *file_private_data, 56 int (*mmap)(struct snd_info_entry *entry, void *file_private_data, [all …]
|
| /include/trace/ |
| D | perf.h | 22 struct trace_event_raw_##call *entry; \ 39 __entry_size = ALIGN(__data_size + sizeof(*entry) + sizeof(u32),\ 43 entry = perf_trace_buf_alloc(__entry_size, &__regs, &rctx); \ 44 if (!entry) \ 53 perf_trace_run_bpf_submit(entry, __entry_size, rctx, \
|
| D | trace_events.h | 218 struct trace_entry *entry; \ 221 entry = iter->ent; \ 223 if (entry->type != event_##call.event.type) { \ 228 field = (typeof(field))entry; \ 264 struct trace_event_raw_##call __maybe_unused *entry; \ 389 struct trace_event_raw_##call *entry; \ 397 entry = trace_event_buffer_reserve(&fbuffer, trace_file, \ 398 sizeof(*entry) + __data_size); \ 400 if (!entry) \
|
| D | trace_custom_events.h | 126 struct trace_custom_event_raw_##call __maybe_unused *entry; \ 148 struct trace_custom_event_raw_##call *entry; \ 156 entry = trace_event_buffer_reserve(&fbuffer, trace_file, \ 157 sizeof(*entry) + __data_size); \ 159 if (!entry) \
|
| /include/linux/netfilter_arp/ |
| D | arp_tables.h | 20 struct arpt_entry entry; member 25 struct arpt_entry entry; member 37 .entry = ARPT_ENTRY_INIT(sizeof(struct arpt_standard)), \ 45 .entry = ARPT_ENTRY_INIT(sizeof(struct arpt_error)), \
|
| /include/linux/netfilter_ipv4/ |
| D | ip_tables.h | 34 struct ipt_entry entry; member 39 struct ipt_entry entry; member 51 .entry = IPT_ENTRY_INIT(sizeof(struct ipt_standard)), \ 59 .entry = IPT_ENTRY_INIT(sizeof(struct ipt_error)), \
|
| /include/net/tc_act/ |
| D | tc_gate.h | 113 struct tcfg_gate_entry *entry; in tcf_gate_get_list() local 120 list_for_each_entry(entry, &p->entries, list) in tcf_gate_get_list() 131 list_for_each_entry(entry, &p->entries, list) { in tcf_gate_get_list() 132 oe[i].gate_state = entry->gate_state; in tcf_gate_get_list() 133 oe[i].interval = entry->interval; in tcf_gate_get_list() 134 oe[i].ipv = entry->ipv; in tcf_gate_get_list() 135 oe[i].maxoctets = entry->maxoctets; in tcf_gate_get_list()
|
| /include/drm/ |
| D | drm_gpuvm.h | 122 struct list_head entry; member 141 struct list_head entry; member 430 if (va && !list_is_last(&va->rb.entry, &va->vm->rb.list)) in __drm_gpuva_next() 431 return list_next_entry(va, rb.entry); in __drm_gpuva_next() 490 list_for_each_entry(va__, &(gpuvm__)->rb.list, rb.entry) 503 list_for_each_entry_safe(va__, next__, &(gpuvm__)->rb.list, rb.entry) 703 } entry; member 772 list_for_each_entry(va__, &(vm_bo)->list.gpuva, gem.entry) 788 list_for_each_entry_safe(va__, next__, &(vm_bo)->list.gpuva, gem.entry) 948 struct list_head entry; member [all …]
|
| /include/uapi/linux/ |
| D | nexthop.h | 23 static inline __u16 nexthop_grp_weight(const struct nexthop_grp *entry) in nexthop_grp_weight() argument 25 return ((entry->weight_high << 8) | entry->weight) + 1; in nexthop_grp_weight()
|
| /include/uapi/linux/netfilter_ipv6/ |
| D | ip6_tables.h | 133 struct ip6t_entry entry; member 138 struct ip6t_entry entry; member 150 .entry = IP6T_ENTRY_INIT(sizeof(struct ip6t_standard)), \ 158 .entry = IP6T_ENTRY_INIT(sizeof(struct ip6t_error)), \
|
| /include/trace/stages/ |
| D | stage5_get_offsets.h | 10 #define __entry entry 54 offsetof(typeof(*entry), __data); \ 75 offsetof(typeof(*entry), __data) - \ 76 offsetof(typeof(*entry), __rel_loc_##item) - \
|
| /include/uapi/linux/netfilter/ |
| D | x_tables.h | 180 #define xt_ematch_foreach(pos, entry) \ argument 181 for ((pos) = (struct xt_entry_match *)entry->elems; \ 182 (pos) < (struct xt_entry_match *)((char *)(entry) + \ 183 (entry)->target_offset); \
|
| /include/linux/platform_data/x86/ |
| D | simatic-ipc.h | 49 u8 entry; in simatic_ipc_get_station_id() member 58 data_entry->entry == SIMATIC_IPC_DMI_ENTRY && in simatic_ipc_get_station_id()
|