Lines Matching refs:entry
65 static inline unsigned long xa_to_value(const void *entry) in xa_to_value() argument
67 return (unsigned long)entry >> 1; in xa_to_value()
77 static inline bool xa_is_value(const void *entry) in xa_is_value() argument
79 return (unsigned long)entry & 1; in xa_is_value()
110 static inline void *xa_untag_pointer(void *entry) in xa_untag_pointer() argument
112 return (void *)((unsigned long)entry & ~3UL); in xa_untag_pointer()
125 static inline unsigned int xa_pointer_tag(void *entry) in xa_pointer_tag() argument
127 return (unsigned long)entry & 3UL; in xa_pointer_tag()
155 static inline unsigned long xa_to_internal(const void *entry) in xa_to_internal() argument
157 return (unsigned long)entry >> 2; in xa_to_internal()
167 static inline bool xa_is_internal(const void *entry) in xa_is_internal() argument
169 return ((unsigned long)entry & 3) == 2; in xa_is_internal()
183 static inline bool xa_is_zero(const void *entry) in xa_is_zero() argument
185 return unlikely(entry == XA_ZERO_ENTRY); in xa_is_zero()
199 static inline bool xa_is_err(const void *entry) in xa_is_err() argument
201 return unlikely(xa_is_internal(entry) && in xa_is_err()
202 entry >= xa_mk_internal(-MAX_ERRNO)); in xa_is_err()
217 static inline int xa_err(void *entry) in xa_err() argument
220 if (xa_is_err(entry)) in xa_err()
221 return (long)entry >> 2; in xa_err()
348 void *xa_store(struct xarray *, unsigned long index, void *entry, gfp_t);
351 void *entry, gfp_t);
442 #define xa_for_each_start(xa, index, entry, start) \ argument
444 entry = xa_find(xa, &index, ULONG_MAX, XA_PRESENT); \
445 entry; \
446 entry = xa_find_after(xa, &index, ULONG_MAX, XA_PRESENT))
469 #define xa_for_each(xa, index, entry) \ argument
470 xa_for_each_start(xa, index, entry, 0)
496 #define xa_for_each_marked(xa, index, entry, filter) \ argument
497 for (index = 0, entry = xa_find(xa, &index, ULONG_MAX, filter); \
498 entry; entry = xa_find_after(xa, &index, ULONG_MAX, filter))
520 void *__xa_store(struct xarray *, unsigned long index, void *entry, gfp_t);
522 void *entry, gfp_t);
524 void *entry, gfp_t);
525 int __must_check __xa_alloc(struct xarray *, u32 *id, void *entry,
527 int __must_check __xa_alloc_cyclic(struct xarray *, u32 *id, void *entry,
547 void *entry, gfp_t gfp) in xa_store_bh() argument
552 curr = __xa_store(xa, index, entry, gfp); in xa_store_bh()
573 void *entry, gfp_t gfp) in xa_store_irq() argument
578 curr = __xa_store(xa, index, entry, gfp); in xa_store_irq()
599 void *entry; in xa_erase_bh() local
602 entry = __xa_erase(xa, index); in xa_erase_bh()
605 return entry; in xa_erase_bh()
623 void *entry; in xa_erase_irq() local
626 entry = __xa_erase(xa, index); in xa_erase_irq()
629 return entry; in xa_erase_irq()
648 void *old, void *entry, gfp_t gfp) in xa_cmpxchg() argument
653 curr = __xa_cmpxchg(xa, index, old, entry, gfp); in xa_cmpxchg()
675 void *old, void *entry, gfp_t gfp) in xa_cmpxchg_bh() argument
680 curr = __xa_cmpxchg(xa, index, old, entry, gfp); in xa_cmpxchg_bh()
702 void *old, void *entry, gfp_t gfp) in xa_cmpxchg_irq() argument
707 curr = __xa_cmpxchg(xa, index, old, entry, gfp); in xa_cmpxchg_irq()
731 unsigned long index, void *entry, gfp_t gfp) in xa_insert() argument
736 err = __xa_insert(xa, index, entry, gfp); in xa_insert()
760 unsigned long index, void *entry, gfp_t gfp) in xa_insert_bh() argument
765 err = __xa_insert(xa, index, entry, gfp); in xa_insert_bh()
789 unsigned long index, void *entry, gfp_t gfp) in xa_insert_irq() argument
794 err = __xa_insert(xa, index, entry, gfp); in xa_insert_irq()
818 void *entry, struct xa_limit limit, gfp_t gfp) in xa_alloc() argument
823 err = __xa_alloc(xa, id, entry, limit, gfp); in xa_alloc()
847 void *entry, struct xa_limit limit, gfp_t gfp) in xa_alloc_bh() argument
852 err = __xa_alloc(xa, id, entry, limit, gfp); in xa_alloc_bh()
876 void *entry, struct xa_limit limit, gfp_t gfp) in xa_alloc_irq() argument
881 err = __xa_alloc(xa, id, entry, limit, gfp); in xa_alloc_irq()
908 static inline int xa_alloc_cyclic(struct xarray *xa, u32 *id, void *entry, in xa_alloc_cyclic() argument
914 err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp); in xa_alloc_cyclic()
941 static inline int xa_alloc_cyclic_bh(struct xarray *xa, u32 *id, void *entry, in xa_alloc_cyclic_bh() argument
947 err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp); in xa_alloc_cyclic_bh()
974 static inline int xa_alloc_cyclic_irq(struct xarray *xa, u32 *id, void *entry, in xa_alloc_cyclic_irq() argument
980 err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp); in xa_alloc_cyclic_irq()
1181 static inline struct xa_node *xa_to_node(const void *entry) in xa_to_node() argument
1183 return (struct xa_node *)((unsigned long)entry - 2); in xa_to_node()
1187 static inline bool xa_is_node(const void *entry) in xa_is_node() argument
1189 return xa_is_internal(entry) && (unsigned long)entry > 4096; in xa_is_node()
1199 static inline unsigned long xa_to_sibling(const void *entry) in xa_to_sibling() argument
1201 return xa_to_internal(entry); in xa_to_sibling()
1210 static inline bool xa_is_sibling(const void *entry) in xa_is_sibling() argument
1212 return IS_ENABLED(CONFIG_XARRAY_MULTI) && xa_is_internal(entry) && in xa_is_sibling()
1213 (entry < xa_mk_sibling(XA_CHUNK_SIZE - 1)); in xa_is_sibling()
1224 static inline bool xa_is_retry(const void *entry) in xa_is_retry() argument
1226 return unlikely(entry == XA_RETRY_ENTRY); in xa_is_retry()
1235 static inline bool xa_is_advanced(const void *entry) in xa_is_advanced() argument
1237 return xa_is_internal(entry) && (entry <= XA_RETRY_ENTRY); in xa_is_advanced()
1447 static inline bool xas_retry(struct xa_state *xas, const void *entry) in xas_retry() argument
1449 if (xa_is_zero(entry)) in xas_retry()
1451 if (!xa_is_retry(entry)) in xas_retry()
1458 void *xas_store(struct xa_state *, void *entry);
1558 void *entry; in xas_next_entry() local
1569 entry = xa_entry(xas->xa, node, xas->xa_offset + 1); in xas_next_entry()
1570 if (unlikely(xa_is_internal(entry))) in xas_next_entry()
1574 } while (!entry); in xas_next_entry()
1576 return entry; in xas_next_entry()
1651 #define xas_for_each(xas, entry, max) \ argument
1652 for (entry = xas_find(xas, max); entry; \
1653 entry = xas_next_entry(xas, max))
1669 #define xas_for_each_marked(xas, entry, max, mark) \ argument
1670 for (entry = xas_find_marked(xas, max, mark); entry; \
1671 entry = xas_next_marked(xas, max, mark))
1686 #define xas_for_each_conflict(xas, entry) \ argument
1687 while ((entry = xas_find_conflict(xas)))