• Home
  • Raw
  • Download

Lines Matching refs:entry

71 static inline unsigned long xa_to_value(const void *entry)  in xa_to_value()  argument
73 return (unsigned long)entry >> 1; in xa_to_value()
83 static inline bool xa_is_value(const void *entry) in xa_is_value() argument
85 return (unsigned long)entry & 1; in xa_is_value()
116 static inline void *xa_untag_pointer(void *entry) in xa_untag_pointer() argument
118 return (void *)((unsigned long)entry & ~3UL); in xa_untag_pointer()
131 static inline unsigned int xa_pointer_tag(void *entry) in xa_pointer_tag() argument
133 return (unsigned long)entry & 3UL; in xa_pointer_tag()
161 static inline unsigned long xa_to_internal(const void *entry) in xa_to_internal() argument
163 return (unsigned long)entry >> 2; in xa_to_internal()
173 static inline bool xa_is_internal(const void *entry) in xa_is_internal() argument
175 return ((unsigned long)entry & 3) == 2; in xa_is_internal()
189 static inline bool xa_is_zero(const void *entry) in xa_is_zero() argument
191 return unlikely(entry == XA_ZERO_ENTRY); in xa_is_zero()
205 static inline bool xa_is_err(const void *entry) in xa_is_err() argument
207 return unlikely(xa_is_internal(entry) && in xa_is_err()
208 entry >= xa_mk_internal(-MAX_ERRNO)); in xa_is_err()
223 static inline int xa_err(void *entry) in xa_err() argument
226 if (xa_is_err(entry)) in xa_err()
227 return (long)entry >> 2; in xa_err()
356 void *xa_store(struct xarray *, unsigned long index, void *entry, gfp_t);
359 void *entry, gfp_t);
451 #define xa_for_each_range(xa, index, entry, start, last) \ argument
453 entry = xa_find(xa, &index, last, XA_PRESENT); \
454 entry; \
455 entry = xa_find_after(xa, &index, last, XA_PRESENT))
480 #define xa_for_each_start(xa, index, entry, start) \ argument
481 xa_for_each_range(xa, index, entry, start, ULONG_MAX)
504 #define xa_for_each(xa, index, entry) \ argument
505 xa_for_each_start(xa, index, entry, 0)
531 #define xa_for_each_marked(xa, index, entry, filter) \ argument
532 for (index = 0, entry = xa_find(xa, &index, ULONG_MAX, filter); \
533 entry; entry = xa_find_after(xa, &index, ULONG_MAX, filter))
563 void *__xa_store(struct xarray *, unsigned long index, void *entry, gfp_t);
565 void *entry, gfp_t);
567 void *entry, gfp_t);
568 int __must_check __xa_alloc(struct xarray *, u32 *id, void *entry,
570 int __must_check __xa_alloc_cyclic(struct xarray *, u32 *id, void *entry,
590 void *entry, gfp_t gfp) in xa_store_bh() argument
596 curr = __xa_store(xa, index, entry, gfp); in xa_store_bh()
617 void *entry, gfp_t gfp) in xa_store_irq() argument
623 curr = __xa_store(xa, index, entry, gfp); in xa_store_irq()
644 void *entry; in xa_erase_bh() local
647 entry = __xa_erase(xa, index); in xa_erase_bh()
650 return entry; in xa_erase_bh()
668 void *entry; in xa_erase_irq() local
671 entry = __xa_erase(xa, index); in xa_erase_irq()
674 return entry; in xa_erase_irq()
693 void *old, void *entry, gfp_t gfp) in xa_cmpxchg() argument
699 curr = __xa_cmpxchg(xa, index, old, entry, gfp); in xa_cmpxchg()
721 void *old, void *entry, gfp_t gfp) in xa_cmpxchg_bh() argument
727 curr = __xa_cmpxchg(xa, index, old, entry, gfp); in xa_cmpxchg_bh()
749 void *old, void *entry, gfp_t gfp) in xa_cmpxchg_irq() argument
755 curr = __xa_cmpxchg(xa, index, old, entry, gfp); in xa_cmpxchg_irq()
779 unsigned long index, void *entry, gfp_t gfp) in xa_insert() argument
785 err = __xa_insert(xa, index, entry, gfp); in xa_insert()
809 unsigned long index, void *entry, gfp_t gfp) in xa_insert_bh() argument
815 err = __xa_insert(xa, index, entry, gfp); in xa_insert_bh()
839 unsigned long index, void *entry, gfp_t gfp) in xa_insert_irq() argument
845 err = __xa_insert(xa, index, entry, gfp); in xa_insert_irq()
872 void *entry, struct xa_limit limit, gfp_t gfp) in xa_alloc() argument
878 err = __xa_alloc(xa, id, entry, limit, gfp); in xa_alloc()
905 void *entry, struct xa_limit limit, gfp_t gfp) in xa_alloc_bh() argument
911 err = __xa_alloc(xa, id, entry, limit, gfp); in xa_alloc_bh()
938 void *entry, struct xa_limit limit, gfp_t gfp) in xa_alloc_irq() argument
944 err = __xa_alloc(xa, id, entry, limit, gfp); in xa_alloc_irq()
974 static inline int xa_alloc_cyclic(struct xarray *xa, u32 *id, void *entry, in xa_alloc_cyclic() argument
981 err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp); in xa_alloc_cyclic()
1011 static inline int xa_alloc_cyclic_bh(struct xarray *xa, u32 *id, void *entry, in xa_alloc_cyclic_bh() argument
1018 err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp); in xa_alloc_cyclic_bh()
1048 static inline int xa_alloc_cyclic_irq(struct xarray *xa, u32 *id, void *entry, in xa_alloc_cyclic_irq() argument
1055 err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp); in xa_alloc_cyclic_irq()
1256 static inline struct xa_node *xa_to_node(const void *entry) in xa_to_node() argument
1258 return (struct xa_node *)((unsigned long)entry - 2); in xa_to_node()
1262 static inline bool xa_is_node(const void *entry) in xa_is_node() argument
1264 return xa_is_internal(entry) && (unsigned long)entry > 4096; in xa_is_node()
1274 static inline unsigned long xa_to_sibling(const void *entry) in xa_to_sibling() argument
1276 return xa_to_internal(entry); in xa_to_sibling()
1285 static inline bool xa_is_sibling(const void *entry) in xa_is_sibling() argument
1287 return IS_ENABLED(CONFIG_XARRAY_MULTI) && xa_is_internal(entry) && in xa_is_sibling()
1288 (entry < xa_mk_sibling(XA_CHUNK_SIZE - 1)); in xa_is_sibling()
1299 static inline bool xa_is_retry(const void *entry) in xa_is_retry() argument
1301 return unlikely(entry == XA_RETRY_ENTRY); in xa_is_retry()
1310 static inline bool xa_is_advanced(const void *entry) in xa_is_advanced() argument
1312 return xa_is_internal(entry) && (entry <= XA_RETRY_ENTRY); in xa_is_advanced()
1526 static inline bool xas_retry(struct xa_state *xas, const void *entry) in xas_retry() argument
1528 if (xa_is_zero(entry)) in xas_retry()
1530 if (!xa_is_retry(entry)) in xas_retry()
1537 void *xas_store(struct xa_state *, void *entry);
1556 void xas_split(struct xa_state *, void *entry, unsigned int order);
1557 void xas_split_alloc(struct xa_state *, void *entry, unsigned int order, gfp_t);
1569 static inline void xas_split(struct xa_state *xas, void *entry, in xas_split() argument
1572 xas_store(xas, entry); in xas_split()
1575 static inline void xas_split_alloc(struct xa_state *xas, void *entry, in xas_split_alloc() argument
1598 void *entry; in xas_reload() local
1605 entry = xa_entry(xas->xa, node, offset); in xas_reload()
1606 if (!xa_is_sibling(entry)) in xas_reload()
1607 return entry; in xas_reload()
1608 offset = xa_to_sibling(entry); in xas_reload()
1701 void *entry; in xas_next_entry() local
1712 entry = xa_entry(xas->xa, node, xas->xa_offset + 1); in xas_next_entry()
1713 if (unlikely(xa_is_internal(entry))) in xas_next_entry()
1717 } while (!entry); in xas_next_entry()
1719 return entry; in xas_next_entry()
1759 void *entry; in xas_next_marked() local
1771 entry = xa_entry(xas->xa, node, offset); in xas_next_marked()
1772 if (!entry) in xas_next_marked()
1774 return entry; in xas_next_marked()
1798 #define xas_for_each(xas, entry, max) \ argument
1799 for (entry = xas_find(xas, max); entry; \
1800 entry = xas_next_entry(xas, max))
1816 #define xas_for_each_marked(xas, entry, max, mark) \ argument
1817 for (entry = xas_find_marked(xas, max, mark); entry; \
1818 entry = xas_next_marked(xas, max, mark))
1832 #define xas_for_each_conflict(xas, entry) \ argument
1833 while ((entry = xas_find_conflict(xas)))