Searched refs:atomic_read (Results 1 – 25 of 61) sorted by relevance
123
28 return atomic_read(&lock->val); in queued_spin_is_locked()54 return atomic_read(&lock->val) & ~_Q_LOCKED_MASK; in queued_spin_is_contended()63 u32 val = atomic_read(&lock->val); in queued_spin_trylock()
42 cnts = atomic_read(&lock->cnts); in queued_read_trylock()61 cnts = atomic_read(&lock->cnts); in queued_write_trylock()
168 #ifndef atomic_read169 #define atomic_read(v) READ_ONCE((v)->counter) macro
184 __entry->n_children = atomic_read(&cookie->n_children);185 __entry->n_active = atomic_read(&cookie->n_active);233 __entry->p_usage = atomic_read(&cookie->parent->usage);234 __entry->p_n_children = atomic_read(&cookie->parent->n_children);263 __entry->usage = atomic_read(&cookie->usage);264 __entry->n_children = atomic_read(&cookie->n_children);265 __entry->n_active = atomic_read(&cookie->n_active);291 __entry->usage = atomic_read(&cookie->usage);292 __entry->n_children = atomic_read(&cookie->n_children);293 __entry->n_active = atomic_read(&cookie->n_active);[all …]
38 __entry->usage_count = atomic_read(44 __entry->child_count = atomic_read(
29 atomic_read(&q->owned_by_drv_count);
190 __entry->wcount = atomic_read(&inode->i_writecount);191 __entry->rcount = atomic_read(&inode->i_readcount);192 __entry->icount = atomic_read(&inode->i_count);
84 __entry->refcnt = atomic_read(&mod->refcnt);
84 __entry->rmem_alloc = atomic_read(&sk->sk_rmem_alloc);118 __entry->rmem_alloc = atomic_read(&sk->sk_rmem_alloc);
127 WARN_ON_ONCE(atomic_read(&ioc->active_ref) <= 0); in get_io_context_active()136 WARN_ON_ONCE(atomic_read(&ioc->nr_tasks) <= 0); in ioc_task_link()
38 return atomic_read(&lock->tail) != OSQ_UNLOCKED_VAL; in osq_is_locked()
197 return atomic_read(&key->enabled); in static_key_count()331 if (atomic_read(&key->enabled) != 0) { in static_key_enable()332 WARN_ON_ONCE(atomic_read(&key->enabled) != 1); in static_key_enable()342 if (atomic_read(&key->enabled) != 1) { in static_key_disable()343 WARN_ON_ONCE(atomic_read(&key->enabled) != 0); in static_key_disable()
112 if (!(atomic_read(&__work->flags) & IRQ_WORK_HARD_IRQ))\117 if (!(atomic_read(&__work->flags) & IRQ_WORK_HARD_IRQ))\
281 if (atomic_read(&css->cgroup->congestion_count)) { in blk_cgroup_congested()534 if (WARN_ON_ONCE(atomic_read(&blkg->use_delay) < 0)) in blkcg_use_delay()542 int old = atomic_read(&blkg->use_delay); in blkcg_unuse_delay()581 int old = atomic_read(&blkg->use_delay); in blkcg_set_delay()598 int old = atomic_read(&blkg->use_delay); in blkcg_clear_delay()
122 if (unlikely(!atomic_read(&kfence_allocation_gate))) in kfence_alloc()
67 return atomic_read(&page->_refcount); in page_ref_count()72 return atomic_read(&compound_head(page)->_refcount); in page_count()
497 int old = atomic_read(index); in sbq_index_atomic_inc()513 ws = &sbq->ws[atomic_read(wait_index)]; in sbq_wait_ptr()
359 (irqs_disabled() && (smp_processor_id() == atomic_read(&kgdb_active)))
704 return atomic_read(&mm->tlb_flush_pending); in mm_tlb_flush_pending()716 return atomic_read(&mm->tlb_flush_pending) > 1; in mm_tlb_flush_nested()
147 return atomic_read(&r->refs); in refcount_read()
42 if (likely(!atomic_read(&system_freezing_cnt))) in freezing()
22 return (atomic_read(&cxl_use_count) != 0); in cxl_ctx_in_use()
62 return atomic_read(&queue->job_count); in spsc_queue_count()
444 return atomic_read(&net->ipv4.rt_genid); in rt_genid_ipv4()450 return atomic_read(&net->ipv6.fib6_sernum); in rt_genid_ipv6()483 return atomic_read(&net->fnhe_genid); in fnhe_genid()
231 return atomic_read(&queue->qlen); in reqsk_queue_len()236 return atomic_read(&queue->young); in reqsk_queue_len_young()