Searched refs:atomic_read (Results 1 – 25 of 54) sorted by relevance
123
36 return atomic_read(&lock->val); in queued_spin_is_locked()52 return !atomic_read(&lock.val); in queued_spin_value_unlocked()62 return atomic_read(&lock->val) & ~_Q_LOCKED_MASK; in queued_spin_is_contended()71 if (!atomic_read(&lock->val) && in queued_spin_trylock()
61 return !(atomic_read(&lock->cnts) & _QW_WMASK); in queued_read_can_lock()70 return !atomic_read(&lock->cnts); in queued_write_can_lock()82 cnts = atomic_read(&lock->cnts); in queued_read_trylock()101 cnts = atomic_read(&lock->cnts); in queued_write_trylock()
176 #ifndef atomic_read177 #define atomic_read(v) READ_ONCE((v)->counter) macro230 c = atomic_read(v); in __atomic_add_unless()
132 WARN_ON_ONCE(atomic_read(&ioc->active_ref) <= 0); in get_io_context_active()141 WARN_ON_ONCE(atomic_read(&ioc->nr_tasks) <= 0); in ioc_task_link()
193 return atomic_read(&key->enabled); in static_key_count()247 if (atomic_read(&key->enabled) != 0) { in static_key_enable()248 WARN_ON_ONCE(atomic_read(&key->enabled) != 1); in static_key_enable()258 if (atomic_read(&key->enabled) != 1) { in static_key_disable()259 WARN_ON_ONCE(atomic_read(&key->enabled) != 0); in static_key_disable()
38 return atomic_read(&lock->tail) != OSQ_UNLOCKED_VAL; in osq_is_locked()
67 return atomic_read(&page->_refcount); in page_ref_count()72 return atomic_read(&compound_head(page)->_refcount); in page_count()
420 int old = atomic_read(index); in sbq_index_atomic_inc()436 ws = &sbq->ws[atomic_read(wait_index)]; in sbq_wait_ptr()
42 return atomic_read(&r->refs); in refcount_read()
587 return atomic_read(&page->_mapcount) < 0; in PageTransCompoundMap()591 return atomic_read(&page->_mapcount) == in PageTransCompoundMap()592 atomic_read(compound_mapcount_ptr(head)); in PageTransCompoundMap()666 return atomic_read(&page->_mapcount) == \671 VM_BUG_ON_PAGE(atomic_read(&page->_mapcount) != -1, page); \
613 return atomic_read(&mm->tlb_flush_pending); in mm_tlb_flush_pending()625 return atomic_read(&mm->tlb_flush_pending) > 1; in mm_tlb_flush_nested()
320 (raw_smp_processor_id() == atomic_read(&kgdb_active))
37 if (likely(!atomic_read(&system_freezing_cnt))) in freezing()
257 if (atomic_read(val) == 0) in wait_on_atomic_t()
102 return (void *)(unsigned long)atomic_read(&poll->event); in proc_sys_poll_event()
49 WARN_ON_ONCE(!atomic_read(&dquot->dq_count)); in dqgrab()59 if (atomic_read(&dquot->dq_count) > 1) in dquot_is_busy()
24 __entry->rmem_alloc = atomic_read(&sk->sk_rmem_alloc);52 __entry->rmem_alloc = atomic_read(&sk->sk_rmem_alloc);
38 __entry->usage_count = atomic_read(44 __entry->child_count = atomic_read(
29 atomic_read(&q->owned_by_drv_count);
84 __entry->refcnt = atomic_read(&mod->refcnt);
187 __entry->wcount = atomic_read(&inode->i_writecount);189 __entry->icount = atomic_read(&inode->i_count);
28 return (atomic_read(&cxl_use_count) != 0); in cxl_ctx_in_use()
230 return atomic_read(&net->count) != 0; in check_net()366 return atomic_read(&net->ipv4.rt_genid); in rt_genid_ipv4()398 return atomic_read(&net->fnhe_genid); in fnhe_genid()
220 return atomic_read(&queue->qlen); in reqsk_queue_len()225 return atomic_read(&queue->young); in reqsk_queue_len_young()
199 return atomic_read(&sk->sk_rmem_alloc) > (sk->sk_rcvbuf >> 1); in dn_congested()