Lines Matching refs:depth_locked
77 return lck->lk.depth_locked != -1; in __kmp_is_tas_lock_nestable()
207 lck->lk.depth_locked += 1; in __kmp_acquire_nested_tas_lock()
212 lck->lk.depth_locked = 1; in __kmp_acquire_nested_tas_lock()
232 retval = ++lck->lk.depth_locked; in __kmp_test_nested_tas_lock()
237 retval = lck->lk.depth_locked = 1; in __kmp_test_nested_tas_lock()
255 if (--(lck->lk.depth_locked) == 0) { in __kmp_release_nested_tas_lock()
280 lck->lk.depth_locked = 0; // >= 0 for nestable locks, -1 for simple locks in __kmp_init_nested_tas_lock()
285 lck->lk.depth_locked = 0; in __kmp_destroy_nested_tas_lock()
314 return lck->lk.depth_locked != -1; in __kmp_is_futex_lock_nestable()
511 lck->lk.depth_locked += 1; in __kmp_acquire_nested_futex_lock()
516 lck->lk.depth_locked = 1; in __kmp_acquire_nested_futex_lock()
536 retval = ++lck->lk.depth_locked; in __kmp_test_nested_futex_lock()
541 retval = lck->lk.depth_locked = 1; in __kmp_test_nested_futex_lock()
559 if (--(lck->lk.depth_locked) == 0) { in __kmp_release_nested_futex_lock()
584 lck->lk.depth_locked = 0; // >= 0 for nestable locks, -1 for simple locks in __kmp_init_nested_futex_lock()
589 lck->lk.depth_locked = 0; in __kmp_destroy_nested_futex_lock()
615 return std::atomic_load_explicit(&lck->lk.depth_locked, in __kmp_is_ticket_lock_nestable()
767 &lck->lk.depth_locked, -1, in __kmp_init_ticket_lock()
783 std::atomic_store_explicit(&lck->lk.depth_locked, -1, in __kmp_destroy_ticket_lock()
812 std::atomic_fetch_add_explicit(&lck->lk.depth_locked, 1, in __kmp_acquire_nested_ticket_lock()
818 std::atomic_store_explicit(&lck->lk.depth_locked, 1, in __kmp_acquire_nested_ticket_lock()
849 retval = std::atomic_fetch_add_explicit(&lck->lk.depth_locked, 1, in __kmp_test_nested_ticket_lock()
855 std::atomic_store_explicit(&lck->lk.depth_locked, 1, in __kmp_test_nested_ticket_lock()
884 if ((std::atomic_fetch_add_explicit(&lck->lk.depth_locked, -1, in __kmp_release_nested_ticket_lock()
919 std::atomic_store_explicit(&lck->lk.depth_locked, 0, in __kmp_init_nested_ticket_lock()
926 std::atomic_store_explicit(&lck->lk.depth_locked, 0, in __kmp_destroy_nested_ticket_lock()
1080 return lck->lk.depth_locked != -1; in __kmp_is_queuing_lock_nestable()
1532 lck->lk.depth_locked = -1; // >= 0 for nestable locks, -1 for simple locks. in __kmp_init_queuing_lock()
1546 lck->lk.depth_locked = -1; in __kmp_destroy_queuing_lock()
1569 lck->lk.depth_locked += 1; in __kmp_acquire_nested_queuing_lock()
1575 lck->lk.depth_locked = 1; in __kmp_acquire_nested_queuing_lock()
1601 retval = ++lck->lk.depth_locked; in __kmp_test_nested_queuing_lock()
1606 retval = lck->lk.depth_locked = 1; in __kmp_test_nested_queuing_lock()
1629 if (--(lck->lk.depth_locked) == 0) { in __kmp_release_nested_queuing_lock()
1660 lck->lk.depth_locked = 0; // >= 0 for nestable locks, -1 for simple locks in __kmp_init_nested_queuing_lock()
1665 lck->lk.depth_locked = 0; in __kmp_destroy_nested_queuing_lock()
2232 return lck->lk.depth_locked != -1; in __kmp_is_drdpa_lock_nestable()
2484 lck->lk.depth_locked = -1; // >= 0 for nestable locks, -1 for simple locks. in __kmp_init_drdpa_lock()
2507 lck->lk.depth_locked = -1; in __kmp_destroy_drdpa_lock()
2530 lck->lk.depth_locked += 1; in __kmp_acquire_nested_drdpa_lock()
2536 lck->lk.depth_locked = 1; in __kmp_acquire_nested_drdpa_lock()
2561 retval = ++lck->lk.depth_locked; in __kmp_test_nested_drdpa_lock()
2566 retval = lck->lk.depth_locked = 1; in __kmp_test_nested_drdpa_lock()
2589 if (--(lck->lk.depth_locked) == 0) { in __kmp_release_nested_drdpa_lock()
2619 lck->lk.depth_locked = 0; // >= 0 for nestable locks, -1 for simple locks in __kmp_init_nested_drdpa_lock()
2624 lck->lk.depth_locked = 0; in __kmp_destroy_nested_drdpa_lock()