/external/llvm-project/openmp/runtime/test/tasking/ |
D | bug_nested_proxy_task.c | 24 typedef int kmp_int32; typedef 29 kmp_int32 reserved_1; /**< might be used in Fortran; see above */ 30 …kmp_int32 flags; /**< also f.flags; KMP_IDENT_xxx flags; KMP_IDENT_KMPC identifies this un… 31 kmp_int32 reserved_2; /**< not really used in Fortran any more; see above */ 36 kmp_int32 reserved_3; /**< source[4] in Fortran, do not use for C++ */ 53 typedef kmp_int32 (* kmp_routine_entry_t)( kmp_int32, struct kmp_task * ); 58 kmp_int32 part_id; /**< part id for the task */ 64 kmp_int32 __kmpc_global_thread_num ( ident_t * ); 66 __kmpc_omp_task_alloc( ident_t *loc_ref, kmp_int32 gtid, kmp_int32 flags, 70 kmp_int32 __kmpc_omp_task_with_deps ( ident_t *loc_ref, kmp_int32 gtid, kmp_task_t * new_task, [all …]
|
D | bug_proxy_task_dep_waiting.c | 21 typedef int kmp_int32; typedef 26 kmp_int32 reserved_1; /**< might be used in Fortran; see above */ 27 …kmp_int32 flags; /**< also f.flags; KMP_IDENT_xxx flags; KMP_IDENT_KMPC identifies this un… 28 kmp_int32 reserved_2; /**< not really used in Fortran any more; see above */ 33 kmp_int32 reserved_3; /**< source[4] in Fortran, do not use for C++ */ 50 typedef kmp_int32 (* kmp_routine_entry_t)( kmp_int32, struct kmp_task * ); 55 kmp_int32 part_id; /**< part id for the task */ 61 kmp_int32 __kmpc_global_thread_num ( ident_t * ); 63 __kmpc_omp_task_alloc( ident_t *loc_ref, kmp_int32 gtid, kmp_int32 flags, 67 kmp_int32 __kmpc_omp_task_with_deps ( ident_t *loc_ref, kmp_int32 gtid, kmp_task_t * new_task, [all …]
|
/external/llvm-project/openmp/runtime/src/ |
D | kmp_omp.h | 29 kmp_int32 offset; 30 kmp_int32 size; 35 kmp_int32 size; 36 kmp_int32 padding; 44 kmp_int32 begin; // Beginning of source line range. 45 kmp_int32 end; // End of source line range. 46 kmp_int32 num_threads; // Specified number of threads. 50 kmp_int32 num; // Number of items in the array. 58 kmp_int32 lib_version; 63 kmp_int32 sizeof_this_structure; [all …]
|
D | kmp.h | 228 kmp_int32 reserved_1; /**< might be used in Fortran; see above */ 229 kmp_int32 flags; /**< also f.flags; KMP_IDENT_xxx flags; KMP_IDENT_KMPC 231 kmp_int32 reserved_2; /**< not really used in Fortran any more; see above */ 236 kmp_int32 reserved_3; /**< source[4] in Fortran, do not use for C++ */ 242 kmp_int32 get_openmp_version() { in get_openmp_version() 349 enum sched_type : kmp_int32 { 453 (s = (enum sched_type)((kmp_int32)s | (kmp_int32)m)) 1477 typedef kmp_int32 kmp_critical_name[8]; 1488 typedef void (*kmpc_micro)(kmp_int32 *global_tid, kmp_int32 *bound_tid, ...); 1489 typedef void (*kmpc_micro_bound)(kmp_int32 *bound_tid, kmp_int32 *bound_nth, [all …]
|
D | kmp_lock.h | 123 std::atomic<kmp_int32> poll; 124 kmp_int32 depth_locked; // depth locked, for nested locks only 144 extern int __kmp_acquire_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 145 extern int __kmp_test_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 146 extern int __kmp_release_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 150 extern int __kmp_acquire_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 151 extern int __kmp_test_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 152 extern int __kmp_release_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 179 volatile kmp_int32 poll; // KMP_LOCK_FREE(futex) => unlocked 182 kmp_int32 depth_locked; // depth locked, for nested locks only [all …]
|
D | kmp_sched.cpp | 71 static void __kmp_for_static_init(ident_t *loc, kmp_int32 global_tid, in __kmp_for_static_init() 72 kmp_int32 schedtype, kmp_int32 *plastiter, in __kmp_for_static_init() 89 kmp_int32 gtid = global_tid; in __kmp_for_static_init() 428 static void __kmp_dist_for_static_init(ident_t *loc, kmp_int32 gtid, in __kmp_dist_for_static_init() 429 kmp_int32 schedule, kmp_int32 *plastiter, in __kmp_dist_for_static_init() 670 static void __kmp_team_static_init(ident_t *loc, kmp_int32 gtid, in __kmp_team_static_init() 671 kmp_int32 *p_last, T *p_lb, T *p_ub, in __kmp_team_static_init() 804 void __kmpc_for_static_init_4(ident_t *loc, kmp_int32 gtid, kmp_int32 schedtype, in __kmpc_for_static_init_4() 805 kmp_int32 *plastiter, kmp_int32 *plower, in __kmpc_for_static_init_4() 806 kmp_int32 *pupper, kmp_int32 *pstride, in __kmpc_for_static_init_4() [all …]
|
D | kmp_os.h | 121 typedef int kmp_int32; typedef 132 kmp_int32 a, b; 157 typedef int kmp_int32; typedef 198 typedef kmp_int32 kmp_int; 201 #define KMP_INT_MAX ((kmp_int32)0x7FFFFFFF) 202 #define KMP_INT_MIN ((kmp_int32)0x80000000) 462 kmp_int32 tmp = InterlockedExchange((volatile long *)p, *(long *)&v); 470 extern kmp_int32 __kmp_test_then_add32(volatile kmp_int32 *p, kmp_int32 v); 481 extern kmp_int32 __kmp_compare_and_store32(volatile kmp_int32 *p, kmp_int32 cv, 482 kmp_int32 sv); [all …]
|
D | kmp_lock.cpp | 72 static kmp_int32 __kmp_get_tas_lock_owner(kmp_tas_lock_t *lck) { in __kmp_get_tas_lock_owner() 81 __kmp_acquire_tas_lock_timed_template(kmp_tas_lock_t *lck, kmp_int32 gtid) { in __kmp_acquire_tas_lock_timed_template() 91 kmp_int32 tas_free = KMP_LOCK_FREE(tas); in __kmp_acquire_tas_lock_timed_template() 92 kmp_int32 tas_busy = KMP_LOCK_BUSY(gtid + 1, tas); in __kmp_acquire_tas_lock_timed_template() 113 int __kmp_acquire_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) { in __kmp_acquire_tas_lock() 120 kmp_int32 gtid) { in __kmp_acquire_tas_lock_with_checks() 132 int __kmp_test_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) { in __kmp_test_tas_lock() 133 kmp_int32 tas_free = KMP_LOCK_FREE(tas); in __kmp_test_tas_lock() 134 kmp_int32 tas_busy = KMP_LOCK_BUSY(gtid + 1, tas); in __kmp_test_tas_lock() 144 kmp_int32 gtid) { in __kmp_test_tas_lock_with_checks() [all …]
|
D | kmp_atomic.h | 362 kmp_int32 gtid) { in __kmp_acquire_atomic_lock() 382 kmp_int32 gtid) { in __kmp_test_atomic_lock() 387 kmp_int32 gtid) { in __kmp_release_atomic_lock() 477 void __kmpc_atomic_fixed4_add(ident_t *id_ref, int gtid, kmp_int32 *lhs, 478 kmp_int32 rhs); 479 void __kmpc_atomic_fixed4_sub(ident_t *id_ref, int gtid, kmp_int32 *lhs, 480 kmp_int32 rhs); 497 void __kmpc_atomic_fixed4_andb(ident_t *id_ref, int gtid, kmp_int32 *lhs, 498 kmp_int32 rhs); 499 void __kmpc_atomic_fixed4_div(ident_t *id_ref, int gtid, kmp_int32 *lhs, [all …]
|
D | kmp_dispatch.h | 58 dispatch_shared_info_template<T> volatile *sh, kmp_int32 *p_last, T *p_lb, 139 kmp_int32 ordered_dummy[KMP_MAX_ORDERED - 3]; 143 kmp_int32 hier_id; 146 kmp_int32 get_hier_id() const { return hier_id; } in get_hier_id() 174 volatile kmp_int32 doacross_buf_idx; // teamwise index 176 kmp_int32 doacross_num_done; // count finished threads 197 __forceinline kmp_int32 test_then_add<kmp_int32>(volatile kmp_int32 *p, 198 kmp_int32 d) { 199 kmp_int32 r; 216 __forceinline kmp_int32 test_then_inc_acq<kmp_int32>(volatile kmp_int32 *p) { [all …]
|
D | kmp_taskdeps.cpp | 33 static std::atomic<kmp_int32> kmp_node_id_seed = ATOMIC_VAR_INIT(0); 60 static inline kmp_int32 __kmp_dephash_hash(kmp_intptr_t addr, size_t hsize) { in __kmp_dephash_hash() 75 kmp_int32 size_to_allocate = in __kmp_dephash_extend() 96 kmp_int32 new_bucket = __kmp_dephash_hash(entry->addr, h->size); in __kmp_dephash_extend() 126 kmp_int32 size = in __kmp_dephash_create() 158 kmp_int32 bucket = __kmp_dephash_hash(addr, h->size); in __kmp_dephash_find() 208 static inline void __kmp_track_dependence(kmp_int32 gtid, kmp_depnode_t *source, in __kmp_track_dependence() 239 static inline kmp_int32 240 __kmp_depnode_link_successor(kmp_int32 gtid, kmp_info_t *thread, in __kmp_depnode_link_successor() 245 kmp_int32 npredecessors = 0; in __kmp_depnode_link_successor() [all …]
|
D | kmp_cancel.cpp | 29 kmp_int32 __kmpc_cancel(ident_t *loc_ref, kmp_int32 gtid, kmp_int32 cncl_kind) { in __kmpc_cancel() 51 kmp_int32 old = cancel_noreq; in __kmpc_cancel() 89 kmp_int32 old = cancel_noreq; in __kmpc_cancel() 135 kmp_int32 __kmpc_cancellationpoint(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_cancellationpoint() 136 kmp_int32 cncl_kind) { in __kmpc_cancellationpoint() 243 kmp_int32 __kmpc_cancel_barrier(ident_t *loc, kmp_int32 gtid) { in __kmpc_cancel_barrier()
|
D | kmp_tasking.cpp | 33 static void __kmp_bottom_half_finish_proxy(kmp_int32 gtid, kmp_task_t *ptask); 44 static void __kmp_trace_task_stack(kmp_int32 gtid, in __kmp_trace_task_stack() 49 kmp_int32 entries = task_stack->ts_entries; in __kmp_trace_task_stack() 98 static void __kmp_init_task_stack(kmp_int32 gtid, in __kmp_init_task_stack() 119 static void __kmp_free_task_stack(kmp_int32 gtid, in __kmp_free_task_stack() 148 static void __kmp_push_task_stack(kmp_int32 gtid, kmp_info_t *thread, in __kmp_push_task_stack() 207 static void __kmp_pop_task_stack(kmp_int32 gtid, kmp_info_t *thread, in __kmp_pop_task_stack() 253 static bool __kmp_task_is_allowed(int gtid, const kmp_int32 is_constrained, in __kmp_task_is_allowed() 265 kmp_int32 level = current->td_level; in __kmp_task_is_allowed() 300 kmp_int32 size = TASK_DEQUE_SIZE(thread_data->td); in __kmp_realloc_task_deque() [all …]
|
D | kmp_dispatch_hier.h | 57 kmp_int32 *small_chunks; 61 void append(enum sched_type sched, kmp_int32 chunk, kmp_hier_layer_e layer) { in append() 65 small_chunks = (kmp_int32 *)__kmp_allocate(sizeof(kmp_int32) * in append() 95 kmp_int32 temp3 = small_chunks[i]; in sort() 141 kmp_int32 status[2]; 154 void set_next_hand_thread(T nlb, T nub, ST nst, kmp_int32 nstatus, in set_next_hand_thread() 161 void set_next(T nlb, T nub, ST nst, kmp_int32 nstatus, kmp_uint64 index) { in set_next() 169 kmp_int32 get_next_status(kmp_uint64 index) const { in get_next_status() 179 kmp_int32 get_curr_status(kmp_uint64 index) const { return status[index]; } in get_curr_status() 242 static void reset_private(kmp_int32 num_active, [all …]
|
D | kmp_dispatch.cpp | 717 __kmp_dispatch_init_hier_runtime<kmp_int32>(ident_t *loc, kmp_int32 lb, in __kmp_dispatch_init_hier_runtime() 718 kmp_int32 ub, kmp_int32 st) { in __kmp_dispatch_init_hier_runtime() 719 __kmp_dispatch_init_hierarchy<kmp_int32>( in __kmp_dispatch_init_hier_runtime() 726 kmp_uint32 ub, kmp_int32 st) { in __kmp_dispatch_init_hier_runtime() 754 reinterpret_cast<dispatch_shared_info_template<kmp_int32> volatile *>( in __kmp_dispatch_free_hierarchies() 1154 kmp_int32 *p_last, T *p_lb, T *p_ub, in __kmp_dispatch_next_algorithm() 1161 kmp_int32 last = 0; in __kmp_dispatch_next_algorithm() 1332 kmp_int32 remaining; in __kmp_dispatch_next_algorithm() 1895 static int __kmp_dispatch_next(ident_t *loc, int gtid, kmp_int32 *p_last, in __kmp_dispatch_next() 1943 kmp_int32 last; in __kmp_dispatch_next() [all …]
|
D | kmp_csupport.cpp | 36 void __kmpc_begin(ident_t *loc, kmp_int32 flags) { in __kmpc_begin() 98 kmp_int32 __kmpc_global_thread_num(ident_t *loc) { in __kmpc_global_thread_num() 99 kmp_int32 gtid = __kmp_entry_gtid(); in __kmpc_global_thread_num() 120 kmp_int32 __kmpc_global_num_threads(ident_t *loc) { in __kmpc_global_num_threads() 133 kmp_int32 __kmpc_bound_thread_num(ident_t *loc) { in __kmpc_bound_thread_num() 143 kmp_int32 __kmpc_bound_num_threads(ident_t *loc) { in __kmpc_bound_num_threads() 155 kmp_int32 __kmpc_ok_to_fork(ident_t *loc) { in __kmpc_ok_to_fork() 217 kmp_int32 __kmpc_in_parallel(ident_t *loc) { in __kmpc_in_parallel() 230 void __kmpc_push_num_threads(ident_t *loc, kmp_int32 global_tid, in __kmpc_push_num_threads() 231 kmp_int32 num_threads) { in __kmpc_push_num_threads() [all …]
|
D | z_Windows_NT-586_asm.asm | 113 ; kmp_int32 114 ; __kmp_test_then_add32( volatile kmp_int32 *p, kmp_int32 d ); 183 ; kmp_int32 184 ; __kmp_compare_and_store32( volatile kmp_int32 *p, kmp_int32 cv, kmp_int32 sv ); 208 ; kmp_int32 287 ; kmp_int32 288 ; __kmp_xchg_fixed32( volatile kmp_int32 *p, kmp_int32 d ); 395 ; kmp_int32 396 ; __kmp_compare_and_store_ret32( volatile kmp_int32 *p, kmp_int32 cv, kmp_int32 sv ); 669 ; kmp_int32 [all …]
|
D | kmp_atomic.cpp | 914 ATOMIC_FIXED_ADD(fixed4, add, kmp_int32, 32, +, 4i, 3, 916 ATOMIC_FIXED_ADD(fixed4, sub, kmp_int32, 32, -, 4i, 3, 994 ATOMIC_CMPXCHG(fixed4, andb, kmp_int32, 32, &, 4i, 3, 996 ATOMIC_CMPXCHG(fixed4, div, kmp_int32, 32, /, 4i, 3, 1000 ATOMIC_CMPXCHG(fixed4, mul, kmp_int32, 32, *, 4i, 3, 1002 ATOMIC_CMPXCHG(fixed4, orb, kmp_int32, 32, |, 4i, 3, 1004 ATOMIC_CMPXCHG(fixed4, shl, kmp_int32, 32, <<, 4i, 3, 1006 ATOMIC_CMPXCHG(fixed4, shr, kmp_int32, 32, >>, 4i, 3, 1010 ATOMIC_CMPXCHG(fixed4, xor, kmp_int32, 32, ^, 4i, 3, 1085 ATOMIC_CMPX_L(fixed4, andl, kmp_int32, 32, &&, 4i, 3, [all …]
|
D | kmp_itt.cpp | 27 kmp_int32 __kmp_barrier_domain_count; 28 kmp_int32 __kmp_region_domain_count; 32 kmp_int32 __kmp_itt_region_team_size[KMP_MAX_FRAME_DOMAINS]; 52 kmp_int32 __kmp_itt_prepare_delay = 0;
|
D | kmp_global.cpp | 62 std::atomic<kmp_int32> __kmp_team_counter = ATOMIC_VAR_INIT(0); 63 std::atomic<kmp_int32> __kmp_task_counter = ATOMIC_VAR_INIT(0); 291 kmp_int32 __kmp_default_device = 0; 294 kmp_int32 __kmp_max_task_priority = 0; 411 kmp_int32 __kmp_use_yield = 1; 413 kmp_int32 __kmp_use_yield_exp_set = 0;
|
D | kmp_itt.h | 30 extern kmp_int32 __kmp_itt_prepare_delay; 274 extern kmp_int32 __kmp_barrier_domain_count; 275 extern kmp_int32 __kmp_region_domain_count; 279 extern kmp_int32 __kmp_itt_region_team_size[KMP_MAX_FRAME_DOMAINS];
|
D | kmp_io.cpp | 153 (kmp_int32)getpid()); in __kmp_vprintf() 184 __kmp_str_buf_print(&__kmp_console_buf, "pid=%d: ", (kmp_int32)getpid()); in __kmp_vprintf() 192 fprintf(stream, "pid=%d: ", (kmp_int32)getpid()); in __kmp_vprintf()
|
D | kmp_taskdeps.h | 27 kmp_int32 n = KMP_ATOMIC_DEC(&node->dn.nrefs) - 1; in __kmp_node_deref() 88 static inline void __kmp_release_deps(kmp_int32 gtid, kmp_taskdata_t *task) { in __kmp_release_deps() 124 kmp_int32 npredecessors = KMP_ATOMIC_DEC(&successor->dn.npredecessors) - 1; in __kmp_release_deps()
|
D | kmp_gsupport.cpp | 30 kmp_int32 num_deps; 63 num_deps = static_cast<kmp_int32>(ndeps); in kmp_gomp_depends_info_t() 65 kmp_int32 get_num_deps() const { return num_deps; } in get_num_deps() 193 kmp_int32 rc = __kmp_enter_single(gtid, &loc, FALSE); in KMP_EXPAND_NAME() 1193 kmp_int32 flags = 0; 1244 kmp_int32 ndeps = gomp_depends.get_num_deps(); 1246 for (kmp_int32 i = 0; i < ndeps; i++) 1271 kmp_int32 ndeps = gomp_depends.get_num_deps(); 1273 for (kmp_int32 i = 0; i < ndeps; i++) 1585 static kmp_int32 __kmp_gomp_to_omp_cancellation_kind(int gomp_kind) { in __kmp_gomp_to_omp_cancellation_kind() [all …]
|
D | z_Windows_NT-586_util.cpp | 53 while (!__kmp_compare_and_store32((volatile kmp_int32 *)p, old_value, in __kmp_test_then_or32() 68 while (!__kmp_compare_and_store32((volatile kmp_int32 *)p, old_value, in __kmp_test_then_and32()
|