Home
last modified time | relevance | path

Searched refs:kmp_int32 (Results 1 – 25 of 36) sorted by relevance

12

/external/llvm-project/openmp/runtime/test/tasking/
Dbug_nested_proxy_task.c24 typedef int kmp_int32; typedef
29 kmp_int32 reserved_1; /**< might be used in Fortran; see above */
30kmp_int32 flags; /**< also f.flags; KMP_IDENT_xxx flags; KMP_IDENT_KMPC identifies this un…
31 kmp_int32 reserved_2; /**< not really used in Fortran any more; see above */
36 kmp_int32 reserved_3; /**< source[4] in Fortran, do not use for C++ */
53 typedef kmp_int32 (* kmp_routine_entry_t)( kmp_int32, struct kmp_task * );
58 kmp_int32 part_id; /**< part id for the task */
64 kmp_int32 __kmpc_global_thread_num ( ident_t * );
66 __kmpc_omp_task_alloc( ident_t *loc_ref, kmp_int32 gtid, kmp_int32 flags,
70 kmp_int32 __kmpc_omp_task_with_deps ( ident_t *loc_ref, kmp_int32 gtid, kmp_task_t * new_task,
[all …]
Dbug_proxy_task_dep_waiting.c21 typedef int kmp_int32; typedef
26 kmp_int32 reserved_1; /**< might be used in Fortran; see above */
27kmp_int32 flags; /**< also f.flags; KMP_IDENT_xxx flags; KMP_IDENT_KMPC identifies this un…
28 kmp_int32 reserved_2; /**< not really used in Fortran any more; see above */
33 kmp_int32 reserved_3; /**< source[4] in Fortran, do not use for C++ */
50 typedef kmp_int32 (* kmp_routine_entry_t)( kmp_int32, struct kmp_task * );
55 kmp_int32 part_id; /**< part id for the task */
61 kmp_int32 __kmpc_global_thread_num ( ident_t * );
63 __kmpc_omp_task_alloc( ident_t *loc_ref, kmp_int32 gtid, kmp_int32 flags,
67 kmp_int32 __kmpc_omp_task_with_deps ( ident_t *loc_ref, kmp_int32 gtid, kmp_task_t * new_task,
[all …]
/external/llvm-project/openmp/runtime/src/
Dkmp_omp.h29 kmp_int32 offset;
30 kmp_int32 size;
35 kmp_int32 size;
36 kmp_int32 padding;
44 kmp_int32 begin; // Beginning of source line range.
45 kmp_int32 end; // End of source line range.
46 kmp_int32 num_threads; // Specified number of threads.
50 kmp_int32 num; // Number of items in the array.
58 kmp_int32 lib_version;
63 kmp_int32 sizeof_this_structure;
[all …]
Dkmp.h228 kmp_int32 reserved_1; /**< might be used in Fortran; see above */
229 kmp_int32 flags; /**< also f.flags; KMP_IDENT_xxx flags; KMP_IDENT_KMPC
231 kmp_int32 reserved_2; /**< not really used in Fortran any more; see above */
236 kmp_int32 reserved_3; /**< source[4] in Fortran, do not use for C++ */
242 kmp_int32 get_openmp_version() { in get_openmp_version()
349 enum sched_type : kmp_int32 {
453 (s = (enum sched_type)((kmp_int32)s | (kmp_int32)m))
1477 typedef kmp_int32 kmp_critical_name[8];
1488 typedef void (*kmpc_micro)(kmp_int32 *global_tid, kmp_int32 *bound_tid, ...);
1489 typedef void (*kmpc_micro_bound)(kmp_int32 *bound_tid, kmp_int32 *bound_nth,
[all …]
Dkmp_lock.h123 std::atomic<kmp_int32> poll;
124 kmp_int32 depth_locked; // depth locked, for nested locks only
144 extern int __kmp_acquire_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid);
145 extern int __kmp_test_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid);
146 extern int __kmp_release_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid);
150 extern int __kmp_acquire_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid);
151 extern int __kmp_test_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid);
152 extern int __kmp_release_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid);
179 volatile kmp_int32 poll; // KMP_LOCK_FREE(futex) => unlocked
182 kmp_int32 depth_locked; // depth locked, for nested locks only
[all …]
Dkmp_sched.cpp71 static void __kmp_for_static_init(ident_t *loc, kmp_int32 global_tid, in __kmp_for_static_init()
72 kmp_int32 schedtype, kmp_int32 *plastiter, in __kmp_for_static_init()
89 kmp_int32 gtid = global_tid; in __kmp_for_static_init()
428 static void __kmp_dist_for_static_init(ident_t *loc, kmp_int32 gtid, in __kmp_dist_for_static_init()
429 kmp_int32 schedule, kmp_int32 *plastiter, in __kmp_dist_for_static_init()
670 static void __kmp_team_static_init(ident_t *loc, kmp_int32 gtid, in __kmp_team_static_init()
671 kmp_int32 *p_last, T *p_lb, T *p_ub, in __kmp_team_static_init()
804 void __kmpc_for_static_init_4(ident_t *loc, kmp_int32 gtid, kmp_int32 schedtype, in __kmpc_for_static_init_4()
805 kmp_int32 *plastiter, kmp_int32 *plower, in __kmpc_for_static_init_4()
806 kmp_int32 *pupper, kmp_int32 *pstride, in __kmpc_for_static_init_4()
[all …]
Dkmp_os.h121 typedef int kmp_int32; typedef
132 kmp_int32 a, b;
157 typedef int kmp_int32; typedef
198 typedef kmp_int32 kmp_int;
201 #define KMP_INT_MAX ((kmp_int32)0x7FFFFFFF)
202 #define KMP_INT_MIN ((kmp_int32)0x80000000)
462 kmp_int32 tmp = InterlockedExchange((volatile long *)p, *(long *)&v);
470 extern kmp_int32 __kmp_test_then_add32(volatile kmp_int32 *p, kmp_int32 v);
481 extern kmp_int32 __kmp_compare_and_store32(volatile kmp_int32 *p, kmp_int32 cv,
482 kmp_int32 sv);
[all …]
Dkmp_lock.cpp72 static kmp_int32 __kmp_get_tas_lock_owner(kmp_tas_lock_t *lck) { in __kmp_get_tas_lock_owner()
81 __kmp_acquire_tas_lock_timed_template(kmp_tas_lock_t *lck, kmp_int32 gtid) { in __kmp_acquire_tas_lock_timed_template()
91 kmp_int32 tas_free = KMP_LOCK_FREE(tas); in __kmp_acquire_tas_lock_timed_template()
92 kmp_int32 tas_busy = KMP_LOCK_BUSY(gtid + 1, tas); in __kmp_acquire_tas_lock_timed_template()
113 int __kmp_acquire_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) { in __kmp_acquire_tas_lock()
120 kmp_int32 gtid) { in __kmp_acquire_tas_lock_with_checks()
132 int __kmp_test_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) { in __kmp_test_tas_lock()
133 kmp_int32 tas_free = KMP_LOCK_FREE(tas); in __kmp_test_tas_lock()
134 kmp_int32 tas_busy = KMP_LOCK_BUSY(gtid + 1, tas); in __kmp_test_tas_lock()
144 kmp_int32 gtid) { in __kmp_test_tas_lock_with_checks()
[all …]
Dkmp_atomic.h362 kmp_int32 gtid) { in __kmp_acquire_atomic_lock()
382 kmp_int32 gtid) { in __kmp_test_atomic_lock()
387 kmp_int32 gtid) { in __kmp_release_atomic_lock()
477 void __kmpc_atomic_fixed4_add(ident_t *id_ref, int gtid, kmp_int32 *lhs,
478 kmp_int32 rhs);
479 void __kmpc_atomic_fixed4_sub(ident_t *id_ref, int gtid, kmp_int32 *lhs,
480 kmp_int32 rhs);
497 void __kmpc_atomic_fixed4_andb(ident_t *id_ref, int gtid, kmp_int32 *lhs,
498 kmp_int32 rhs);
499 void __kmpc_atomic_fixed4_div(ident_t *id_ref, int gtid, kmp_int32 *lhs,
[all …]
Dkmp_dispatch.h58 dispatch_shared_info_template<T> volatile *sh, kmp_int32 *p_last, T *p_lb,
139 kmp_int32 ordered_dummy[KMP_MAX_ORDERED - 3];
143 kmp_int32 hier_id;
146 kmp_int32 get_hier_id() const { return hier_id; } in get_hier_id()
174 volatile kmp_int32 doacross_buf_idx; // teamwise index
176 kmp_int32 doacross_num_done; // count finished threads
197 __forceinline kmp_int32 test_then_add<kmp_int32>(volatile kmp_int32 *p,
198 kmp_int32 d) {
199 kmp_int32 r;
216 __forceinline kmp_int32 test_then_inc_acq<kmp_int32>(volatile kmp_int32 *p) {
[all …]
Dkmp_taskdeps.cpp33 static std::atomic<kmp_int32> kmp_node_id_seed = ATOMIC_VAR_INIT(0);
60 static inline kmp_int32 __kmp_dephash_hash(kmp_intptr_t addr, size_t hsize) { in __kmp_dephash_hash()
75 kmp_int32 size_to_allocate = in __kmp_dephash_extend()
96 kmp_int32 new_bucket = __kmp_dephash_hash(entry->addr, h->size); in __kmp_dephash_extend()
126 kmp_int32 size = in __kmp_dephash_create()
158 kmp_int32 bucket = __kmp_dephash_hash(addr, h->size); in __kmp_dephash_find()
208 static inline void __kmp_track_dependence(kmp_int32 gtid, kmp_depnode_t *source, in __kmp_track_dependence()
239 static inline kmp_int32
240 __kmp_depnode_link_successor(kmp_int32 gtid, kmp_info_t *thread, in __kmp_depnode_link_successor()
245 kmp_int32 npredecessors = 0; in __kmp_depnode_link_successor()
[all …]
Dkmp_cancel.cpp29 kmp_int32 __kmpc_cancel(ident_t *loc_ref, kmp_int32 gtid, kmp_int32 cncl_kind) { in __kmpc_cancel()
51 kmp_int32 old = cancel_noreq; in __kmpc_cancel()
89 kmp_int32 old = cancel_noreq; in __kmpc_cancel()
135 kmp_int32 __kmpc_cancellationpoint(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_cancellationpoint()
136 kmp_int32 cncl_kind) { in __kmpc_cancellationpoint()
243 kmp_int32 __kmpc_cancel_barrier(ident_t *loc, kmp_int32 gtid) { in __kmpc_cancel_barrier()
Dkmp_tasking.cpp33 static void __kmp_bottom_half_finish_proxy(kmp_int32 gtid, kmp_task_t *ptask);
44 static void __kmp_trace_task_stack(kmp_int32 gtid, in __kmp_trace_task_stack()
49 kmp_int32 entries = task_stack->ts_entries; in __kmp_trace_task_stack()
98 static void __kmp_init_task_stack(kmp_int32 gtid, in __kmp_init_task_stack()
119 static void __kmp_free_task_stack(kmp_int32 gtid, in __kmp_free_task_stack()
148 static void __kmp_push_task_stack(kmp_int32 gtid, kmp_info_t *thread, in __kmp_push_task_stack()
207 static void __kmp_pop_task_stack(kmp_int32 gtid, kmp_info_t *thread, in __kmp_pop_task_stack()
253 static bool __kmp_task_is_allowed(int gtid, const kmp_int32 is_constrained, in __kmp_task_is_allowed()
265 kmp_int32 level = current->td_level; in __kmp_task_is_allowed()
300 kmp_int32 size = TASK_DEQUE_SIZE(thread_data->td); in __kmp_realloc_task_deque()
[all …]
Dkmp_dispatch_hier.h57 kmp_int32 *small_chunks;
61 void append(enum sched_type sched, kmp_int32 chunk, kmp_hier_layer_e layer) { in append()
65 small_chunks = (kmp_int32 *)__kmp_allocate(sizeof(kmp_int32) * in append()
95 kmp_int32 temp3 = small_chunks[i]; in sort()
141 kmp_int32 status[2];
154 void set_next_hand_thread(T nlb, T nub, ST nst, kmp_int32 nstatus, in set_next_hand_thread()
161 void set_next(T nlb, T nub, ST nst, kmp_int32 nstatus, kmp_uint64 index) { in set_next()
169 kmp_int32 get_next_status(kmp_uint64 index) const { in get_next_status()
179 kmp_int32 get_curr_status(kmp_uint64 index) const { return status[index]; } in get_curr_status()
242 static void reset_private(kmp_int32 num_active,
[all …]
Dkmp_dispatch.cpp717 __kmp_dispatch_init_hier_runtime<kmp_int32>(ident_t *loc, kmp_int32 lb, in __kmp_dispatch_init_hier_runtime()
718 kmp_int32 ub, kmp_int32 st) { in __kmp_dispatch_init_hier_runtime()
719 __kmp_dispatch_init_hierarchy<kmp_int32>( in __kmp_dispatch_init_hier_runtime()
726 kmp_uint32 ub, kmp_int32 st) { in __kmp_dispatch_init_hier_runtime()
754 reinterpret_cast<dispatch_shared_info_template<kmp_int32> volatile *>( in __kmp_dispatch_free_hierarchies()
1154 kmp_int32 *p_last, T *p_lb, T *p_ub, in __kmp_dispatch_next_algorithm()
1161 kmp_int32 last = 0; in __kmp_dispatch_next_algorithm()
1332 kmp_int32 remaining; in __kmp_dispatch_next_algorithm()
1895 static int __kmp_dispatch_next(ident_t *loc, int gtid, kmp_int32 *p_last, in __kmp_dispatch_next()
1943 kmp_int32 last; in __kmp_dispatch_next()
[all …]
Dkmp_csupport.cpp36 void __kmpc_begin(ident_t *loc, kmp_int32 flags) { in __kmpc_begin()
98 kmp_int32 __kmpc_global_thread_num(ident_t *loc) { in __kmpc_global_thread_num()
99 kmp_int32 gtid = __kmp_entry_gtid(); in __kmpc_global_thread_num()
120 kmp_int32 __kmpc_global_num_threads(ident_t *loc) { in __kmpc_global_num_threads()
133 kmp_int32 __kmpc_bound_thread_num(ident_t *loc) { in __kmpc_bound_thread_num()
143 kmp_int32 __kmpc_bound_num_threads(ident_t *loc) { in __kmpc_bound_num_threads()
155 kmp_int32 __kmpc_ok_to_fork(ident_t *loc) { in __kmpc_ok_to_fork()
217 kmp_int32 __kmpc_in_parallel(ident_t *loc) { in __kmpc_in_parallel()
230 void __kmpc_push_num_threads(ident_t *loc, kmp_int32 global_tid, in __kmpc_push_num_threads()
231 kmp_int32 num_threads) { in __kmpc_push_num_threads()
[all …]
Dz_Windows_NT-586_asm.asm113 ; kmp_int32
114 ; __kmp_test_then_add32( volatile kmp_int32 *p, kmp_int32 d );
183 ; kmp_int32
184 ; __kmp_compare_and_store32( volatile kmp_int32 *p, kmp_int32 cv, kmp_int32 sv );
208 ; kmp_int32
287 ; kmp_int32
288 ; __kmp_xchg_fixed32( volatile kmp_int32 *p, kmp_int32 d );
395 ; kmp_int32
396 ; __kmp_compare_and_store_ret32( volatile kmp_int32 *p, kmp_int32 cv, kmp_int32 sv );
669 ; kmp_int32
[all …]
Dkmp_atomic.cpp914 ATOMIC_FIXED_ADD(fixed4, add, kmp_int32, 32, +, 4i, 3,
916 ATOMIC_FIXED_ADD(fixed4, sub, kmp_int32, 32, -, 4i, 3,
994 ATOMIC_CMPXCHG(fixed4, andb, kmp_int32, 32, &, 4i, 3,
996 ATOMIC_CMPXCHG(fixed4, div, kmp_int32, 32, /, 4i, 3,
1000 ATOMIC_CMPXCHG(fixed4, mul, kmp_int32, 32, *, 4i, 3,
1002 ATOMIC_CMPXCHG(fixed4, orb, kmp_int32, 32, |, 4i, 3,
1004 ATOMIC_CMPXCHG(fixed4, shl, kmp_int32, 32, <<, 4i, 3,
1006 ATOMIC_CMPXCHG(fixed4, shr, kmp_int32, 32, >>, 4i, 3,
1010 ATOMIC_CMPXCHG(fixed4, xor, kmp_int32, 32, ^, 4i, 3,
1085 ATOMIC_CMPX_L(fixed4, andl, kmp_int32, 32, &&, 4i, 3,
[all …]
Dkmp_itt.cpp27 kmp_int32 __kmp_barrier_domain_count;
28 kmp_int32 __kmp_region_domain_count;
32 kmp_int32 __kmp_itt_region_team_size[KMP_MAX_FRAME_DOMAINS];
52 kmp_int32 __kmp_itt_prepare_delay = 0;
Dkmp_global.cpp62 std::atomic<kmp_int32> __kmp_team_counter = ATOMIC_VAR_INIT(0);
63 std::atomic<kmp_int32> __kmp_task_counter = ATOMIC_VAR_INIT(0);
291 kmp_int32 __kmp_default_device = 0;
294 kmp_int32 __kmp_max_task_priority = 0;
411 kmp_int32 __kmp_use_yield = 1;
413 kmp_int32 __kmp_use_yield_exp_set = 0;
Dkmp_itt.h30 extern kmp_int32 __kmp_itt_prepare_delay;
274 extern kmp_int32 __kmp_barrier_domain_count;
275 extern kmp_int32 __kmp_region_domain_count;
279 extern kmp_int32 __kmp_itt_region_team_size[KMP_MAX_FRAME_DOMAINS];
Dkmp_io.cpp153 (kmp_int32)getpid()); in __kmp_vprintf()
184 __kmp_str_buf_print(&__kmp_console_buf, "pid=%d: ", (kmp_int32)getpid()); in __kmp_vprintf()
192 fprintf(stream, "pid=%d: ", (kmp_int32)getpid()); in __kmp_vprintf()
Dkmp_taskdeps.h27 kmp_int32 n = KMP_ATOMIC_DEC(&node->dn.nrefs) - 1; in __kmp_node_deref()
88 static inline void __kmp_release_deps(kmp_int32 gtid, kmp_taskdata_t *task) { in __kmp_release_deps()
124 kmp_int32 npredecessors = KMP_ATOMIC_DEC(&successor->dn.npredecessors) - 1; in __kmp_release_deps()
Dkmp_gsupport.cpp30 kmp_int32 num_deps;
63 num_deps = static_cast<kmp_int32>(ndeps); in kmp_gomp_depends_info_t()
65 kmp_int32 get_num_deps() const { return num_deps; } in get_num_deps()
193 kmp_int32 rc = __kmp_enter_single(gtid, &loc, FALSE); in KMP_EXPAND_NAME()
1193 kmp_int32 flags = 0;
1244 kmp_int32 ndeps = gomp_depends.get_num_deps();
1246 for (kmp_int32 i = 0; i < ndeps; i++)
1271 kmp_int32 ndeps = gomp_depends.get_num_deps();
1273 for (kmp_int32 i = 0; i < ndeps; i++)
1585 static kmp_int32 __kmp_gomp_to_omp_cancellation_kind(int gomp_kind) { in __kmp_gomp_to_omp_cancellation_kind()
[all …]
Dz_Windows_NT-586_util.cpp53 while (!__kmp_compare_and_store32((volatile kmp_int32 *)p, old_value, in __kmp_test_then_or32()
68 while (!__kmp_compare_and_store32((volatile kmp_int32 *)p, old_value, in __kmp_test_then_and32()

12