/external/angle/src/third_party/volk/ |
D | volk.c | 30 static void volkGenLoadLoader(void* context, PFN_vkVoidFunction (*load)(void*, const char*)); 31 static void volkGenLoadInstance(void* context, PFN_vkVoidFunction (*load)(void*, const char*)); 32 static void volkGenLoadDevice(void* context, PFN_vkVoidFunction (*load)(void*, const char*)); 33 …eTable(struct VolkDeviceTable* table, void* context, PFN_vkVoidFunction (*load)(void*, const char*… 134 static void volkGenLoadLoader(void* context, PFN_vkVoidFunction (*load)(void*, const char*)) in volkGenLoadLoader() 138 vkCreateInstance = (PFN_vkCreateInstance)load(context, "vkCreateInstance"); in volkGenLoadLoader() 139 …vkEnumerateInstanceExtensionProperties = (PFN_vkEnumerateInstanceExtensionProperties)load(context,… in volkGenLoadLoader() 140 …vkEnumerateInstanceLayerProperties = (PFN_vkEnumerateInstanceLayerProperties)load(context, "vkEnum… in volkGenLoadLoader() 143 …vkEnumerateInstanceVersion = (PFN_vkEnumerateInstanceVersion)load(context, "vkEnumerateInstanceVer… in volkGenLoadLoader() 148 static void volkGenLoadInstance(void* context, PFN_vkVoidFunction (*load)(void*, const char*)) in volkGenLoadInstance() [all …]
|
/external/llvm/test/CodeGen/Mips/ |
D | ra-allocatable.ll | 101 %0 = load i32, i32* @a0, align 4 102 %1 = load i32*, i32** @b0, align 4 104 %2 = load i32, i32* @a1, align 4 105 %3 = load i32*, i32** @b1, align 4 107 %4 = load i32, i32* @a2, align 4 108 %5 = load i32*, i32** @b2, align 4 110 %6 = load i32, i32* @a3, align 4 111 %7 = load i32*, i32** @b3, align 4 113 %8 = load i32, i32* @a4, align 4 114 %9 = load i32*, i32** @b4, align 4 [all …]
|
/external/gemmlowp/meta/generators/ |
D | transform_kernels_common.py | 63 load = [registers.QuadRegister() for unused_i in range(register_count)] 64 emitter.EmitVLoadAE(8, elements, load, input_address, None) 67 for register in load: 70 for register in load: 74 emitter.EmitVStoreAE(8, elements, load, output_address, None) 76 registers.FreeRegisters(load) 109 load = [registers.QuadRegister() for unused_i in range(register_count)] 110 emitter.EmitVLoadAE(8, elements, load, input_address, None) 113 if len(load) is 1: 114 emitter.EmitVMovl('u8', load[0], load[0]) [all …]
|
/external/rust/crates/async-task/tests/ |
D | waker_pending.rs | 108 assert_eq!(POLL.load(Ordering::SeqCst), 2); in wake_during_run() 109 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 2); in wake_during_run() 110 assert_eq!(DROP_F.load(Ordering::SeqCst), 0); in wake_during_run() 111 assert_eq!(DROP_S.load(Ordering::SeqCst), 0); in wake_during_run() 118 assert_eq!(POLL.load(Ordering::SeqCst), 2); in wake_during_run() 119 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 1); in wake_during_run() 120 assert_eq!(DROP_F.load(Ordering::SeqCst), 0); in wake_during_run() 121 assert_eq!(DROP_S.load(Ordering::SeqCst), 0); in wake_during_run() 126 assert_eq!(POLL.load(Ordering::SeqCst), 2); in wake_during_run() 127 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 2); in wake_during_run() [all …]
|
D | join.rs | 103 assert_eq!(DROP_T.load(Ordering::SeqCst), 0); in drop_and_join() 106 assert_eq!(DROP_T.load(Ordering::SeqCst), 0); in drop_and_join() 109 assert_eq!(POLL.load(Ordering::SeqCst), 0); in drop_and_join() 110 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 0); in drop_and_join() 111 assert_eq!(DROP_F.load(Ordering::SeqCst), 1); in drop_and_join() 112 assert_eq!(DROP_S.load(Ordering::SeqCst), 1); in drop_and_join() 113 assert_eq!(DROP_T.load(Ordering::SeqCst), 0); in drop_and_join() 122 assert_eq!(DROP_T.load(Ordering::SeqCst), 0); in run_and_join() 125 assert_eq!(DROP_T.load(Ordering::SeqCst), 0); in run_and_join() 128 assert_eq!(POLL.load(Ordering::SeqCst), 1); in run_and_join() [all …]
|
D | basic.rs | 84 assert_eq!(POLL.load(Ordering::SeqCst), 0); in drop_and_detach() 85 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 0); in drop_and_detach() 86 assert_eq!(DROP_F.load(Ordering::SeqCst), 0); in drop_and_detach() 87 assert_eq!(DROP_S.load(Ordering::SeqCst), 0); in drop_and_detach() 90 assert_eq!(POLL.load(Ordering::SeqCst), 0); in drop_and_detach() 91 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 0); in drop_and_detach() 92 assert_eq!(DROP_F.load(Ordering::SeqCst), 1); in drop_and_detach() 93 assert_eq!(DROP_S.load(Ordering::SeqCst), 0); in drop_and_detach() 96 assert_eq!(POLL.load(Ordering::SeqCst), 0); in drop_and_detach() 97 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 0); in drop_and_detach() [all …]
|
D | waker_ready.rs | 110 assert_eq!(POLL.load(Ordering::SeqCst), 1); in wake() 111 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 0); in wake() 112 assert_eq!(DROP_F.load(Ordering::SeqCst), 0); in wake() 113 assert_eq!(DROP_S.load(Ordering::SeqCst), 0); in wake() 118 assert_eq!(POLL.load(Ordering::SeqCst), 1); in wake() 119 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 1); in wake() 120 assert_eq!(DROP_F.load(Ordering::SeqCst), 0); in wake() 121 assert_eq!(DROP_S.load(Ordering::SeqCst), 0); in wake() 125 assert_eq!(POLL.load(Ordering::SeqCst), 2); in wake() 126 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 1); in wake() [all …]
|
D | waker_panic.rs | 122 assert_eq!(POLL.load(Ordering::SeqCst), 2); in wake_during_run() 123 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 1); in wake_during_run() 124 assert_eq!(DROP_F.load(Ordering::SeqCst), 1); in wake_during_run() 125 assert_eq!(DROP_S.load(Ordering::SeqCst), 1); in wake_during_run() 133 assert_eq!(POLL.load(Ordering::SeqCst), 2); in wake_during_run() 134 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 1); in wake_during_run() 135 assert_eq!(DROP_F.load(Ordering::SeqCst), 0); in wake_during_run() 136 assert_eq!(DROP_S.load(Ordering::SeqCst), 0); in wake_during_run() 141 assert_eq!(POLL.load(Ordering::SeqCst), 2); in wake_during_run() 142 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 1); in wake_during_run() [all …]
|
D | panic.rs | 92 assert_eq!(POLL.load(Ordering::SeqCst), 1); in cancel_during_run() 93 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 0); in cancel_during_run() 94 assert_eq!(DROP_F.load(Ordering::SeqCst), 1); in cancel_during_run() 95 assert_eq!(DROP_S.load(Ordering::SeqCst), 1); in cancel_during_run() 101 assert_eq!(POLL.load(Ordering::SeqCst), 1); in cancel_during_run() 102 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 0); in cancel_during_run() 103 assert_eq!(DROP_F.load(Ordering::SeqCst), 0); in cancel_during_run() 104 assert_eq!(DROP_S.load(Ordering::SeqCst), 0); in cancel_during_run() 116 assert_eq!(POLL.load(Ordering::SeqCst), 1); in run_and_join() 117 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 0); in run_and_join() [all …]
|
D | ready.rs | 104 assert_eq!(POLL.load(Ordering::SeqCst), 1); in cancel_during_run() 105 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 0); in cancel_during_run() 106 assert_eq!(DROP_F.load(Ordering::SeqCst), 1); in cancel_during_run() 107 assert_eq!(DROP_S.load(Ordering::SeqCst), 1); in cancel_during_run() 108 assert_eq!(DROP_T.load(Ordering::SeqCst), 1); in cancel_during_run() 113 assert_eq!(POLL.load(Ordering::SeqCst), 1); in cancel_during_run() 114 assert_eq!(SCHEDULE.load(Ordering::SeqCst), 0); in cancel_during_run() 115 assert_eq!(DROP_F.load(Ordering::SeqCst), 0); in cancel_during_run() 116 assert_eq!(DROP_S.load(Ordering::SeqCst), 0); in cancel_during_run() 117 assert_eq!(DROP_T.load(Ordering::SeqCst), 0); in cancel_during_run() [all …]
|
/external/autotest/metadata/ |
D | tests.star | 9 load('//metadata/tests/graphics.star', define_graphics = 'define_tests') 10 load('//metadata/tests/platform.star', define_platform = 'define_tests') 11 load('//metadata/tests/desktopui.star', define_desktopui = 'define_tests') 12 load('//metadata/tests/policy.star', define_policy = 'define_tests') 13 load('//metadata/tests/security.star', define_security = 'define_tests') 14 load('//metadata/tests/network.star', define_network = 'define_tests') 15 load('//metadata/tests/dummy.star', define_dummy = 'define_tests') 16 load('//metadata/tests/power.star', define_power = 'define_tests') 17 load('//metadata/tests/kernel.star', define_kernel = 'define_tests') 18 load('//metadata/tests/hardware.star', define_hardware = 'define_tests') [all …]
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | load-constant-i1.ll | 13 %load = load i1, i1 addrspace(2)* %in 14 store i1 %load, i1 addrspace(1)* %out 20 %load = load <2 x i1>, <2 x i1> addrspace(2)* %in 21 store <2 x i1> %load, <2 x i1> addrspace(1)* %out 27 %load = load <3 x i1>, <3 x i1> addrspace(2)* %in 28 store <3 x i1> %load, <3 x i1> addrspace(1)* %out 34 %load = load <4 x i1>, <4 x i1> addrspace(2)* %in 35 store <4 x i1> %load, <4 x i1> addrspace(1)* %out 41 %load = load <8 x i1>, <8 x i1> addrspace(2)* %in 42 store <8 x i1> %load, <8 x i1> addrspace(1)* %out [all …]
|
D | load-global-i1.ll | 13 %load = load i1, i1 addrspace(1)* %in 14 store i1 %load, i1 addrspace(1)* %out 20 %load = load <2 x i1>, <2 x i1> addrspace(1)* %in 21 store <2 x i1> %load, <2 x i1> addrspace(1)* %out 27 %load = load <3 x i1>, <3 x i1> addrspace(1)* %in 28 store <3 x i1> %load, <3 x i1> addrspace(1)* %out 34 %load = load <4 x i1>, <4 x i1> addrspace(1)* %in 35 store <4 x i1> %load, <4 x i1> addrspace(1)* %out 41 %load = load <8 x i1>, <8 x i1> addrspace(1)* %in 42 store <8 x i1> %load, <8 x i1> addrspace(1)* %out [all …]
|
D | load-local-i1.ll | 14 %load = load i1, i1 addrspace(3)* %in 15 store i1 %load, i1 addrspace(3)* %out 21 %load = load <2 x i1>, <2 x i1> addrspace(3)* %in 22 store <2 x i1> %load, <2 x i1> addrspace(3)* %out 28 %load = load <3 x i1>, <3 x i1> addrspace(3)* %in 29 store <3 x i1> %load, <3 x i1> addrspace(3)* %out 35 %load = load <4 x i1>, <4 x i1> addrspace(3)* %in 36 store <4 x i1> %load, <4 x i1> addrspace(3)* %out 42 %load = load <8 x i1>, <8 x i1> addrspace(3)* %in 43 store <8 x i1> %load, <8 x i1> addrspace(3)* %out [all …]
|
D | load-local-i8.ll | 14 %ld = load i8, i8 addrspace(3)* %in 27 %ld = load <2 x i8>, <2 x i8> addrspace(3)* %in 38 %ld = load <3 x i8>, <3 x i8> addrspace(3)* %in 49 %ld = load <4 x i8>, <4 x i8> addrspace(3)* %in 61 %ld = load <8 x i8>, <8 x i8> addrspace(3)* %in 76 %ld = load <16 x i8>, <16 x i8> addrspace(3)* %in 88 %a = load i8, i8 addrspace(3)* %in 102 %ld = load i8, i8 addrspace(3)* %in 110 %load = load <1 x i8>, <1 x i8> addrspace(3)* %in 111 %ext = zext <1 x i8> %load to <1 x i32> [all …]
|
D | load-constant-i8.ll | 14 %ld = load i8, i8 addrspace(2)* %in 26 %ld = load <2 x i8>, <2 x i8> addrspace(2)* %in 37 %ld = load <3 x i8>, <3 x i8> addrspace(2)* %in 48 %ld = load <4 x i8>, <4 x i8> addrspace(2)* %in 59 %ld = load <8 x i8>, <8 x i8> addrspace(2)* %in 70 %ld = load <16 x i8>, <16 x i8> addrspace(2)* %in 81 %a = load i8, i8 addrspace(2)* %in 95 %ld = load i8, i8 addrspace(2)* %in 103 %load = load <1 x i8>, <1 x i8> addrspace(2)* %in 104 %ext = zext <1 x i8> %load to <1 x i32> [all …]
|
D | load-global-i8.ll | 15 %ld = load i8, i8 addrspace(1)* %in 27 %ld = load <2 x i8>, <2 x i8> addrspace(1)* %in 39 %ld = load <3 x i8>, <3 x i8> addrspace(1)* %in 51 %ld = load <4 x i8>, <4 x i8> addrspace(1)* %in 63 %ld = load <8 x i8>, <8 x i8> addrspace(1)* %in 76 %ld = load <16 x i8>, <16 x i8> addrspace(1)* %in 87 %a = load i8, i8 addrspace(1)* %in 101 %ld = load i8, i8 addrspace(1)* %in 109 %load = load <1 x i8>, <1 x i8> addrspace(1)* %in 110 %ext = zext <1 x i8> %load to <1 x i32> [all …]
|
/external/linux-kselftest/tools/testing/selftests/powerpc/ptrace/ |
D | ptrace-vsx.h | 13 int validate_vsx(unsigned long *vsx, unsigned long *load) in validate_vsx() argument 18 if (vsx[i] != load[2 * i + 1]) { in validate_vsx() 20 i, vsx[i], 2 * i + 1, load[2 * i + 1]); in validate_vsx() 31 int validate_vmx(unsigned long vmx[][2], unsigned long *load) in validate_vmx() argument 37 if ((vmx[i][0] != load[64 + 2 * i]) || in validate_vmx() 38 (vmx[i][1] != load[65 + 2 * i])) { in validate_vmx() 41 load[64 + 2 * i]); in validate_vmx() 44 load[65 + 2 * i]); in validate_vmx() 51 if ((vmx[i][0] != load[65 + 2 * i]) || in validate_vmx() 52 (vmx[i][1] != load[64 + 2 * i])) { in validate_vmx() [all …]
|
/external/llvm/test/Analysis/CostModel/X86/ |
D | load_store.ll | 36 ;CHECK: cost of 1 {{.*}} load 37 load i8, i8* undef, align 4 38 ;CHECK: cost of 1 {{.*}} load 39 load i16, i16* undef, align 4 40 ;CHECK: cost of 1 {{.*}} load 41 load i32, i32* undef, align 4 42 ;CHECK: cost of 1 {{.*}} load 43 load i64, i64* undef, align 4 44 ;CHECK: cost of 2 {{.*}} load 45 load i128, i128* undef, align 4 [all …]
|
/external/mesa3d/src/compiler/nir/tests/ |
D | load_store_vectorizer_tests.cpp | 30 #define EXPECT_INSTR_SWIZZLES(instr, load, expected_swizzle) \ argument 31 EXPECT_EQ((instr)->src[0].src.ssa, &(load)->dest.ssa); \ 218 nir_intrinsic_instr *load = nir_intrinsic_instr_create(b->shader, intrinsic); in create_indirect_load() local 219 nir_ssa_dest_init(&load->instr, &load->dest, components, bit_size, NULL); in create_indirect_load() 220 load->num_components = components; in create_indirect_load() 222 load->src[0] = nir_src_for_ssa(res); in create_indirect_load() 223 load->src[1] = nir_src_for_ssa(offset); in create_indirect_load() 225 load->src[0] = nir_src_for_ssa(offset); in create_indirect_load() 230 nir_intrinsic_set_align(load, byte_size, 0); in create_indirect_load() 231 nir_intrinsic_set_access(load, (gl_access_qualifier)access); in create_indirect_load() [all …]
|
/external/rust/crates/grpcio-sys/grpc/third_party/abseil-cpp/absl/container/internal/ |
D | hashtablez_sampler_test.cc | 66 res.push_back(info.size.load(std::memory_order_acquire)); in GetSizes() 84 EXPECT_EQ(info.capacity.load(), 0); in TEST() 85 EXPECT_EQ(info.size.load(), 0); in TEST() 86 EXPECT_EQ(info.num_erases.load(), 0); in TEST() 87 EXPECT_EQ(info.num_rehashes.load(), 0); in TEST() 88 EXPECT_EQ(info.max_probe_length.load(), 0); in TEST() 89 EXPECT_EQ(info.total_probe_length.load(), 0); in TEST() 90 EXPECT_EQ(info.hashes_bitwise_or.load(), 0); in TEST() 91 EXPECT_EQ(info.hashes_bitwise_and.load(), ~size_t{}); in TEST() 92 EXPECT_EQ(info.hashes_bitwise_xor.load(), 0); in TEST() [all …]
|
/external/llvm/test/Transforms/EarlyCSE/ |
D | invariant-loads.ll | 7 ; CHECK: %val0 = load i32, i32* %ptr, !invariant.load !0 13 %val0 = load i32, i32* %ptr, !invariant.load !{} 15 %val1 = load i32, i32* %ptr, !invariant.load !{} 17 %val2 = load i32, i32* %ptr, !invariant.load !{} 24 ; invariant load has executed, the location loaded from is known to be 28 ; CHECK: %val0 = load i32, i32* %ptr, !invariant.load !0 32 %val0 = load i32, i32* %ptr, !invariant.load !{} 34 %val1 = load i32, i32* %ptr 40 ; Negative test -- we can't forward a non-invariant load into an 41 ; invariant load. [all …]
|
/external/llvm/test/Transforms/Inline/ |
D | inline-cold.ll | 20 %a1 = load volatile i32, i32* @a 22 %a2 = load volatile i32, i32* @a 24 %a3 = load volatile i32, i32* @a 26 %a4 = load volatile i32, i32* @a 28 %a5 = load volatile i32, i32* @a 30 %a6 = load volatile i32, i32* @a 32 %a7 = load volatile i32, i32* @a 34 %a8 = load volatile i32, i32* @a 36 %a9 = load volatile i32, i32* @a 38 %a10 = load volatile i32, i32* @a [all …]
|
/external/llvm/test/Transforms/GVN/ |
D | volatile.ll | 5 ; Check that we can bypass a volatile load when searching 6 ; for dependencies of a non-volatile load 9 ; CHECK: %0 = load volatile i32, i32* %q 12 %x = load i32, i32* %p 13 load volatile i32, i32* %q 14 %y = load i32, i32* %p 20 ; volatile, this would be (in effect) removing the volatile load 23 ; CHECK: %x = load i32, i32* %p 24 ; CHECK-NEXT: %y = load volatile i32, i32* %p 27 %x = load i32, i32* %p [all …]
|
/external/libavc/common/arm/ |
D | ih264_default_weighted_pred_a9q.s | 124 vld1.32 d0[0], [r0], r3 @load row 1 in source 1 125 vld1.32 d0[1], [r0], r3 @load row 2 in source 1 126 vld1.32 d2[0], [r1], r4 @load row 1 in source 2 127 vld1.32 d2[1], [r1], r4 @load row 2 in source 2 129 vld1.32 d1[0], [r0], r3 @load row 3 in source 1 130 vld1.32 d1[1], [r0], r3 @load row 4 in source 1 132 vld1.32 d3[0], [r1], r4 @load row 3 in source 2 133 vld1.32 d3[1], [r1], r4 @load row 4 in source 2 136 vst1.32 d0[0], [r2], r5 @load row 1 in destination 137 vst1.32 d0[1], [r2], r5 @load row 2 in destination [all …]
|