/external/deqp/external/openglcts/data/mustpass/gles/aosp_mustpass/3.2.4.x/src/ |
D | gles31-hw-issues.txt | 2 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_less.clamp_to_edge_repeat 3 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_less.repeat_mirrored_rep… 4 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_less.mirrored_repeat_cla… 5 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_greater.clamp_to_edge_re… 6 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_greater.repeat_mirrored_… 7 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_greater.mirrored_repeat_… 8 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_npot.compare_less.clamp_to_edge_repe… 9 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_npot.compare_less.repeat_mirrored_re… 10 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_npot.compare_less.mirrored_repeat_cl… 11 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_npot.compare_greater.clamp_to_edge_r… [all …]
|
/external/deqp/external/openglcts/data/mustpass/gles/aosp_mustpass/3.2.5.x/src/ |
D | gles31-hw-issues.txt | 2 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_less.clamp_to_edge_repeat 3 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_less.repeat_mirrored_rep… 4 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_less.mirrored_repeat_cla… 5 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_greater.clamp_to_edge_re… 6 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_greater.repeat_mirrored_… 7 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_greater.mirrored_repeat_… 8 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_npot.compare_less.clamp_to_edge_repe… 9 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_npot.compare_less.repeat_mirrored_re… 10 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_npot.compare_less.mirrored_repeat_cl… 11 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_npot.compare_greater.clamp_to_edge_r… [all …]
|
/external/deqp/external/openglcts/data/mustpass/gles/aosp_mustpass/3.2.3.x/src/ |
D | gles31-hw-issues.txt | 2 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_less.clamp_to_edge_repeat 3 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_less.repeat_mirrored_rep… 4 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_less.mirrored_repeat_cla… 5 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_greater.clamp_to_edge_re… 6 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_greater.repeat_mirrored_… 7 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_pot.compare_greater.mirrored_repeat_… 8 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_npot.compare_less.clamp_to_edge_repe… 9 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_npot.compare_less.repeat_mirrored_re… 10 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_npot.compare_less.mirrored_repeat_cl… 11 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.size_npot.compare_greater.clamp_to_edge_r… [all …]
|
/external/tensorflow/tensorflow/python/kernel_tests/ |
D | gather_op_test.py | 55 gather_t = array_ops.gather(params, indices_tf) 70 gather_t = array_ops.gather(params, indices, axis=axis) 86 gather_t = array_ops.gather(params, indices, axis=axis) 108 gather = array_ops.gather(tf_params, tf_indices, axis=tf_axis) 109 gather_negative_axis = array_ops.gather( 112 [gather, gather_negative_axis]) 118 self.assertEqual(expected_shape, gather.shape) 123 *gather.get_shape().as_list()).astype(dtype.as_numpy_dtype) 127 gather, [tf_params, tf_indices, tf_axis], gather_grad) 160 array_ops.gather(params, 1, axis=0).eval()) [all …]
|
/external/tensorflow/tensorflow/python/ops/parallel_for/ |
D | array_test.py | 42 x_i = array_ops.gather(x, i) 46 outputs.append(array_ops.gather(y, 2, axis=axis)) 47 outputs.append(array_ops.gather(y, i, axis=axis)) 48 outputs.append(array_ops.gather(y, [i], axis=axis)) 49 outputs.append(array_ops.gather(y, [i, 2], axis=axis)) 50 outputs.append(array_ops.gather(y, [[2, i], [i, 1]], axis=axis)) 59 x_i = array_ops.gather(x, i) 68 x_i = array_ops.gather(x, i) 77 x_i = array_ops.gather(x, i) 87 x_i = array_ops.gather(x, i) [all …]
|
D | math_test.py | 104 x1 = array_ops.gather(x, i) 133 return op(array_ops.gather(x, i)) 198 x1 = array_ops.gather(x, i) 199 y1 = array_ops.gather(y, i) 213 x1 = array_ops.gather(x, i) 214 y1 = array_ops.gather(y, i) 225 x1 = array_ops.gather(x, i) 247 a = array_ops.gather(x, i) if stack_a else x 248 b = array_ops.gather(y, i) if stack_b else y 272 a = array_ops.gather(x, i) if stack_a else x [all …]
|
D | control_flow_ops_test.py | 69 x_i = array_ops.gather(x, i) 86 return array_ops.gather(x, i) 112 x_i = array_ops.gather(x, i) 126 x_i = array_ops.gather(x, i) 138 x_i = array_ops.gather(x, i) 155 x_i = array_ops.gather(x, i) 168 x_i = array_ops.gather(x, i) 181 x_i = array_ops.gather(x, i) 198 x1 = array_ops.gather(x, i) 219 x1 = array_ops.gather(x, i) [all …]
|
/external/deqp/android/cts/master/src/ |
D | gles31-hw-issues.txt | 35 dEQP-GLES31.functional.texture.gather.basic.2d.depth32f.size_pot.compare_greater.mirrored_repeat_cl… 36 dEQP-GLES31.functional.texture.gather.basic.2d.depth32f.size_pot.compare_greater.repeat_mirrored_re… 37 dEQP-GLES31.functional.texture.gather.basic.2d.depth32f.size_pot.compare_less.mirrored_repeat_clamp… 38 dEQP-GLES31.functional.texture.gather.basic.2d.depth32f.size_pot.compare_less.repeat_mirrored_repeat 41 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.base_level.level_1 42 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.base_level.level_2 43 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.filter_mode.min_linear_mag_linear 44 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.filter_mode.min_linear_mipmap_linear_mag_… 45 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.filter_mode.min_linear_mipmap_nearest_mag… 46 dEQP-GLES31.functional.texture.gather.basic.cube.depth32f.filter_mode.min_nearest_mipmap_linear_mag… [all …]
|
/external/tensorflow/tensorflow/python/ops/ragged/ |
D | ragged_gather_op_test.py | 43 ragged_gather_ops.gather(params, ragged_indices), 46 ragged_gather_ops.gather(ragged_params, indices), 49 ragged_gather_ops.gather(ragged_params, ragged_indices), 56 ragged_gather_ops.gather(params, indices), [b'c', b'a', b'c', b'b']) 57 self.assertIsInstance(ragged_gather_ops.gather(params, indices), ops.Tensor) 64 ragged_gather_ops.gather(params, indices), 71 ragged_gather_ops.gather(params, indices), 79 ragged_gather_ops.gather(params, indices), 90 ragged_gather_ops.gather(params, indices), [b'c', b'd', b'e']) 97 ragged_gather_ops.gather(params, indices), [all …]
|
D | ragged_gather_ops.py | 36 def gather(params, indices, validate_indices=None, axis=0, batch_dims=0, function 101 return indices.with_values(gather(params, indices.values)) 104 return array_ops.gather(params, indices) 228 return gather(params, flattened_index_tuples) 238 flattened_index_tuples = array_ops.gather(params.row_splits, 252 flattened_index_tuples = array_ops.gather( 258 return gather(flattened_params, flattened_index_tuples)
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/NewGVN/ |
D | 2016-08-30-MaskedScatterGather.ll | 5 declare <2 x i32> @llvm.masked.gather.v2i32.v2p0i32(<2 x i32*>, i32, <2 x i1>, <2 x i32>) 7 ; This test ensures that masked scatter and gather operations, which take vectors of pointers, 9 ; No scatter/gather calls should end up eliminated 10 ; CHECK: llvm.masked.gather 11 ; CHECK: llvm.masked.gather 13 ; CHECK: llvm.masked.gather 15 ; CHECK: llvm.masked.gather 24 …%in1.v = call <2 x i32> @llvm.masked.gather.v2i32.v2p0i32(<2 x i32*> %in1, i32 1, <2 x i1> <i1 tru… 25 …%in2.v = call <2 x i32> @llvm.masked.gather.v2i32.v2p0i32(<2 x i32*> %in2, i32 1, <2 x i1> <i1 tru… 29 ; This gather should alias the scatter we just saw [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/GVN/ |
D | 2016-08-30-MaskedScatterGather.ll | 4 declare <2 x i32> @llvm.masked.gather.v2i32.v2p0i32(<2 x i32*>, i32, <2 x i1>, <2 x i32>) 6 ; This test ensures that masked scatter and gather operations, which take vectors of pointers, 8 ; No scatter/gather calls should end up eliminated 9 ; CHECK: llvm.masked.gather 10 ; CHECK: llvm.masked.gather 12 ; CHECK: llvm.masked.gather 14 ; CHECK: llvm.masked.gather 23 …%in1.v = call <2 x i32> @llvm.masked.gather.v2i32.v2p0i32(<2 x i32*> %in1, i32 1, <2 x i1> <i1 tru… 24 …%in2.v = call <2 x i32> @llvm.masked.gather.v2i32.v2p0i32(<2 x i32*> %in2, i32 1, <2 x i1> <i1 tru… 28 ; This gather should alias the scatter we just saw [all …]
|
/external/llvm/test/Analysis/CostModel/X86/ |
D | masked-intrinsic-cost.ll | 73 ; AVX2: Found an estimated cost of 7 {{.*}}.gather 76 ; KNL: Found an estimated cost of 7 {{.*}}.gather 79 ; SKX: Found an estimated cost of 7 {{.*}}.gather 81 %res = call <2 x double> @llvm.masked.gather.v2f64(<2 x double*> %ptrs, i32 4, <2 x i1> %mask, <2 x… 84 declare <2 x double> @llvm.masked.gather.v2f64(<2 x double*> %ptrs, i32, <2 x i1> %mask, <2 x doubl… 89 ; AVX2: Found an estimated cost of 16 {{.*}}.gather 92 ; KNL: Found an estimated cost of 16 {{.*}}.gather 95 ; SKX: Found an estimated cost of 6 {{.*}}.gather 97 %res = call <4 x i32> @llvm.masked.gather.v4i32(<4 x i32*> %ptrs, i32 4, <4 x i1> %mask, <4 x i32> … 104 ; AVX2: Found an estimated cost of 8 {{.*}}.gather [all …]
|
/external/llvm/test/Analysis/BasicAA/ |
D | pr18573.ll | 3 ; Check that llvm.x86.avx2.gather.d.ps.256 intrinsic is not eliminated as gather and store memory a… 8 declare <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float>, i8*, <8 x i32>, <8 x float>, i8) #0 16 …%v1 = tail call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> undef, i8* %arr.ptr, <8 x i… 19 …%v2 = tail call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> undef, i8* %arr.ptr, <8 x i… 25 ; CHECK: llvm.x86.avx2.gather.d.ps.256 27 ; CHECK: llvm.x86.avx2.gather.d.ps.256 29 ; Check that second gather is eliminated as gather and store memory accesses are based on different… 37 …%v1 = tail call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> undef, i8* %arr.ptr, <8 x i… 40 …%v2 = tail call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> undef, i8* %arr.ptr, <8 x i… 46 ; CHECK: llvm.x86.avx2.gather.d.ps.256 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Analysis/BasicAA/ |
D | pr18573.ll | 3 ; Check that llvm.x86.avx2.gather.d.ps.256 intrinsic is not eliminated as gather and store memory a… 8 declare <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float>, i8*, <8 x i32>, <8 x float>, i8) #0 16 …%v1 = tail call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> undef, i8* %arr.ptr, <8 x i… 19 …%v2 = tail call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> undef, i8* %arr.ptr, <8 x i… 25 ; CHECK: llvm.x86.avx2.gather.d.ps.256 27 ; CHECK: llvm.x86.avx2.gather.d.ps.256 29 ; Check that second gather is eliminated as gather and store memory accesses are based on different… 37 …%v1 = tail call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> undef, i8* %arr.ptr, <8 x i… 40 …%v2 = tail call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> undef, i8* %arr.ptr, <8 x i… 46 ; CHECK: llvm.x86.avx2.gather.d.ps.256 [all …]
|
/external/tensorflow/tensorflow/compiler/tests/ |
D | gather_test.py | 52 gather_t = array_ops.gather(params, indices_tf) 66 gather_t = array_ops.gather(params, indices, axis=axis) 82 gather_t = array_ops.gather(params, indices, axis=axis) 102 gather_t = array_ops.gather(params, indices, axis=axis) 123 gather = array_ops.gather(tf_params, tf_indices, axis=axis) 124 gather_value = sess.run(gather, feed_dict={tf_params: params}) 135 gather = array_ops.gather(params, indices) 137 7, gather.eval(feed_dict={params: [4, 7, 2], indices: 1})) 139 [7], gather.eval(feed_dict={params: [4, 7, 2], indices: [1]})) 141 [[7]], gather.eval(feed_dict={params: [4, 7, 2], indices: [[1]]})) [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | avx2-gather.ll | 4 declare <4 x float> @llvm.x86.avx2.gather.d.ps(<4 x float>, i8*, 9 %res = call <4 x float> @llvm.x86.avx2.gather.d.ps(<4 x float> undef, 21 declare <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double>, i8*, 26 %res = call <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double> undef, 36 declare <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float>, i8*, 41 %res = call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> undef, 49 declare <4 x double> @llvm.x86.avx2.gather.d.pd.256(<4 x double>, i8*, 54 %res = call <4 x double> @llvm.x86.avx2.gather.d.pd.256(<4 x double> undef,
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Verifier/ |
D | scatter_gather.ll | 6 …%res = call <16 x float> @llvm.masked.gather.v16f32.v16p0f32(<16 x float*> %ptrs, i32 4, <16 x i1>… 9 declare <16 x float> @llvm.masked.gather.v16f32.v16p0f32(<16 x float*>, i32, <16 x i1>*, <16 x floa… 14 …%res = call <8 x float> @llvm.masked.gather.v8f32.v8p0f32(<8 x float*> %ptrs, i32 4, <16 x i1> %ma… 17 declare <8 x float> @llvm.masked.gather.v8f32.v8p0f32(<8 x float*>, i32, <16 x i1>, <8 x float>) 22 …%res = call <8 x float>* @llvm.masked.gather.p0v8f32.v8p0f32(<8 x float*> %ptrs, i32 4, <8 x i1> %… 25 declare <8 x float>* @llvm.masked.gather.p0v8f32.v8p0f32(<8 x float*>, i32, <8 x i1>, <8 x float>) 30 …%res = call <8 x float> @llvm.masked.gather.v8f32.p0v8p0f32(<8 x float*>* %ptrs, i32 4, <8 x i1> %… 33 declare <8 x float> @llvm.masked.gather.v8f32.p0v8p0f32(<8 x float*>*, i32, <8 x i1>, <8 x float>) 38 …%res = call <8 x float> @llvm.masked.gather.v8f32.v8f32(<8 x float> %ptrs, i32 4, <8 x i1> %mask, … 41 declare <8 x float> @llvm.masked.gather.v8f32.v8f32(<8 x float>, i32, <8 x i1>, <8 x float>) [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/ |
D | avx2-gather.ll | 5 declare <4 x float> @llvm.x86.avx2.gather.d.ps(<4 x float>, i8*, 23 %res = call <4 x float> @llvm.x86.avx2.gather.d.ps(<4 x float> undef, 28 declare <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double>, i8*, 46 %res = call <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double> undef, 51 declare <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float>, i8*, 69 %res = call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> undef, 74 declare <4 x double> @llvm.x86.avx2.gather.d.pd.256(<4 x double>, i8*, 92 %res = call <4 x double> @llvm.x86.avx2.gather.d.pd.256(<4 x double> undef, 117 …%call = call <4 x i32> @llvm.x86.avx2.gather.d.d(<4 x i32> zeroinitializer, i8* %arg0, <4 x i32> %… 121 declare <4 x i32> @llvm.x86.avx2.gather.d.d(<4 x i32>, i8*, <4 x i32>, <4 x i32>, i8) nounwind read… [all …]
|
D | pr33772.ll | 5 ; make sure we don't crash if scale for gather isn't constant. 7 ; CHECK: LLVM ERROR: Cannot select: intrinsic %llvm.x86.avx512.gather.dpi.512 8 declare <16 x i32> @llvm.x86.avx512.gather.dpi.512(<16 x i32>, i8*, <16 x i32>, i16, i32) 13 …%res = tail call <16 x i32> @llvm.x86.avx512.gather.dpi.512(<16 x i32> undef, i8* %ptr, <16 x i32>…
|
D | speculative-load-hardening-gather.ll | 4 declare <4 x float> @llvm.x86.avx2.gather.d.ps(<4 x float>, i8*, <4 x i32>, <4 x float>, i8) 23 …%v = call <4 x float> @llvm.x86.avx2.gather.d.ps(<4 x float> zeroinitializer, i8* %b, <4 x i32> %i… 27 declare <4 x float> @llvm.x86.avx2.gather.q.ps(<4 x float>, i8*, <2 x i64>, <4 x float>, i8) 46 …%v = call <4 x float> @llvm.x86.avx2.gather.q.ps(<4 x float> zeroinitializer, i8* %b, <2 x i64> %i… 50 declare <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double>, i8*, <4 x i32>, <2 x double>, i8) 69 …%v = call <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double> zeroinitializer, i8* %b, <4 x i32> … 73 declare <2 x double> @llvm.x86.avx2.gather.q.pd(<2 x double>, i8*, <2 x i64>, <2 x double>, i8) 92 …%v = call <2 x double> @llvm.x86.avx2.gather.q.pd(<2 x double> zeroinitializer, i8* %b, <2 x i64> … 96 declare <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float>, i8*, <8 x i32>, <8 x float>, i8) 115 …%v = call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> zeroinitializer, i8* %b, <8 x i32… [all …]
|
/external/tensorflow/tensorflow/contrib/opt/python/training/ |
D | lazy_adam_optimizer.py | 63 beta1_t * array_ops.gather(m, grad.indices) + 70 beta2_t * array_ops.gather(v, grad.indices) + 75 m_t_slice = array_ops.gather(m_t, grad.indices) 76 v_t_slice = array_ops.gather(v_t, grad.indices) 95 m_t_slice = beta1_t * array_ops.gather(m, indices) + (1 - beta1_t) * grad 102 v_t_slice = (beta2_t * array_ops.gather(v, indices) +
|
D | lazy_adam_gs_optimizer.py | 63 beta1_t * array_ops.gather(m, grad.indices) + 70 beta2_t * array_ops.gather(v, grad.indices) + 75 m_t_slice = array_ops.gather(m_t, grad.indices) 76 v_t_slice = array_ops.gather(v_t, grad.indices) 95 m_t_slice = beta1_t * array_ops.gather(m, indices) + (1 - beta1_t) * grad 102 v_t_slice = (beta2_t * array_ops.gather(v, indices) +
|
/external/tensorflow/tensorflow/contrib/image/python/ops/ |
D | dense_image_warp.py | 143 def gather(y_coords, x_coords, name): function 146 gathered_values = array_ops.gather(flattened_grid, linear_coordinates) 151 top_left = gather(floors[0], floors[1], 'top_left') 152 top_right = gather(floors[0], ceils[1], 'top_right') 153 bottom_left = gather(ceils[0], floors[1], 'bottom_left') 154 bottom_right = gather(ceils[0], ceils[1], 'bottom_right')
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | gather_op.cc | 183 xla::XlaOp gather; in Compile() local 187 builder, &gather)); in Compile() 188 context->SetOutput(0, gather); in Compile() 224 xla::XlaOp gather; in Compile() local 228 indices_type, builder, &gather)); in Compile() 229 context->SetOutput(0, gather); in Compile()
|