Home
last modified time | relevance | path

Searched refs:compute (Results 1 – 25 of 1408) sorted by relevance

12345678910>>...57

/external/tensorflow/tensorflow/python/keras/layers/preprocessing/
Dpreprocessing_test_utils.py70 single_compute = combiner.compute(data)
73 combiner.compute(data_0),
74 combiner.compute(data_1),
75 combiner.compute(data_2)
84 combiner.compute(data_1),
85 combiner.compute(data_2),
86 combiner.compute(data_0)
95 combiner.compute(data_1),
96 combiner.merge([combiner.compute(data_2),
97 combiner.compute(data_0)])
[all …]
/external/XNNPACK/src/
Doperator-run.c717 switch (op->compute.type) { in xnn_run_operator()
721 assert(op->compute.range[0] != 0); in xnn_run_operator()
724 op->compute.task_1d, in xnn_run_operator()
726 op->compute.range[0], in xnn_run_operator()
730 assert(op->compute.range[0] != 0); in xnn_run_operator()
731 assert(op->compute.tile[0] != 0); in xnn_run_operator()
734 op->compute.task_1d_tile_1d, in xnn_run_operator()
736 op->compute.range[0], in xnn_run_operator()
737 op->compute.tile[0], in xnn_run_operator()
741 assert(op->compute.range[0] != 0); in xnn_run_operator()
[all …]
Dclamp-nc.c218 clamp_op->compute.type = xnn_parallelization_type_1d_tile_1d; in xnn_setup_clamp_nc_u8()
219 …clamp_op->compute.task_1d_tile_1d = (pthreadpool_task_1d_tile_1d_t) xnn_compute_univector_contiguo… in xnn_setup_clamp_nc_u8()
220 clamp_op->compute.range[0] = batch_size * channels * sizeof(uint8_t); in xnn_setup_clamp_nc_u8()
221 clamp_op->compute.tile[0] = block_size; in xnn_setup_clamp_nc_u8()
232 clamp_op->compute.type = xnn_parallelization_type_1d_tile_1d; in xnn_setup_clamp_nc_u8()
233 … clamp_op->compute.task_1d_tile_1d = (pthreadpool_task_1d_tile_1d_t) xnn_compute_univector_strided; in xnn_setup_clamp_nc_u8()
234 clamp_op->compute.range[0] = batch_size; in xnn_setup_clamp_nc_u8()
235 clamp_op->compute.tile[0] = 1; in xnn_setup_clamp_nc_u8()
278 clamp_op->compute.type = xnn_parallelization_type_1d_tile_1d; in xnn_setup_clamp_nc_f32()
279 …clamp_op->compute.task_1d_tile_1d = (pthreadpool_task_1d_tile_1d_t) xnn_compute_univector_contiguo… in xnn_setup_clamp_nc_f32()
[all …]
Dsigmoid-nc.c263 sigmoid_op->compute.type = xnn_parallelization_type_1d_tile_1d; in xnn_setup_sigmoid_nc_q8()
264 … sigmoid_op->compute.task_1d_tile_1d = (pthreadpool_task_1d_tile_1d_t) xnn_compute_lut_contiguous; in xnn_setup_sigmoid_nc_q8()
265 sigmoid_op->compute.range[0] = batch_size * channels * sizeof(uint8_t); in xnn_setup_sigmoid_nc_q8()
266 sigmoid_op->compute.tile[0] = block_size; in xnn_setup_sigmoid_nc_q8()
277 sigmoid_op->compute.type = xnn_parallelization_type_1d; in xnn_setup_sigmoid_nc_q8()
278 sigmoid_op->compute.task_1d = (pthreadpool_task_1d_t) xnn_compute_lut_strided; in xnn_setup_sigmoid_nc_q8()
279 sigmoid_op->compute.range[0] = batch_size; in xnn_setup_sigmoid_nc_q8()
280 sigmoid_op->compute.tile[0] = 0; in xnn_setup_sigmoid_nc_q8()
322 sigmoid_op->compute.type = xnn_parallelization_type_1d_tile_1d; in xnn_setup_sigmoid_nc_f32()
323 …sigmoid_op->compute.task_1d_tile_1d = (pthreadpool_task_1d_tile_1d_t) xnn_compute_univector_contig… in xnn_setup_sigmoid_nc_f32()
[all …]
Dadd-nc.c287 add_op->compute.type = xnn_parallelization_type_1d_tile_1d; in xnn_setup_add_nc_q8()
288 add_op->compute.task_1d_tile_1d = (pthreadpool_task_1d_tile_1d_t) xnn_compute_add_contiguous; in xnn_setup_add_nc_q8()
289 add_op->compute.range[0] = batch_size * channels * sizeof(uint8_t); in xnn_setup_add_nc_q8()
290 add_op->compute.tile[0] = block_size; in xnn_setup_add_nc_q8()
303 add_op->compute.type = xnn_parallelization_type_1d_tile_1d; in xnn_setup_add_nc_q8()
304 add_op->compute.task_1d_tile_1d = (pthreadpool_task_1d_tile_1d_t) xnn_compute_add_strided; in xnn_setup_add_nc_q8()
305 add_op->compute.range[0] = batch_size; in xnn_setup_add_nc_q8()
306 add_op->compute.tile[0] = 1; in xnn_setup_add_nc_q8()
350 add_op->compute.type = xnn_parallelization_type_1d_tile_1d; in xnn_setup_add_nc_f32()
351 add_op->compute.task_1d_tile_1d = (pthreadpool_task_1d_tile_1d_t) xnn_compute_add_contiguous; in xnn_setup_add_nc_f32()
[all …]
Dhardswish-nc.c120 hardswish_op->compute.type = xnn_parallelization_type_1d_tile_1d; in xnn_setup_hardswish_nc_f32()
121 …hardswish_op->compute.task_1d_tile_1d = (pthreadpool_task_1d_tile_1d_t) xnn_compute_univector_cont… in xnn_setup_hardswish_nc_f32()
122 hardswish_op->compute.range[0] = batch_size * channels * sizeof(float); in xnn_setup_hardswish_nc_f32()
123 hardswish_op->compute.tile[0] = block_size; in xnn_setup_hardswish_nc_f32()
134 hardswish_op->compute.type = xnn_parallelization_type_1d_tile_1d; in xnn_setup_hardswish_nc_f32()
135 …hardswish_op->compute.task_1d_tile_1d = (pthreadpool_task_1d_tile_1d_t) xnn_compute_univector_stri… in xnn_setup_hardswish_nc_f32()
136 hardswish_op->compute.range[0] = batch_size; in xnn_setup_hardswish_nc_f32()
137 hardswish_op->compute.tile[0] = 1; in xnn_setup_hardswish_nc_f32()
Ddeconvolution-nhwc.c630 deconvolution_op->compute.type = xnn_parallelization_type_3d_tile_2d; in setup_conv_path()
631 deconvolution_op->compute.task_3d_tile_2d = (pthreadpool_task_3d_tile_2d_t) xnn_compute_igemm; in setup_conv_path()
632 deconvolution_op->compute.range[0] = batch_size; in setup_conv_path()
633 deconvolution_op->compute.range[1] = output_size; in setup_conv_path()
634 deconvolution_op->compute.range[2] = group_output_channels; in setup_conv_path()
635 deconvolution_op->compute.tile[0] = mr; in setup_conv_path()
636 deconvolution_op->compute.tile[1] = nc; in setup_conv_path()
638 deconvolution_op->compute.type = xnn_parallelization_type_4d_tile_2d; in setup_conv_path()
639 deconvolution_op->compute.task_4d_tile_2d = (pthreadpool_task_4d_tile_2d_t) xnn_compute_gigemm; in setup_conv_path()
640 deconvolution_op->compute.range[0] = batch_size; in setup_conv_path()
[all …]
Dconvolution-nhwc.c836 convolution_op->compute.type = xnn_parallelization_type_2d_tile_2d; in setup_convolution2d_nhwc()
837 convolution_op->compute.task_2d_tile_2d = (pthreadpool_task_2d_tile_2d_t) xnn_compute_gemm; in setup_convolution2d_nhwc()
838 convolution_op->compute.range[0] = batch_output_size; in setup_convolution2d_nhwc()
839 convolution_op->compute.range[1] = group_output_channels; in setup_convolution2d_nhwc()
840 convolution_op->compute.tile[0] = mr; in setup_convolution2d_nhwc()
841 convolution_op->compute.tile[1] = nc; in setup_convolution2d_nhwc()
843 convolution_op->compute.type = xnn_parallelization_type_3d_tile_2d; in setup_convolution2d_nhwc()
844 convolution_op->compute.task_3d_tile_2d = (pthreadpool_task_3d_tile_2d_t) xnn_compute_ggemm; in setup_convolution2d_nhwc()
845 convolution_op->compute.range[0] = groups; in setup_convolution2d_nhwc()
846 convolution_op->compute.range[1] = batch_output_size; in setup_convolution2d_nhwc()
[all …]
Dleaky-relu-nc.c195 leaky_relu_op->compute.type = xnn_parallelization_type_1d_tile_1d; in xnn_setup_leaky_relu_nc_q8()
196 …leaky_relu_op->compute.task_1d_tile_1d = (pthreadpool_task_1d_tile_1d_t) xnn_compute_lut_contiguou… in xnn_setup_leaky_relu_nc_q8()
197 leaky_relu_op->compute.range[0] = batch_size * channels * sizeof(uint8_t); in xnn_setup_leaky_relu_nc_q8()
198 leaky_relu_op->compute.tile[0] = block_size; in xnn_setup_leaky_relu_nc_q8()
209 leaky_relu_op->compute.type = xnn_parallelization_type_1d; in xnn_setup_leaky_relu_nc_q8()
210 leaky_relu_op->compute.task_1d = (pthreadpool_task_1d_t) xnn_compute_lut_strided; in xnn_setup_leaky_relu_nc_q8()
211 leaky_relu_op->compute.range[0] = batch_size; in xnn_setup_leaky_relu_nc_q8()
212 leaky_relu_op->compute.tile[0] = 0; in xnn_setup_leaky_relu_nc_q8()
/external/guava/android/guava-tests/test/com/google/common/math/
DQuantilesTest.java127 assertThat(median().compute(SIXTEEN_SQUARES_DOUBLES)) in testMedian_compute_doubleCollection()
139 assertThat(quartiles().index(1).compute(SIXTEEN_SQUARES_DOUBLES)) in testQuartiles_index_compute_doubleCollection()
153 assertThat(quartiles().indexes(1, 3).compute(SIXTEEN_SQUARES_DOUBLES)) in testQuartiles_indexes_varargs_compute_doubleCollection()
169 assertThat(Quantiles.scale(10).index(1).compute(SIXTEEN_SQUARES_DOUBLES)) in testScale_index_compute_doubleCollection()
175 assertThat(Quantiles.scale(10).index(1).compute(SIXTEEN_SQUARES_LONGS)) in testScale_index_compute_longCollection()
181 assertThat(Quantiles.scale(10).index(1).compute(SIXTEEN_SQUARES_INTEGERS)) in testScale_index_compute_integerCollection()
188 assertThat(Quantiles.scale(10).index(1).compute(dataset)) in testScale_index_compute_doubleVarargs()
199 assertThat(Quantiles.scale(10).index(1).compute(dataset)) in testScale_index_compute_longVarargs()
207 assertThat(Quantiles.scale(10).index(1).compute(dataset)) in testScale_index_compute_intVarargs()
229 assertThat(Quantiles.scale(10).indexes(0, 10, 5, 1, 8, 1).compute(SIXTEEN_SQUARES_DOUBLES)) in testScale_indexes_varargs_compute_doubleCollection()
[all …]
/external/guava/guava-tests/test/com/google/common/math/
DQuantilesTest.java127 assertThat(median().compute(SIXTEEN_SQUARES_DOUBLES)) in testMedian_compute_doubleCollection()
139 assertThat(quartiles().index(1).compute(SIXTEEN_SQUARES_DOUBLES)) in testQuartiles_index_compute_doubleCollection()
153 assertThat(quartiles().indexes(1, 3).compute(SIXTEEN_SQUARES_DOUBLES)) in testQuartiles_indexes_varargs_compute_doubleCollection()
169 assertThat(Quantiles.scale(10).index(1).compute(SIXTEEN_SQUARES_DOUBLES)) in testScale_index_compute_doubleCollection()
175 assertThat(Quantiles.scale(10).index(1).compute(SIXTEEN_SQUARES_LONGS)) in testScale_index_compute_longCollection()
181 assertThat(Quantiles.scale(10).index(1).compute(SIXTEEN_SQUARES_INTEGERS)) in testScale_index_compute_integerCollection()
188 assertThat(Quantiles.scale(10).index(1).compute(dataset)) in testScale_index_compute_doubleVarargs()
199 assertThat(Quantiles.scale(10).index(1).compute(dataset)) in testScale_index_compute_longVarargs()
207 assertThat(Quantiles.scale(10).index(1).compute(dataset)) in testScale_index_compute_intVarargs()
229 assertThat(Quantiles.scale(10).indexes(0, 10, 5, 1, 8, 1).compute(SIXTEEN_SQUARES_DOUBLES)) in testScale_indexes_varargs_compute_doubleCollection()
[all …]
/external/mesa3d/src/mesa/drivers/dri/i965/
Dbrw_compute.c102 brw->compute.num_work_groups_bo = NULL; in brw_dispatch_compute()
103 brw->compute.num_work_groups = num_groups; in brw_dispatch_compute()
104 brw->compute.group_size = NULL; in brw_dispatch_compute()
121 brw->compute.num_work_groups_bo = bo; in brw_dispatch_compute_indirect()
122 brw->compute.num_work_groups_offset = indirect; in brw_dispatch_compute_indirect()
123 brw->compute.num_work_groups = indirect_group_counts; in brw_dispatch_compute_indirect()
124 brw->compute.group_size = NULL; in brw_dispatch_compute_indirect()
137 brw->compute.num_work_groups_bo = NULL; in brw_dispatch_compute_group_size()
138 brw->compute.num_work_groups = num_groups; in brw_dispatch_compute_group_size()
139 brw->compute.group_size = group_size; in brw_dispatch_compute_group_size()
/external/deqp/external/openglcts/data/mustpass/gl/khronos_mustpass_single/4.6.1.x/
Dgl45-khr-single.txt3 KHR-Single-GL45.subgroups.builtin_var.compute.subgroupsize_compute
4 KHR-Single-GL45.subgroups.builtin_var.compute.subgroupinvocationid_compute
5 KHR-Single-GL45.subgroups.builtin_var.compute.numsubgroups
6 KHR-Single-GL45.subgroups.builtin_var.compute.subgroupid
20 KHR-Single-GL45.subgroups.builtin_mask_var.compute.subgroupeqmask
21 KHR-Single-GL45.subgroups.builtin_mask_var.compute.subgroupgemask
22 KHR-Single-GL45.subgroups.builtin_mask_var.compute.subgroupgtmask
23 KHR-Single-GL45.subgroups.builtin_mask_var.compute.subgrouplemask
24 KHR-Single-GL45.subgroups.builtin_mask_var.compute.subgroupltmask
50 KHR-Single-GL45.subgroups.basic.compute.subgroupelect
[all …]
Dgl46-khr-single.txt3 KHR-Single-GL46.subgroups.builtin_var.compute.subgroupsize_compute
4 KHR-Single-GL46.subgroups.builtin_var.compute.subgroupinvocationid_compute
5 KHR-Single-GL46.subgroups.builtin_var.compute.numsubgroups
6 KHR-Single-GL46.subgroups.builtin_var.compute.subgroupid
20 KHR-Single-GL46.subgroups.builtin_mask_var.compute.subgroupeqmask
21 KHR-Single-GL46.subgroups.builtin_mask_var.compute.subgroupgemask
22 KHR-Single-GL46.subgroups.builtin_mask_var.compute.subgroupgtmask
23 KHR-Single-GL46.subgroups.builtin_mask_var.compute.subgrouplemask
24 KHR-Single-GL46.subgroups.builtin_mask_var.compute.subgroupltmask
50 KHR-Single-GL46.subgroups.basic.compute.subgroupelect
[all …]
/external/deqp/external/openglcts/data/mustpass/gles/khronos_mustpass_single/master/
Dgles32-khr-single.txt3 KHR-Single-GLES32.subgroups.builtin_var.compute.subgroupsize_compute
4 KHR-Single-GLES32.subgroups.builtin_var.compute.subgroupinvocationid_compute
5 KHR-Single-GLES32.subgroups.builtin_var.compute.numsubgroups
6 KHR-Single-GLES32.subgroups.builtin_var.compute.subgroupid
20 KHR-Single-GLES32.subgroups.builtin_mask_var.compute.subgroupeqmask
21 KHR-Single-GLES32.subgroups.builtin_mask_var.compute.subgroupgemask
22 KHR-Single-GLES32.subgroups.builtin_mask_var.compute.subgroupgtmask
23 KHR-Single-GLES32.subgroups.builtin_mask_var.compute.subgrouplemask
24 KHR-Single-GLES32.subgroups.builtin_mask_var.compute.subgroupltmask
50 KHR-Single-GLES32.subgroups.basic.compute.subgroupelect
[all …]
/external/deqp/external/openglcts/data/mustpass/gles/khronos_mustpass_single/3.2.6.x/
Dgles32-khr-single.txt3 KHR-Single-GLES32.subgroups.builtin_var.compute.subgroupsize_compute
4 KHR-Single-GLES32.subgroups.builtin_var.compute.subgroupinvocationid_compute
5 KHR-Single-GLES32.subgroups.builtin_var.compute.numsubgroups
6 KHR-Single-GLES32.subgroups.builtin_var.compute.subgroupid
20 KHR-Single-GLES32.subgroups.builtin_mask_var.compute.subgroupeqmask
21 KHR-Single-GLES32.subgroups.builtin_mask_var.compute.subgroupgemask
22 KHR-Single-GLES32.subgroups.builtin_mask_var.compute.subgroupgtmask
23 KHR-Single-GLES32.subgroups.builtin_mask_var.compute.subgrouplemask
24 KHR-Single-GLES32.subgroups.builtin_mask_var.compute.subgroupltmask
50 KHR-Single-GLES32.subgroups.basic.compute.subgroupelect
[all …]
/external/universal-tween-engine/java/api/src/aurelienribon/tweenengine/equations/
DBounce.java13 public final float compute(float t) {
14 return 1 - OUT.compute(1-t);
25 public final float compute(float t) {
45 public final float compute(float t) {
46 if (t < 0.5f) return IN.compute(t*2) * .5f;
47 else return OUT.compute(t*2-1) * .5f + 0.5f;
/external/skqp/src/compute/sk/
DSkSurface_Compute.cpp33 SkSurface_Compute::SkSurface_Compute(sk_sp<SkContext_Compute> compute, in SkSurface_Compute() argument
36 compute(compute) in SkSurface_Compute()
60 SkDevice_Compute * const device_compute = new SkDevice_Compute(compute,w,h); in onNewCanvas()
87 return sk_make_sp<SkSurface_Compute>(compute,info.width(),info.height()); in onNewSurface()
107 return sk_make_sp<SkImage_Compute>(compute,snap,w,h); in onNewImageSnapshot()
131 sk_sp<SkContext_Compute> compute)
182 return sk_make_sp<SkSurface_Compute>(compute,desc.fWidth,desc.fHeight);
/external/guava/guava-testlib/src/com/google/common/collect/testing/testers/
DMapComputeTester.java47 .compute( in testCompute_absentToPresent()
63 .compute( in testCompute_absentToAbsent()
81 .compute( in testCompute_presentToPresent()
98 .compute( in testCompute_presentToAbsent()
119 .compute( in testCompute_presentNullToPresentNonnull()
139 .compute( in testCompute_presentNullToNull()
158 .compute( in testCompute_nullKeyPresentToPresent()
175 .compute( in testCompute_presentFunctionThrows()
192 .compute( in testCompute_absentFunctionThrows()
/external/llvm/test/CodeGen/MIR/X86/
Dused-physical-register-info.mir7 define i32 @compute(i32 %a) #0 {
15 %b = call i32 @compute(i32 %a)
21 %b = call i32 @compute(i32 %a)
27 %b = call i32 @compute(i32 %a)
35 # CHECK: name: compute
39 name: compute
69 … CALL64pcrel32 @compute, csr_64, implicit %rsp, implicit %edi, implicit-def %rsp, implicit-def %eax
89 … CALL64pcrel32 @compute, csr_64, implicit %rsp, implicit %edi, implicit-def %rsp, implicit-def %eax
106 … CALL64pcrel32 @compute, csr_64, implicit %rsp, implicit %edi, implicit-def %rsp, implicit-def %eax
/external/eigen/test/
Dnomalloc.cpp121 Eigen::LLT<Matrix> LLT; LLT.compute(A); in ctms_decompositions()
124 Eigen::LDLT<Matrix> LDLT; LDLT.compute(A); in ctms_decompositions()
129 Eigen::HessenbergDecomposition<ComplexMatrix> hessDecomp; hessDecomp.compute(complexA); in ctms_decompositions()
130 Eigen::ComplexSchur<ComplexMatrix> cSchur(size); cSchur.compute(complexA); in ctms_decompositions()
131 Eigen::ComplexEigenSolver<ComplexMatrix> cEigSolver; cEigSolver.compute(complexA); in ctms_decompositions()
132 Eigen::EigenSolver<Matrix> eigSolver; eigSolver.compute(A); in ctms_decompositions()
133 Eigen::SelfAdjointEigenSolver<Matrix> saEigSolver(size); saEigSolver.compute(saA); in ctms_decompositions()
134 Eigen::Tridiagonalization<Matrix> tridiag; tridiag.compute(saA); in ctms_decompositions()
137 Eigen::PartialPivLU<Matrix> ppLU; ppLU.compute(A); in ctms_decompositions()
140 Eigen::FullPivLU<Matrix> fpLU; fpLU.compute(A); in ctms_decompositions()
[all …]
Dsvd_common.h288 svd.compute(MatrixType::Constant(10,10,some_inf), ComputeFullU | ComputeFullV); in svd_inf_nan()
292 svd.compute(MatrixType::Constant(10,10,nan), ComputeFullU | ComputeFullV); in svd_inf_nan()
296 svd.compute(m, ComputeFullU | ComputeFullV); in svd_inf_nan()
300 svd.compute(m, ComputeFullU | ComputeFullV); in svd_inf_nan()
307 svd.compute(m, ComputeFullU | ComputeFullV); in svd_inf_nan()
314 svd.compute(m, ComputeFullU | ComputeFullV); in svd_inf_nan()
331 svd.compute(M,ComputeFullU|ComputeFullV); in svd_underoverflow()
342 svd.compute(M,ComputeFullU|ComputeFullV); in svd_underoverflow()
366 svd3.compute(M3,ComputeFullU|ComputeFullV); // just check we don't loop indefinitely in svd_underoverflow()
407 svd.compute(m); in svd_preallocate()
[all …]
/external/eigen/bench/
Ddense_solvers.cpp18 solver.compute(A.transpose()*A); in compute_norm_equation()
20 solver.compute(A); in compute_norm_equation()
25 void compute(Solver &solver, const MatrixType &A) { in compute() function
26 solver.compute(A); in compute()
64 BENCH(t_qr, tries, rep, compute(qr,A)); in bench()
65 BENCH(t_cpqr, tries, rep, compute(cpqr,A)); in bench()
66 BENCH(t_cod, tries, rep, compute(cod,A)); in bench()
68 BENCH(t_fpqr, tries, rep, compute(fpqr,A)); in bench()
70 BENCH(t_jsvd, tries, rep, jsvd.compute(A,svd_opt)); in bench()
72 BENCH(t_bdcsvd, tries, rep, bdcsvd.compute(A,svd_opt)); in bench()
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/MIR/X86/
Dregister-mask-operands.mir6 define i32 @compute(i32 %a) #0 {
14 %b = call i32 @compute(i32 %a)
22 name: compute
34 …; CHECK-NEXT: CALL64pcrel32 @compute, csr_64, implicit $rsp, implicit $edi, implicit-def $rsp, imp…
36 … CALL64pcrel32 @compute, csr_64, implicit $rsp, implicit $edi, implicit-def $rsp, implicit-def $eax
/external/deqp/doc/testspecs/VK/
Dsparse_resources.txt27 …es - one supporting sparse binding operations, the second one supporting compute and transfer oper…
46 …es - one supporting sparse binding operations, the second one supporting compute and transfer oper…
66 …es - one supporting sparse binding operations, the second one supporting compute and transfer oper…
71 The second queue is used to perform compute and transfer operations. A compute shader is invoked to…
75 from compute shader. For parts of the data that correspond to the regions of sparse buffer that hav…
76 against expected output from compute shader. For parts that correspond to gaps, the data is random …
85 …es - one supporting sparse binding operations, the second one supporting compute and transfer oper…
90 The second queue is used to perform compute and transfer operations. A compute shader is invoked to…
94 from compute shader. For parts of the data that correspond to the regions of image that have device…
95 against expected output from compute shader. For parts that correspond to gaps, the data is random …
[all …]

12345678910>>...57