/external/tensorflow/tensorflow/core/kernels/ |
D | unary_ops_composition.cc | 160 #define REGISTER_COMPUTE_FN_HELPER(name, functor) \ argument 161 static_assert(std::is_same<functor::in_type, functor::out_type>::value, \ 165 *out = in.unaryExpr(functor::func()); \ 168 return Eigen::internal::functor_traits<functor::func>::Cost; \ 177 auto relu = functor::Relu<Eigen::DefaultDevice, T>(); \ 186 auto relu6 = functor::Relu6<Eigen::DefaultDevice, T>(); \ 195 auto elu = functor::Elu<Eigen::DefaultDevice, T>(); \ 204 auto selu = functor::Selu<Eigen::DefaultDevice, T>(); \ 261 REGISTER_COMPUTE_FN_HELPER(Abs, functor::abs<T>); 262 REGISTER_COMPUTE_FN_HELPER(Acos, functor::acos<T>); [all …]
|
D | strided_slice_op_gpu_impl.h | 34 template struct functor::StridedSlice<GPUDevice, T, 1>; \ 35 template struct functor::StridedSlice<GPUDevice, T, 2>; \ 36 template struct functor::StridedSlice<GPUDevice, T, 3>; \ 37 template struct functor::StridedSlice<GPUDevice, T, 4>; \ 38 template struct functor::StridedSlice<GPUDevice, T, 5>; \ 39 template struct functor::StridedSlice<GPUDevice, T, 6>; \ 40 template struct functor::StridedSlice<GPUDevice, T, 7>; \ 41 template struct functor::StridedSlice<GPUDevice, T, 8>; \ 42 template struct functor::StridedSliceGrad<GPUDevice, T, 1>; \ 43 template struct functor::StridedSliceGrad<GPUDevice, T, 2>; \ [all …]
|
D | segment_reduction_ops_impl_3.cc | 30 functor::UnsortedSegmentFunctor<CPUDevice, type, index_type, \ 36 functor::Zero<type>, \ 37 functor::SumOp<type>); \ 39 functor::Lowest<type>, \ 40 functor::MaxOp<type>); \ 42 functor::Highest<type>, \ 43 functor::MinOp<type>); \ 45 functor::One<type>, \ 46 functor::ProdOp<type>); 50 functor::Zero<type>, \ [all …]
|
D | segment_reduction_ops_impl_4.cc | 30 functor::UnsortedSegmentFunctor<CPUDevice, type, index_type, \ 36 functor::Zero<type>, \ 37 functor::SumOp<type>); \ 39 functor::Lowest<type>, \ 40 functor::MaxOp<type>); \ 42 functor::Highest<type>, \ 43 functor::MinOp<type>); \ 45 functor::One<type>, \ 46 functor::ProdOp<type>); 50 functor::Zero<type>, \ [all …]
|
D | cwise_op_div.cc | 19 REGISTER6(BinaryOp, CPU, "Div", functor::div, float, Eigen::half, double, 21 REGISTER5(BinaryOp, CPU, "Div", functor::safe_div, uint8, uint16, int16, int32, 23 REGISTER5(BinaryOp, CPU, "TruncateDiv", functor::safe_div, uint8, uint16, int16, 25 REGISTER6(BinaryOp, CPU, "RealDiv", functor::div, float, Eigen::half, double, 27 REGISTER5(BinaryOp, CPU, "DivNoNan", functor::div_no_nan, Eigen::half, float, 34 REGISTER9(BinaryOp, GPU, "Div", functor::div, float, Eigen::half, double, uint8, 36 REGISTER5(BinaryOp, GPU, "RealDiv", functor::div, float, Eigen::half, double, 38 REGISTER4(BinaryOp, GPU, "TruncateDiv", functor::div, uint8, uint16, int16, 41 REGISTER4(BinaryOp, GPU, "Div", functor::div, uint8, uint16, complex64, 43 REGISTER2(BinaryOp, GPU, "RealDiv", functor::div, complex64, complex128); [all …]
|
D | segment_reduction_ops_impl_2.cc | 21 #define REGISTER_CPU_KERNEL_SEGMENT(name, functor, type, index_type, \ argument 28 SegmentReductionOp<CPUDevice, type, index_type, functor, default_value>) 76 functor::SegmentReductionFunctor< \ 82 "SegmentSum", type, index_type, functor::Zero<type>, \ 83 functor::NonAtomicSumOpGpu<type>, functor::AtomicSumOpGpu<type>); \ 85 "SegmentProd", type, index_type, functor::One<type>, \ 86 functor::NonAtomicProdOpGpu<type>, functor::AtomicProdOpGpu<type>); \ 88 "SegmentMin", type, index_type, functor::Highest<type>, \ 89 functor::NonAtomicMinOpGpu<type>, functor::AtomicMinOpGpu<type>); \ 91 "SegmentMax", type, index_type, functor::Lowest<type>, \ [all …]
|
D | relu_op.h | 39 functor::Relu<Device, T> functor; in Operate() local 40 functor(context->eigen_device<Device>(), input.flat<T>(), in Operate() 87 functor::ReluGrad<Device, T> functor; in OperateNoTemplate() local 88 functor(context->eigen_device<Device>(), g.flat<T>(), a.flat<T>(), in OperateNoTemplate() 98 functor::Relu6<Device, T> functor; in Operate() local 99 functor(context->eigen_device<Device>(), input.flat<T>(), in Operate() 129 functor::Relu6Grad<Device, T> functor; in OperateNoTemplate() local 130 functor(context->eigen_device<Device>(), g.flat<T>(), a.flat<T>(), in OperateNoTemplate() 145 functor::LeakyRelu<Device, T> functor; in Operate() local 146 functor({context->eigen_device<Device>(), input.flat<T>(), alpha_, in Operate() [all …]
|
D | pad_op_gpu.cu.cc | 31 template struct functor::Pad<GPUDevice, T, Tpadding, 0>; \ 32 template struct functor::Pad<GPUDevice, T, Tpadding, 1>; \ 33 template struct functor::Pad<GPUDevice, T, Tpadding, 2>; \ 34 template struct functor::Pad<GPUDevice, T, Tpadding, 3>; \ 35 template struct functor::Pad<GPUDevice, T, Tpadding, 4>; \ 36 template struct functor::Pad<GPUDevice, T, Tpadding, 5>; \ 37 template struct functor::Pad<GPUDevice, T, Tpadding, 6>; \ 38 template struct functor::Pad<GPUDevice, T, Tpadding, 7>; \ 39 template struct functor::Pad<GPUDevice, T, Tpadding, 8>;
|
D | segment_reduction_ops_impl_1.cc | 71 #define REGISTER_CPU_KERNEL_SEGMENT(name, functor, type, index_type, \ argument 78 SegmentReductionOp<CPUDevice, type, index_type, functor, default_value>) 126 functor::SegmentReductionFunctor< \ 132 "SegmentSum", type, index_type, functor::Zero<type>, \ 133 functor::NonAtomicSumOpGpu<type>, functor::AtomicSumOpGpu<type>); \ 135 "SegmentProd", type, index_type, functor::One<type>, \ 136 functor::NonAtomicProdOpGpu<type>, functor::AtomicProdOpGpu<type>); \ 138 "SegmentMin", type, index_type, functor::Highest<type>, \ 139 functor::NonAtomicMinOpGpu<type>, functor::AtomicMinOpGpu<type>); \ 141 "SegmentMax", type, index_type, functor::Lowest<type>, \ [all …]
|
D | segment_reduction_ops_gpu.cu.cc | 129 namespace functor { namespace 220 template struct SegmentReductionFunctor<T, Index, functor::Zero<T>, \ 221 functor::NonAtomicSumOpGpu<T>, \ 222 functor::AtomicSumOpGpu<T>>; \ 223 template struct SegmentReductionFunctor<T, Index, functor::One<T>, \ 224 functor::NonAtomicProdOpGpu<T>, \ 225 functor::AtomicProdOpGpu<T>>; \ 226 template struct SegmentReductionFunctor<T, Index, functor::Highest<T>, \ 227 functor::NonAtomicMinOpGpu<T>, \ 228 functor::AtomicMinOpGpu<T>>; \ [all …]
|
D | slice_op_gpu.cu.cc | 31 template struct functor::Slice<GPUDevice, T, 1>; \ 32 template struct functor::Slice<GPUDevice, T, 2>; \ 33 template struct functor::Slice<GPUDevice, T, 3>; \ 34 template struct functor::Slice<GPUDevice, T, 4>; \ 35 template struct functor::Slice<GPUDevice, T, 5>; \ 36 template struct functor::Slice<GPUDevice, T, 6>; \ 37 template struct functor::Slice<GPUDevice, T, 7>; \ 38 template struct functor::Slice<GPUDevice, T, 8>;
|
D | cwise_op_reciprocal.cc | 19 REGISTER5(UnaryOp, CPU, "Inv", functor::inverse, float, Eigen::half, double, 22 REGISTER4(UnaryOp, GPU, "Inv", functor::inverse, float, Eigen::half, double, 26 REGISTER5(SimpleBinaryOp, CPU, "InvGrad", functor::inverse_grad, float, 29 REGISTER3(SimpleBinaryOp, GPU, "InvGrad", functor::inverse_grad, float, 33 REGISTER6(UnaryOp, CPU, "Reciprocal", functor::inverse, float, Eigen::half, 36 REGISTER4(UnaryOp, GPU, "Reciprocal", functor::inverse, float, Eigen::half, 40 REGISTER6(SimpleBinaryOp, CPU, "ReciprocalGrad", functor::inverse_grad, float, 43 REGISTER3(SimpleBinaryOp, GPU, "ReciprocalGrad", functor::inverse_grad, float,
|
D | training_ops_gpu.cu.cc | 28 namespace functor { namespace 1086 template struct functor::ApplyGradientDescent<GPUDevice, Eigen::half>; 1087 template struct functor::ApplyGradientDescent<GPUDevice, float>; 1088 template struct functor::ApplyGradientDescent<GPUDevice, double>; 1091 template struct functor::ApplyGradientDescent<GPUDevice, complex64>; 1092 template struct functor::ApplyGradientDescent<GPUDevice, complex128>; 1095 template struct functor::ApplyAdagrad<GPUDevice, Eigen::half>; 1096 template struct functor::ApplyAdagrad<GPUDevice, float>; 1097 template struct functor::ApplyAdagrad<GPUDevice, double>; 1100 template struct functor::ApplyAdagrad<GPUDevice, complex64>; [all …]
|
D | cwise_op_mul_1.cc | 20 REGISTER6(BinaryOp, CPU, "Mul", functor::mul, float, Eigen::half, double, uint8, 22 REGISTER6(BinaryOp, CPU, "MulNoNan", functor::mul_no_nan, Eigen::half, float, 29 REGISTER(BinaryOp, CPU, "Mul", functor::mul, int32); 35 REGISTER4(BinaryOp, GPU, "Mul", functor::mul, Eigen::half, float, double, 38 REGISTER(BinaryOp, GPU, "Mul", functor::mul, uint8); 49 BinaryOp<CPUDevice, functor::mul<int32>>); 53 REGISTER5(BinaryOp, GPU, "MulNoNan", functor::mul_no_nan, Eigen::half, float,
|
D | cwise_op_add_1.cc | 19 REGISTER6(BinaryOp, CPU, "Add", functor::add, float, Eigen::half, double, int32, 24 REGISTER6(BinaryOp, CPU, "AddV2", functor::add, float, Eigen::half, double, 27 REGISTER(BinaryOp, CPU, "AddV2", functor::add, bfloat16); 34 REGISTER3(BinaryOp, GPU, "Add", functor::add, float, Eigen::half, double); 35 REGISTER3(BinaryOp, GPU, "AddV2", functor::add, float, Eigen::half, double); 47 BinaryOp<CPUDevice, functor::add<int32>>); 54 BinaryOp<CPUDevice, functor::add<int32>>);
|
D | cwise_op_igammas.cc | 20 REGISTER2(BinaryOp, CPU, "Igamma", functor::igamma, float, double); 21 REGISTER2(BinaryOp, CPU, "IgammaGradA", functor::igamma_grad_a, float, double); 22 REGISTER2(BinaryOp, CPU, "Igammac", functor::igammac, float, double); 24 REGISTER2(BinaryOp, GPU, "Igamma", functor::igamma, float, double); 25 REGISTER2(BinaryOp, GPU, "IgammaGradA", functor::igamma_grad_a, float, double); 26 REGISTER2(BinaryOp, GPU, "Igammac", functor::igammac, float, double);
|
D | quantize_and_dequantize_op_gpu.cu.cc | 29 namespace functor { namespace 90 template struct functor::QuantizeAndDequantizeOneScaleFunctor<GPUDevice, float>; 91 template struct functor::QuantizeAndDequantizeOneScaleFunctor<GPUDevice, 94 template struct functor::QuantizeAndDequantizePerChannelFunctor<GPUDevice, 96 template struct functor::QuantizeAndDequantizePerChannelFunctor<GPUDevice, 99 template struct functor::QuantizeAndDequantizeOneScaleGradientFunctor<GPUDevice, 101 template struct functor::QuantizeAndDequantizeOneScaleGradientFunctor<GPUDevice, 103 template struct functor::QuantizeAndDequantizePerChannelGradientFunctor< 105 template struct functor::QuantizeAndDequantizePerChannelGradientFunctor<
|
D | spacetodepth_op.cc | 135 functor::SpaceToDepthOpFunctor<Device, int32, FORMAT_NCHW> functor; in Compute() local 136 functor(context->eigen_device<Device>(), Tinput_v, block_size_, in Compute() 140 functor::SpaceToDepthOpFunctor<Device, RT, FORMAT_NCHW> functor; in Compute() local 141 functor(context->eigen_device<Device>(), input.tensor<RT, 4>(), in Compute() 145 functor::SpaceToDepthOpFunctor<Device, RT, FORMAT_NHWC> functor; in Compute() local 146 functor(context->eigen_device<Device>(), input.tensor<RT, 4>(), in Compute() 152 functor::SpaceToDepthOpFunctor<Device, T, FORMAT_NHWC> functor; in Compute() local 153 functor(context->eigen_device<Device>(), input.tensor<T, 4>(), in Compute() 164 namespace functor { namespace
|
D | cwise_op_mod.cc | 19 REGISTER2(BinaryOp, CPU, "Mod", functor::safe_mod, int32, int64); 20 REGISTER2(BinaryOp, CPU, "Mod", functor::fmod, float, double); 21 REGISTER2(BinaryOp, CPU, "TruncateMod", functor::safe_mod, int32, int64); 22 REGISTER2(BinaryOp, CPU, "TruncateMod", functor::fmod, float, double); 34 BinaryOp<CPUDevice, functor::safe_mod<int32>>); 41 BinaryOp<CPUDevice, functor::safe_mod<int32>>);
|
/external/tensorflow/tensorflow/core/kernels/special_math/ |
D | special_math_op_bessel.cc | 20 REGISTER3(UnaryOp, CPU, "BesselI0", functor::bessel_i0, Eigen::half, float, 22 REGISTER3(UnaryOp, CPU, "BesselI1", functor::bessel_i1, Eigen::half, float, 24 REGISTER3(UnaryOp, CPU, "BesselI0e", functor::bessel_i0e, Eigen::half, float, 26 REGISTER3(UnaryOp, CPU, "BesselI1e", functor::bessel_i1e, Eigen::half, float, 29 REGISTER3(UnaryOp, CPU, "BesselK0", functor::bessel_k0, Eigen::half, float, 31 REGISTER3(UnaryOp, CPU, "BesselK1", functor::bessel_k1, Eigen::half, float, 33 REGISTER3(UnaryOp, CPU, "BesselK0e", functor::bessel_k0e, Eigen::half, float, 35 REGISTER3(UnaryOp, CPU, "BesselK1e", functor::bessel_k1e, Eigen::half, float, 38 REGISTER3(UnaryOp, CPU, "BesselJ0", functor::bessel_j0, Eigen::half, float, 40 REGISTER3(UnaryOp, CPU, "BesselJ1", functor::bessel_j1, Eigen::half, float, [all …]
|
/external/tensorflow/tensorflow/core/kernels/image/ |
D | mirror_pad_op_gpu.cu.cc | 28 template struct functor::MirrorPad<GpuDevice, T, int32, 1>; \ 29 template struct functor::MirrorPad<GpuDevice, T, int32, 2>; \ 30 template struct functor::MirrorPad<GpuDevice, T, int32, 3>; \ 31 template struct functor::MirrorPad<GpuDevice, T, int32, 4>; \ 32 template struct functor::MirrorPad<GpuDevice, T, int32, 5>; \ 33 template struct functor::MirrorPad<GpuDevice, T, int64, 1>; \ 34 template struct functor::MirrorPad<GpuDevice, T, int64, 2>; \ 35 template struct functor::MirrorPad<GpuDevice, T, int64, 3>; \ 36 template struct functor::MirrorPad<GpuDevice, T, int64, 4>; \ 37 template struct functor::MirrorPad<GpuDevice, T, int64, 5>; \ [all …]
|
D | adjust_contrast_op_gpu.cu.cc | 29 template struct functor::AdjustContrastv2<GPUDevice, float>; 30 template struct functor::AdjustContrastv2<GPUDevice, Eigen::half>; 33 template struct functor::AdjustContrast<GPUDevice, uint8>; 34 template struct functor::AdjustContrast<GPUDevice, int8>; 35 template struct functor::AdjustContrast<GPUDevice, int16>; 36 template struct functor::AdjustContrast<GPUDevice, int32>; 37 template struct functor::AdjustContrast<GPUDevice, int64>; 38 template struct functor::AdjustContrast<GPUDevice, float>; 39 template struct functor::AdjustContrast<GPUDevice, double>;
|
/external/webrtc/rtc_base/ |
D | callback.h | 77 Callback0(const T& functor) in Callback0() argument 78 : helper_(new RefCountedObject<HelperImpl<T> >(functor)) {} in Callback0() 93 explicit HelperImpl(const T& functor) : functor_(functor) {} in HelperImpl() 106 Callback1(const T& functor) in Callback1() argument 107 : helper_(new RefCountedObject<HelperImpl<T> >(functor)) {} in Callback1() 122 explicit HelperImpl(const T& functor) : functor_(functor) {} in HelperImpl() 135 Callback2(const T& functor) in Callback2() argument 136 : helper_(new RefCountedObject<HelperImpl<T> >(functor)) {} in Callback2() 151 explicit HelperImpl(const T& functor) : functor_(functor) {} in HelperImpl() 164 Callback3(const T& functor) in Callback3() argument [all …]
|
/external/llvm-project/clang/test/Analysis/ |
D | cxx-uninitialized-object.cpp | 750 Callable functor; member 752 LambdaThisTest(const Callable &functor, int) : functor(functor) { in LambdaThisTest() 766 Callable functor; member 768 LambdaTest1(const Callable &functor, int) : functor(functor) { in LambdaTest1() 781 Callable functor; member 783 …LambdaTest2(const Callable &functor, int) : functor(functor) {} // expected-warning{{1 uninitializ… in LambdaTest2() 794 Callable functor; member 796 LambdaTest2(const Callable &functor, int) : functor(functor) {} in LambdaTest2() 817 Callable functor; member 819 …LambdaTest3(const Callable &functor, int) : functor(functor) {} // expected-warning{{2 uninitializ… in LambdaTest3() [all …]
|
/external/tensorflow/tensorflow/core/kernels/sparse/ |
D | transpose_op.h | 24 namespace functor { 34 functor::UnaryFunctor<Device, functor::conj<complex64>> conj; 43 functor::UnaryFunctor<Device, functor::conj<complex128>> conj; 57 functor::UnaryFunctor<Device, functor::conj<complex64>> conj; 65 functor::UnaryFunctor<Device, functor::conj<complex128>> conj;
|