/external/python/cpython3/Lib/test/ |
D | test_tokenize.py | 2 from tokenize import (tokenize, _tokenize, untokenize, NUMBER, NAME, OP, 403 OP '(' (1, 7) (1, 8) 405 OP ',' (1, 9) (1, 10) 407 OP ',' (1, 12) (1, 13) 409 OP '=' (1, 15) (1, 16) 411 OP ',' (1, 17) (1, 18) 413 OP '=' (1, 20) (1, 21) 415 OP ',' (1, 22) (1, 23) 416 OP '*' (1, 24) (1, 25) 418 OP ')' (1, 26) (1, 27) [all …]
|
/external/jemalloc_new/include/jemalloc/internal/ |
D | mutex_prof.h | 9 OP(background_thread) \ 10 OP(ctl) \ 11 OP(prof) 14 #define OP(mtx) global_prof_mutex_##mtx, macro 16 #undef OP 21 OP(large) \ 22 OP(extent_avail) \ 23 OP(extents_dirty) \ 24 OP(extents_muzzy) \ 25 OP(extents_retained) \ [all …]
|
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
D | TensorSyclExtractFunctors.h | 47 template <template <class, class> class UnaryCategory, typename OP, typename RHSExpr, typename Dev> 48 struct FunctorExtractor<TensorEvaluator<const UnaryCategory<OP, RHSExpr>, Dev> > { 50 OP func; 51 FunctorExtractor(const TensorEvaluator<const UnaryCategory<OP, RHSExpr>, Dev>& expr) 56 template <template <class, class> class UnaryCategory, typename OP, typename RHSExpr, typename Dev> 57 struct FunctorExtractor<TensorEvaluator<UnaryCategory<OP, RHSExpr>, Dev> > 58 : FunctorExtractor<TensorEvaluator<const UnaryCategory<OP, RHSExpr>, Dev> >{}; 62 template <template<class, class, class> class BinaryCategory, typename OP, typename LHSExpr, typena… 63 struct FunctorExtractor<TensorEvaluator<const BinaryCategory<OP, LHSExpr, RHSExpr>, Dev> > { 66 OP func; [all …]
|
D | TensorSyclExtractAccessor.h | 68 template <template<class, class> class UnaryCategory, typename OP, typename RHSExpr, typename Dev> 69 struct ExtractAccessor<TensorEvaluator<const UnaryCategory<OP, RHSExpr>, Dev> > { 70 …static inline auto getTuple(cl::sycl::handler& cgh, const TensorEvaluator<const UnaryCategory<OP, … 77 template <template<class, class> class UnaryCategory, typename OP, typename RHSExpr, typename Dev> 78 struct ExtractAccessor<TensorEvaluator<UnaryCategory<OP, RHSExpr>, Dev> > 79 : ExtractAccessor<TensorEvaluator<const UnaryCategory<OP, RHSExpr>, Dev> > {}; 82 template <template<class, class, class> class BinaryCategory, typename OP, typename LHSExpr, typen… 83 struct ExtractAccessor<TensorEvaluator<const BinaryCategory<OP, LHSExpr, RHSExpr>, Dev> > { 84 …static inline auto getTuple(cl::sycl::handler& cgh, const TensorEvaluator<const BinaryCategory<OP,… 90 template <template<class, class, class> class BinaryCategory, typename OP, typename LHSExpr, typen… [all …]
|
D | TensorSyclLeafCount.h | 57 template <template <class, class...> class CategoryExpr, typename OP, typename... RHSExpr> 58 struct LeafCount<const CategoryExpr<OP, RHSExpr...> >: CategoryCount<RHSExpr...> {}; 60 template <template <class, class...> class CategoryExpr, typename OP, typename... RHSExpr> 61 struct LeafCount<CategoryExpr<OP, RHSExpr...> > :LeafCount<const CategoryExpr<OP, RHSExpr...> >{}; 97 template <typename OP, typename Dim, typename Expr> 98 struct LeafCount<const TensorReductionOp<OP, Dim, Expr> > { 103 template <typename OP, typename Dim, typename Expr> 104 struct LeafCount<TensorReductionOp<OP, Dim, Expr> >: LeafCount<const TensorReductionOp<OP, Dim, Exp…
|
D | TensorSyclExprConstructor.h | 64 template <template<class, class> class UnaryCategory, typename OP, typename OrigRHSExpr, typename R… 65 struct ExprConstructor<CVQual UnaryCategory<OP, OrigRHSExpr>, CVQual UnaryCategory<OP, RHSExpr>, Pa… 68 typedef CVQual UnaryCategory<OP, typename my_type::Type> Type;\ 82 template <template<class, class, class> class BinaryCategory, typename OP, typename OrigLHSExpr, ty… 84 struct ExprConstructor<CVQual BinaryCategory<OP, OrigLHSExpr, OrigRHSExpr>, CVQual BinaryCategory<… 87 …typedef CVQual BinaryCategory<OP, typename my_left_type::Type, typename my_right_type::Type> Type… 103 template <template <class, class, class, class> class TernaryCategory, typename OP, typename OrigAr… 105 … ExprConstructor<CVQual TernaryCategory<OP, OrigArg1Expr, OrigArg2Expr, OrigArg3Expr>, CVQual Tern… 109 …typedef CVQual TernaryCategory<OP, typename my_arg1_type::Type, typename my_arg2_type::Type, type… 211 template <typename OP, typename Dim, typename OrigExpr, typename DevExpr, size_t N, typename... Par… [all …]
|
D | TensorSyclPlaceHolderExpr.h | 73 template<template<class...> class Category , class OP, class TPL> 76 template<template<class...> class Category , class OP, class ...T > 77 struct CategoryHelper<Category, OP, utility::tuple::Tuple<T...> > { 78 typedef Category<OP, T... > Type; 89 template <template <class, class... > class Category, typename OP, typename... SubExpr, size_t N>\ 90 struct PlaceHolderExpression<CVQual Category<OP, SubExpr...>, N>{\ 91 …typedef CVQual typename CategoryHelper<Category, OP, typename CalculateIndex<N, SubExpr...>::ArgsT… 162 template <typename OP, typename Dims, typename Expr, size_t N>\ 163 struct PlaceHolderExpression<CVQual TensorReductionOp<OP, Dims, Expr>, N>{\ 164 typedef CVQual PlaceHolder<CVQual TensorReductionOp<OP, Dims,Expr>, N> Type;\
|
D | TensorSyclConvertToDeviceExpression.h | 63 template <template<class, class...> class Category, typename OP, typename... subExprs>\ 64 struct ConvertToDeviceExpression<CVQual Category<OP, subExprs...> > {\ 65 typedef CVQual Category<OP, typename ConvertToDeviceExpression<subExprs>::Type... > Type;\ 108 template <typename OP, typename Dim, typename subExpr, template <class> class MakePointer_>\ 109 struct ConvertToDeviceExpression<CVQual TensorReductionOp<OP, Dim, subExpr, MakePointer_> > {\ 110 …typedef CVQual TensorReductionOp<OP, Dim, typename ConvertToDeviceExpression<subExpr>::Type, MakeG…
|
/external/python/cpython2/Modules/ |
D | operator.c | 13 #define spam1(OP,AOP) static PyObject *OP(PyObject *s, PyObject *a1) { \ argument 16 #define spam2(OP,AOP) static PyObject *OP(PyObject *s, PyObject *a) { \ argument 18 if(! PyArg_UnpackTuple(a,#OP,2,2,&a1,&a2)) return NULL; \ 21 #define spamoi(OP,AOP) static PyObject *OP(PyObject *s, PyObject *a) { \ argument 23 if(! PyArg_ParseTuple(a,"Oi:" #OP,&a1,&a2)) return NULL; \ 26 #define spam2n(OP,AOP) static PyObject *OP(PyObject *s, PyObject *a) { \ argument 28 if(! PyArg_UnpackTuple(a,#OP,2,2,&a1,&a2)) return NULL; \ 33 #define spam3n(OP,AOP) static PyObject *OP(PyObject *s, PyObject *a) { \ argument 35 if(! PyArg_UnpackTuple(a,#OP,3,3,&a1,&a2,&a3)) return NULL; \ 40 #define spami(OP,AOP) static PyObject *OP(PyObject *s, PyObject *a1) { \ argument [all …]
|
/external/llvm/test/Transforms/InstCombine/ |
D | fold-fops-into-selects.ll | 18 ; CHECK: [[OP:%.*]] = fadd float %B, 2.000000e+00 19 ; CHECK: select i1 %A, float 3.000000e+00, float [[OP]] 28 ; CHECK: [[OP:%.*]] = fsub float 2.000000e+00, %B 29 ; CHECK: select i1 %A, float 1.000000e+00, float [[OP]] 38 ; CHECK: [[OP:%.*]] = fmul float %B, 2.000000e+00 39 ; CHECK: select i1 %A, float 2.000000e+00, float [[OP]] 48 ; CHECK: [[OP:%.*]] = fdiv float 2.000000e+00, %B 49 ; CHECK: select i1 %A, float 2.000000e+00, float [[OP]] 58 ; CHECK: [[OP:%.*]] = fmul float %B, 5.000000e-01 59 ; CHECK: select i1 %A, float 5.000000e-01, float [[OP]] [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/InstCombine/ |
D | fold-fops-into-selects.ll | 18 ; CHECK: [[OP:%.*]] = fadd float %B, 2.000000e+00 19 ; CHECK: select i1 %A, float 3.000000e+00, float [[OP]] 28 ; CHECK: [[OP:%.*]] = fsub float 2.000000e+00, %B 29 ; CHECK: select i1 %A, float 1.000000e+00, float [[OP]] 38 ; CHECK: [[OP:%.*]] = fmul float %B, 2.000000e+00 39 ; CHECK: select i1 %A, float 2.000000e+00, float [[OP]] 48 ; CHECK: [[OP:%.*]] = fdiv float 2.000000e+00, %B 49 ; CHECK: select i1 %A, float 2.000000e+00, float [[OP]] 58 ; CHECK: [[OP:%.*]] = fmul float %B, 5.000000e-01 59 ; CHECK: select i1 %A, float 5.000000e-01, float [[OP]] [all …]
|
/external/tensorflow/tensorflow/core/kernels/ |
D | cwise_ops_common.h | 543 #define REGISTER(OP, D, N, F, T) \ 545 OP<D##Device, F<T>>); 547 #define REGISTER_VARIANT(OP, D, N, ENUM) \ 550 OP<D##Device, ENUM>); 562 #define REGISTER2(OP, D, N, F, T0, T1) REGISTER(OP, D, N, F, T0) 563 #define REGISTER3(OP, D, N, F, T0, T1, T2) REGISTER(OP, D, N, F, T0) 564 #define REGISTER4(OP, D, N, F, T0, T1, T2, T3) REGISTER(OP, D, N, F, T0) 565 #define REGISTER5(OP, D, N, F, T0, T1, T2, T3, T4) REGISTER(OP, D, N, F, T0) 566 #define REGISTER6(OP, D, N, F, T0, T1, T2, T3, T4, T5) REGISTER(OP, D, N, F, T0) 567 #define REGISTER7(OP, D, N, F, T0, T1, T2, T3, T4, T5, T6) \ [all …]
|
D | spacetobatch_benchmark_test.cc | 59 #define BM_SpaceToBatchDev(OP, DEVICE, DTYPE, B, H, W, D, BS, P00, P01, P10, … argument 62 …BM_##OP##_##DEVICE##_##DTYPE##_##B##_##H##_##W##_##D##_bs##BS##_pad##P00##_##P01##_##P10##_##P11( \ 67 … #OP, TensorShape({B, H, W, D}), BS, DTYPE, \ 72 … BM_##OP##_##DEVICE##_##DTYPE##_##B##_##H##_##W##_##D##_bs##BS##_pad##P00##_##P01##_##P10##_##P11); 73 #define BM_SpaceToBatch(OP, ...) \ argument 74 BM_Expand(BM_SpaceToBatchDev(OP, cpu, DT_FLOAT, __VA_ARGS__)); \ 75 BM_Expand(BM_SpaceToBatchDev(OP, gpu, DT_FLOAT, __VA_ARGS__)); \ 76 BM_Expand(BM_SpaceToBatchDev(OP, cpu, DT_HALF, __VA_ARGS__)); \ 77 BM_Expand(BM_SpaceToBatchDev(OP, gpu, DT_HALF, __VA_ARGS__));
|
/external/eigen/unsupported/Eigen/src/Eigenvalues/ |
D | ArpackSelfAdjointEigenSolver.h | 34 template<typename MatrixSolver, typename MatrixType, typename Scalar, bool BisSPD> struct OP; 472 MatrixSolver OP; in compute() local 476 OP.compute(B); in compute() 482 OP.compute(A); in compute() 494 OP.compute(AminusSigmaB); in compute() 499 OP.compute(AminusSigmaB); in compute() 504 if (!(mode == 1 && isBempty) && !(mode == 2 && isBempty) && OP.info() != Success) in compute() 541 internal::OP<MatrixSolver, MatrixType, Scalar, BisSPD>::applyOP(OP, A, n, in, out); in compute() 551 Matrix<Scalar, Dynamic, 1>::Map(out, n) = OP.solve(Matrix<Scalar, Dynamic, 1>::Map(in, n)); in compute() 559 … Matrix<Scalar, Dynamic, 1>::Map(out, n) = OP.solve(Matrix<Scalar, Dynamic, 1>::Map(in, n)); in compute() [all …]
|
/external/iproute2/include/ |
D | bpf_util.h | 70 #define BPF_ALU64_REG(OP, DST, SRC) \ argument 72 .code = BPF_ALU64 | BPF_OP(OP) | BPF_X, \ 78 #define BPF_ALU32_REG(OP, DST, SRC) \ argument 80 .code = BPF_ALU | BPF_OP(OP) | BPF_X, \ 88 #define BPF_ALU64_IMM(OP, DST, IMM) \ argument 90 .code = BPF_ALU64 | BPF_OP(OP) | BPF_K, \ 96 #define BPF_ALU32_IMM(OP, DST, IMM) \ argument 98 .code = BPF_ALU | BPF_OP(OP) | BPF_K, \ 209 #define BPF_JMP_REG(OP, DST, SRC, OFF) \ argument 211 .code = BPF_JMP | BPF_OP(OP) | BPF_X, \ [all …]
|
/external/v8/src/base/ |
D | safe_math.h | 198 #define BASE_NUMERIC_ARITHMETIC_OPERATORS(NAME, OP, COMPOUND_OP) \ argument 201 CheckedNumeric<typename ArithmeticPromotion<T>::type> operator OP( \ 206 return CheckedNumeric<T>(lhs.ValueUnsafe() OP rhs.ValueUnsafe()); \ 209 lhs.ValueUnsafe() OP rhs.ValueUnsafe(), \ 223 *this = CheckedNumeric<T>::cast(*this) OP CheckedNumeric<Src>::cast(rhs); \ 228 CheckedNumeric<typename ArithmeticPromotion<T, Src>::type> operator OP( \ 233 lhs.ValueUnsafe() OP rhs.ValueUnsafe(), \ 236 OP CheckedNumeric<Promotion>::cast(rhs); \ 240 CheckedNumeric<typename ArithmeticPromotion<T, Src>::type> operator OP( \ 244 return CheckedNumeric<Promotion>(lhs.ValueUnsafe() OP rhs, \ [all …]
|
/external/bcc/src/cc/ |
D | libbpf.h | 127 #define BPF_ALU64_REG(OP, DST, SRC) \ argument 129 .code = BPF_ALU64 | BPF_OP(OP) | BPF_X, \ 135 #define BPF_ALU32_REG(OP, DST, SRC) \ argument 137 .code = BPF_ALU | BPF_OP(OP) | BPF_X, \ 145 #define BPF_ALU64_IMM(OP, DST, IMM) \ argument 147 .code = BPF_ALU64 | BPF_OP(OP) | BPF_K, \ 153 #define BPF_ALU32_IMM(OP, DST, IMM) \ argument 155 .code = BPF_ALU | BPF_OP(OP) | BPF_K, \ 248 #define BPF_JMP_REG(OP, DST, SRC, OFF) \ argument 250 .code = BPF_JMP | BPF_OP(OP) | BPF_X, \ [all …]
|
/external/bcc/src/cc/includes/ |
D | libbpf.h | 127 #define BPF_ALU64_REG(OP, DST, SRC) \ argument 129 .code = BPF_ALU64 | BPF_OP(OP) | BPF_X, \ 135 #define BPF_ALU32_REG(OP, DST, SRC) \ argument 137 .code = BPF_ALU | BPF_OP(OP) | BPF_X, \ 145 #define BPF_ALU64_IMM(OP, DST, IMM) \ argument 147 .code = BPF_ALU64 | BPF_OP(OP) | BPF_K, \ 153 #define BPF_ALU32_IMM(OP, DST, IMM) \ argument 155 .code = BPF_ALU | BPF_OP(OP) | BPF_K, \ 248 #define BPF_JMP_REG(OP, DST, SRC, OFF) \ argument 250 .code = BPF_JMP | BPF_OP(OP) | BPF_X, \ [all …]
|
/external/openssh/ |
D | ttymodes.c | 320 #define TTYCHAR(NAME, OP) \ in tty_make_modes() argument 321 buffer_put_char(&buf, OP); \ in tty_make_modes() 324 #define TTYMODE(NAME, FIELD, OP) \ in tty_make_modes() argument 325 buffer_put_char(&buf, OP); \ in tty_make_modes() 405 #define TTYCHAR(NAME, OP) \ in tty_parse_modes() argument 406 case OP: \ in tty_parse_modes() 410 #define TTYMODE(NAME, FIELD, OP) \ in tty_parse_modes() argument 411 case OP: \ in tty_parse_modes()
|
/external/eigen/Eigen/src/plugins/ |
D | ArrayCwiseBinaryOps.h | 108 #define EIGEN_MAKE_CWISE_COMP_OP(OP, COMPARATOR) \ argument 111 OP(const EIGEN_CURRENT_STORAGE_BASE_CLASS<OtherDerived> &other) const \ 118 OP(const Scalar& s) const { \ 119 return this->OP(Derived::PlainObject::Constant(rows(), cols(), s)); \ 122 OP(const Scalar& s, const Derived& d) { \ 123 return Derived::PlainObject::Constant(d.rows(), d.cols(), s).OP(d); \ 126 #define EIGEN_MAKE_CWISE_COMP_R_OP(OP, R_OP, RCOMPARATOR) \ argument 129 OP(const EIGEN_CURRENT_STORAGE_BASE_CLASS<OtherDerived> &other) const \ 135 OP(const Scalar& s) const { \ 139 OP(const Scalar& s, const Derived& d) { \
|
/external/eigen/blas/ |
D | level3_impl.h | 48 if(OP(*opa)==INVALID) info = 1; in EIGEN_BLAS_FUNC() 49 else if(OP(*opb)==INVALID) info = 2; in EIGEN_BLAS_FUNC() 53 else if(*lda<std::max(1,(OP(*opa)==NOTR)?*m:*k)) info = 8; in EIGEN_BLAS_FUNC() 54 else if(*ldb<std::max(1,(OP(*opb)==NOTR)?*k:*n)) info = 10; in EIGEN_BLAS_FUNC() 73 int code = OP(*opa) | (OP(*opb) << 2); in EIGEN_BLAS_FUNC() 149 else if(OP(*opa)==INVALID) info = 3; in EIGEN_BLAS_FUNC() 161 int code = OP(*opa) | (SIDE(*side) << 2) | (UPLO(*uplo) << 3) | (DIAG(*diag) << 4); in EIGEN_BLAS_FUNC() 254 else if(OP(*opa)==INVALID) info = 3; in EIGEN_BLAS_FUNC() 263 int code = OP(*opa) | (SIDE(*side) << 2) | (UPLO(*uplo) << 3) | (DIAG(*diag) << 4); in EIGEN_BLAS_FUNC() 388 else if(OP(*op)==INVALID || (ISCOMPLEX && OP(*op)==ADJ) ) info = 2; in EIGEN_BLAS_FUNC() [all …]
|
D | level2_impl.h | 48 if(OP(*opa)==INVALID) info = 1; in EIGEN_BLAS_FUNC() 62 int code = OP(*opa); in EIGEN_BLAS_FUNC() 125 else if(OP(*opa)==INVALID) info = 2; in EIGEN_BLAS_FUNC() 135 int code = OP(*opa) | (UPLO(*uplo) << 2) | (DIAG(*diag) << 3); in EIGEN_BLAS_FUNC() 184 else if(OP(*opa)==INVALID) info = 2; in EIGEN_BLAS_FUNC() 199 int code = OP(*opa) | (UPLO(*uplo) << 2) | (DIAG(*diag) << 3); in EIGEN_BLAS_FUNC() 229 if(OP(*trans)==INVALID) info = 1; in EIGEN_BLAS_FUNC() 245 if(OP(*trans)!=NOTR) in EIGEN_BLAS_FUNC() 266 if(OP(*trans)==NOTR) in EIGEN_BLAS_FUNC() 268 else if(OP(*trans)==TR) in EIGEN_BLAS_FUNC() [all …]
|
/external/clang/test/SemaCXX/ |
D | warn-self-assign.cpp | 20 #define OP = in false_positives() macro 25 a OP a; in false_positives() 28 LHS OP RHS; in false_positives() 29 #undef OP in false_positives()
|
/external/eigen/test/ |
D | exceptions.cpp | 80 #define CHECK_MEMLEAK(OP) { \ argument 84 try { OP; } \ 87 …Y(ScalarWithExceptions::instances==before && "memory leak detected in " && EIGEN_MAKESTRING(OP)); \ 89 VERIFY(exception_thrown && " no exception thrown in " && EIGEN_MAKESTRING(OP)); \ 107 …ScalarWithExceptions::instances==0 && "global memory leak detected in " && EIGEN_MAKESTRING(OP)); \ in memoryleak()
|
/external/swiftshader/third_party/LLVM/test/CodeGen/Thumb2/ |
D | 2009-11-11-ScavengerAssert.ll | 3 %struct.OP = type { %struct.OP*, %struct.OP*, %struct.OP* ()*, i32, i16, i16, i8, i8 } 8 define %struct.OP* @Perl_pp_complement() nounwind { 80 ret %struct.OP* undef
|