/external/tensorflow/tensorflow/lite/kernels/internal/ |
D | tensor_utils.h | 142 inline void ApplyReluToVector(const float* __restrict__ vector, int v_size, in ApplyReluToVector() 150 inline void ApplyRelu1ToVector(const float* __restrict__ vector, int v_size, in ApplyRelu1ToVector() 158 inline void ApplyRelu6ToVector(const float* __restrict__ vector, int v_size, in ApplyRelu6ToVector() 166 inline void ApplyTanhToVector(const float* __restrict__ vector, int v_size, in ApplyTanhToVector() 175 inline void ApplySignbitToVector(const float* __restrict__ vector, int v_size, in ApplySignbitToVector() 183 inline void ApplySigmoidToVector(const float* __restrict__ vector, int v_size, in ApplySigmoidToVector() 193 int v_size, in ApplyActivationToVector()
|
D | tensor_utils_common.h | 313 const T* __restrict__ vector2, int v_size, in VectorVectorCwiseProduct() 325 int v_size, in VectorVectorCwiseProductAccumulate() 352 int v_size, int n_batch, in BatchVectorBatchVectorDotProduct() 368 inline void VectorBatchVectorCwiseProduct(const T* vector, int v_size, in VectorBatchVectorCwiseProduct() 383 inline void VectorBatchVectorCwiseProductAccumulate(const T* vector, int v_size, in VectorBatchVectorCwiseProductAccumulate() 402 void VectorBatchVectorAdd(const T* vector, int v_size, int n_batch, in VectorBatchVectorAdd() 414 void VectorBatchVectorAssign(const T* vector, int v_size, int n_batch, in VectorBatchVectorAssign()
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
D | portable_tensor_utils_impl.h | 34 bool PortableIsZeroVector(const T* vector, int v_size) { in PortableIsZeroVector() 176 void PortableCwiseClipping(T* vector, const int v_size, in PortableCwiseClipping()
|
D | portable_tensor_utils.h | 28 bool IsZeroVector(const float* vector, int v_size) { in IsZeroVector() 33 bool IsZeroVector(const int8_t* vector, int v_size) { in IsZeroVector() 230 void CwiseClipping(float* vector, const int v_size, in CwiseClipping() 235 void CwiseClipping(int16_t* vector, const int v_size, in CwiseClipping() 240 void CwiseClipping(int8_t* vector, const int v_size, in CwiseClipping() 245 void VectorBatchVectorCwiseProductAccumulate(const int16_t* vector, int v_size, in VectorBatchVectorCwiseProductAccumulate() 254 int v_size) { in VectorVectorDotProduct() 259 const int16_t* vector2, int v_size, in BatchVectorBatchVectorDotProduct() 265 void Sub1Vector(const float* vector, int v_size, float* result) { in Sub1Vector() 269 void Sub1Vector(const int16_t* vector, int v_size, int16_t* result) { in Sub1Vector() [all …]
|
D | portable_tensor_utils.cc | 653 int v_size) { in PortableVectorVectorDotProduct() 663 const int16_t* vector2, int v_size) { in VectorVectorDotProduct() 674 int v_size, int n_batch, in PortableBatchVectorBatchVectorDotProduct() 684 const int16_t* vector, int v_size, const int16_t* batch_vector, int n_batch, in PortableVectorBatchVectorCwiseProductAccumulate() 698 void PortableSub1Vector(const float* vector, int v_size, float* result) { in PortableSub1Vector() 704 void PortableSub1Vector(const int16_t* vector, int v_size, int16_t* result) { in PortableSub1Vector() 711 void PortableVectorScalarMultiply(const int8_t* vector, const int v_size, in PortableVectorScalarMultiply() 720 int v_size, int n_batch) { in PortableMeanStddevNormalization()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | sse_tensor_utils.h | 206 void CwiseClipping(float* vector, const int v_size, in CwiseClipping() 211 void CwiseClipping(int16_t* vector, const int v_size, in CwiseClipping() 216 void CwiseClipping(int8_t* vector, const int v_size, in CwiseClipping() 222 const int16_t* vector2, int v_size, in BatchVectorBatchVectorDotProduct() 228 void VectorBatchVectorCwiseProductAccumulate(const int16_t* vector, int v_size, in VectorBatchVectorCwiseProductAccumulate() 237 int v_size) { in VectorVectorDotProduct() 241 void Sub1Vector(const float* vector, int v_size, float* result) { in Sub1Vector() 245 void Sub1Vector(const int16_t* vector, int v_size, int16_t* result) { in Sub1Vector() 250 bool IsZeroVector(const float* vector, int v_size) { in IsZeroVector() 255 bool IsZeroVector(const int8_t* vector, int v_size) { in IsZeroVector() [all …]
|
D | neon_tensor_utils.h | 198 void CwiseClipping(float* vector, const int v_size, in CwiseClipping() 202 void CwiseClipping(int16_t* vector, const int v_size, in CwiseClipping() 206 void CwiseClipping(int8_t* vector, const int v_size, in CwiseClipping() 212 const int16_t* vector2, int v_size, in BatchVectorBatchVectorDotProduct() 218 void VectorBatchVectorCwiseProductAccumulate(const int16_t* vector, int v_size, in VectorBatchVectorCwiseProductAccumulate() 227 int v_size) { in VectorVectorDotProduct() 231 void Sub1Vector(const float* vector, int v_size, float* result) { in Sub1Vector() 235 void Sub1Vector(const int16_t* vector, int v_size, int16_t* result) { in Sub1Vector() 240 bool IsZeroVector(const float* vector, int v_size) { in IsZeroVector() 245 bool IsZeroVector(const int8_t* vector, int v_size) { in IsZeroVector() [all …]
|
D | neon_tensor_utils.cc | 1838 void NeonCwiseClipping(float* vector, const int v_size, in NeonCwiseClipping() 1859 void NeonCwiseClipping(int16_t* vector, const int v_size, in NeonCwiseClipping() 1882 void NeonCwiseClipping(int8_t* vector, const int v_size, in NeonCwiseClipping() 2047 void NeonSub1Vector(const float* vector, int v_size, float* result) { in NeonSub1Vector() 2069 void NeonSub1Vector(const int16_t* vector, int v_size, int16_t* result) { in NeonSub1Vector() 2123 bool NeonIsZeroVector(const float* vector, int v_size) { in NeonIsZeroVector() 2142 bool NeonIsZeroVector(const int8_t* vector, int v_size) { in NeonIsZeroVector() 2161 void NeonVectorScalarMultiply(const int8_t* vector, const int v_size, in NeonVectorScalarMultiply() 2443 int v_size) { in NeonVectorVectorDotProduct() 2518 const int16_t* vector, int v_size, const int16_t* batch_vector, int n_batch, in NeonVectorBatchVectorCwiseProductAccumulate() [all …]
|
/external/webp/tests/fuzzer/ |
D | simple_api_fuzzer.c | 69 size_t v_size = uv_stride * (h + 1) / 2; in LLVMFuzzerTestOneInput() local
|
/external/drm_hwcomposer/bufferinfo/legacy/ |
D | BufferInfoMaliHisi.cpp | 107 int v_size = vu_stride * (adjusted_height / 2); in ConvertBoInfo() local
|
/external/webp/src/dec/ |
D | buffer_dec.c | 58 const uint64_t v_size = MIN_BUFFER_SIZE(uv_width, uv_height, v_stride); in CheckDecBuffer() local
|
D | idec_dec.c | 718 uint8_t* v, size_t v_size, int v_stride, in WebPINewYUVA() 761 uint8_t* v, size_t v_size, int v_stride) { in WebPINewYUV()
|
D | webp_dec.c | 578 uint8_t* v, size_t v_size, int v_stride) { in WebPDecodeYUVInto()
|
/external/webp/include/webp/ |
D | decode.h | 195 size_t u_size, v_size; // chroma planes size member
|
/external/kernel-headers/original/uapi/linux/ |
D | nilfs2_api.h | 165 __u16 v_size; /* size of members */ member
|
D | rkisp1-config.h | 208 __u16 v_size; member
|
/external/webp/src/webp/ |
D | decode.h | 195 size_t u_size, v_size; // chroma planes size member
|
/external/brotli/c/enc/ |
D | brotli_bit_stream.c | 578 static size_t IndexOf(const uint8_t* v, size_t v_size, uint8_t value) { in IndexOf() 596 const size_t v_size, in MoveToFrontTransform()
|
/external/tensorflow/tensorflow/core/grappler/utils/ |
D | graph_view_internal.h | 451 int v_size = v->size(); in AddOrUpdateAtIndex() local
|