/external/tensorflow/tensorflow/core/kernels/ |
D | maxpooling_op_gpu.cu.cc | 72 const int pad_l, dtype* __restrict__ top_data, int64* __restrict__ mask, in MaxPoolForwardNCHW() argument 80 int wstart = pw * stride_w - pad_l; in MaxPoolForwardNCHW() 115 const int stride_h, const int stride_w, const int pad_t, const int pad_l, in MaxPoolForwardNoMaskKernel_NCHW_VECT_C() argument 125 int wstart = pw * stride_w - pad_l; in MaxPoolForwardNoMaskKernel_NCHW_VECT_C() 148 const int stride_h, const int stride_w, const int pad_t, const int pad_l, in MaxPoolForwardNHWC() argument 155 int wstart = (n % pooled_width) * stride_w - pad_l; in MaxPoolForwardNHWC() 235 const int stride_w, const int pad_t, const int pad_l, in MaxPoolGradBackwardNoMaskNCHW() argument 244 int wstart = pw * stride_w - pad_l; in MaxPoolGradBackwardNoMaskNCHW() 276 const int stride_w, const int pad_t, const int pad_l, in MaxPoolGradBackwardNoMaskNHWC() argument 283 int wstart = (n % pooled_width) * stride_w - pad_l; in MaxPoolGradBackwardNoMaskNHWC() [all …]
|
D | avgpooling_op_gpu.cu.cc | 48 const int pad_l, dtype* const __restrict__ bottom_diff) { in DEFINE_GPU_KERNELS() 53 const int w = index / channels % width + pad_l; in DEFINE_GPU_KERNELS() 67 int wstart = pw * stride_w - pad_l; in DEFINE_GPU_KERNELS() 88 const int pad_l, T* const bottom_diff, in RunAvePoolBackwardNHWC() argument 105 const int stride_h, const int stride_w, const int pad_t, const int pad_l, 111 const int stride_h, const int stride_w, const int pad_t, const int pad_l, 117 const int stride_h, const int stride_w, const int pad_t, const int pad_l,
|
D | pooling_ops_3d_gpu.cu.cc | 37 const int pad_l, const dtype* __restrict__ top_diff, in MaxPoolGradBackwardNoMaskNCDHW() argument 48 int wstart = pw * stride_w - pad_l; in MaxPoolGradBackwardNoMaskNCDHW() 88 const int pad_l, const dtype* __restrict__ top_diff, in MaxPoolGradBackwardNoMaskNDHWC() argument 95 int wstart = (n % pooled_width) * stride_w - pad_l; in MaxPoolGradBackwardNoMaskNDHWC() 143 const int stride_w, const int pad_p, const int pad_t, const int pad_l, in operator ()() argument 154 stride_w, pad_p, pad_t, pad_l, top_diff, bottom_diff)); in operator ()() 161 stride_w, pad_p, pad_t, pad_l, top_diff, bottom_diff)); in operator ()()
|
D | maxpooling_op_gpu.h | 41 const int pad_t, const int pad_l, T* top_data, int64_t* mask, 51 const int pad_t, const int pad_l, int32* top_data, 78 const int stride_w, const int pad_t, const int pad_l,
|
D | pooling_ops_3d_gpu.h | 41 const int pad_l, const T* top_diff, T* bottom_diff,
|
D | avgpooling_op.h | 71 const int pad_l, T* const bottom_diff,
|
D | conv_grad_input_ops.h | 80 const int pad_t, const int pad_l, const int pad_b, const int pad_r, in Col2im() argument 83 int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1; in Col2im() 86 int w_pad = -pad_l; in Col2im()
|
D | conv_grad_ops_3d.cc | 81 const int pad_t, const int pad_l, const int pad_pb, const int pad_b, in Col2im() argument 86 const int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1; in Col2im() 91 int w_pad = -pad_l; in Col2im() 130 const int pad_t, const int pad_l, const int pad_pb, const int pad_b, in Im2col() argument 135 const int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1; in Im2col() 141 int w_pad = -pad_l; in Im2col()
|
D | conv_grad_filter_ops.cc | 76 const int pad_t, const int pad_l, const int pad_b, const int pad_r, in Im2col() argument 79 int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1; in Im2col() 83 int w_pad = -pad_l; in Im2col()
|
/external/tensorflow/tensorflow/core/kernels/mkl/ |
D | mkl_conv_ops.h | 348 memory::dims* pad_l, memory::dims* pad_r, bool is_grouped_convolution, 352 DCHECK(pad_l); 432 pad_top = static_cast<int64_t>((*pad_l)[0]); 433 pad_left = static_cast<int64_t>((*pad_l)[1]); 451 pad_front = static_cast<int64>((*pad_l)[0]); 452 pad_top = static_cast<int64>((*pad_l)[1]); 453 pad_left = static_cast<int64>((*pad_l)[2]); 479 *pad_l = {static_cast<int>(pad_top), static_cast<int>(pad_left)}; 487 *pad_l = {static_cast<int>(pad_front), static_cast<int>(pad_top), 542 memory::dims* output_dims_mkl_order, memory::dims* pad_l, in GetOutputAndPadSizeInMklOrder() argument [all …]
|
/external/ComputeLibrary/src/cpu/kernels/pool2d/neon/nchw/ |
D | all.cpp | 59 float16x4_t read_4_boundary_aware_fp16(int srcw, int srch, int pad_l, int pad_t, int x, int y, cons… in read_4_boundary_aware_fp16() argument 65 if(row_in_bounds && (x + i >= pad_l) && (x + i < (srcw + pad_l))) in read_4_boundary_aware_fp16() 170 auto read_2_boundary_aware(int srcw, int srch, int pad_l, int pad_t, int x, int y, const T *ptr, T … in read_2_boundary_aware() argument 176 if(row_in_bounds && (x + i >= pad_l) && (x + i < (srcw + pad_l))) in read_2_boundary_aware()
|
/external/tensorflow/tensorflow/security/advisory/ |
D | tfsa-2021-077.md | 13 int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1;
|
/external/ComputeLibrary/src/cpu/kernels/pool2d/neon/ |
D | quantized.h | 329 auto load16_boundary_aware(int srcw, int srch, int pad_l, int pad_r, int pad_t, int pad_b, int x, i… in load16_boundary_aware() argument 337 if(row_in_bounds && (x + i >= pad_l) && (x + i < (srcw + pad_l))) in load16_boundary_aware()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | optimized_ops.h | 4902 const int pad_t, const int pad_l, const int pad_b, const int pad_r, in Col2im() argument 4906 int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1; in Col2im() 4909 int w_pad = -pad_l; in Col2im() 8150 const int pad_t, const int pad_l, const int pad_pb, const int pad_b, in Col2im() argument 8155 const int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1; in Col2im() 8160 int w_pad = -pad_l; in Col2im()
|