Home
last modified time | relevance | path

Searched refs:pad_l (Results 1 – 14 of 14) sorted by relevance

/external/tensorflow/tensorflow/core/kernels/
Dmaxpooling_op_gpu.cu.cc72 const int pad_l, dtype* __restrict__ top_data, int64* __restrict__ mask, in MaxPoolForwardNCHW() argument
80 int wstart = pw * stride_w - pad_l; in MaxPoolForwardNCHW()
115 const int stride_h, const int stride_w, const int pad_t, const int pad_l, in MaxPoolForwardNoMaskKernel_NCHW_VECT_C() argument
125 int wstart = pw * stride_w - pad_l; in MaxPoolForwardNoMaskKernel_NCHW_VECT_C()
148 const int stride_h, const int stride_w, const int pad_t, const int pad_l, in MaxPoolForwardNHWC() argument
155 int wstart = (n % pooled_width) * stride_w - pad_l; in MaxPoolForwardNHWC()
235 const int stride_w, const int pad_t, const int pad_l, in MaxPoolGradBackwardNoMaskNCHW() argument
244 int wstart = pw * stride_w - pad_l; in MaxPoolGradBackwardNoMaskNCHW()
276 const int stride_w, const int pad_t, const int pad_l, in MaxPoolGradBackwardNoMaskNHWC() argument
283 int wstart = (n % pooled_width) * stride_w - pad_l; in MaxPoolGradBackwardNoMaskNHWC()
[all …]
Davgpooling_op_gpu.cu.cc48 const int pad_l, dtype* const __restrict__ bottom_diff) { in DEFINE_GPU_KERNELS()
53 const int w = index / channels % width + pad_l; in DEFINE_GPU_KERNELS()
67 int wstart = pw * stride_w - pad_l; in DEFINE_GPU_KERNELS()
88 const int pad_l, T* const bottom_diff, in RunAvePoolBackwardNHWC() argument
105 const int stride_h, const int stride_w, const int pad_t, const int pad_l,
111 const int stride_h, const int stride_w, const int pad_t, const int pad_l,
117 const int stride_h, const int stride_w, const int pad_t, const int pad_l,
Dpooling_ops_3d_gpu.cu.cc37 const int pad_l, const dtype* __restrict__ top_diff, in MaxPoolGradBackwardNoMaskNCDHW() argument
48 int wstart = pw * stride_w - pad_l; in MaxPoolGradBackwardNoMaskNCDHW()
88 const int pad_l, const dtype* __restrict__ top_diff, in MaxPoolGradBackwardNoMaskNDHWC() argument
95 int wstart = (n % pooled_width) * stride_w - pad_l; in MaxPoolGradBackwardNoMaskNDHWC()
143 const int stride_w, const int pad_p, const int pad_t, const int pad_l, in operator ()() argument
154 stride_w, pad_p, pad_t, pad_l, top_diff, bottom_diff)); in operator ()()
161 stride_w, pad_p, pad_t, pad_l, top_diff, bottom_diff)); in operator ()()
Dmaxpooling_op_gpu.h41 const int pad_t, const int pad_l, T* top_data, int64_t* mask,
51 const int pad_t, const int pad_l, int32* top_data,
78 const int stride_w, const int pad_t, const int pad_l,
Dpooling_ops_3d_gpu.h41 const int pad_l, const T* top_diff, T* bottom_diff,
Davgpooling_op.h71 const int pad_l, T* const bottom_diff,
Dconv_grad_input_ops.h80 const int pad_t, const int pad_l, const int pad_b, const int pad_r, in Col2im() argument
83 int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1; in Col2im()
86 int w_pad = -pad_l; in Col2im()
Dconv_grad_ops_3d.cc81 const int pad_t, const int pad_l, const int pad_pb, const int pad_b, in Col2im() argument
86 const int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1; in Col2im()
91 int w_pad = -pad_l; in Col2im()
130 const int pad_t, const int pad_l, const int pad_pb, const int pad_b, in Im2col() argument
135 const int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1; in Im2col()
141 int w_pad = -pad_l; in Im2col()
Dconv_grad_filter_ops.cc76 const int pad_t, const int pad_l, const int pad_b, const int pad_r, in Im2col() argument
79 int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1; in Im2col()
83 int w_pad = -pad_l; in Im2col()
/external/tensorflow/tensorflow/core/kernels/mkl/
Dmkl_conv_ops.h348 memory::dims* pad_l, memory::dims* pad_r, bool is_grouped_convolution,
352 DCHECK(pad_l);
432 pad_top = static_cast<int64_t>((*pad_l)[0]);
433 pad_left = static_cast<int64_t>((*pad_l)[1]);
451 pad_front = static_cast<int64>((*pad_l)[0]);
452 pad_top = static_cast<int64>((*pad_l)[1]);
453 pad_left = static_cast<int64>((*pad_l)[2]);
479 *pad_l = {static_cast<int>(pad_top), static_cast<int>(pad_left)};
487 *pad_l = {static_cast<int>(pad_front), static_cast<int>(pad_top),
542 memory::dims* output_dims_mkl_order, memory::dims* pad_l, in GetOutputAndPadSizeInMklOrder() argument
[all …]
/external/ComputeLibrary/src/cpu/kernels/pool2d/neon/nchw/
Dall.cpp59 float16x4_t read_4_boundary_aware_fp16(int srcw, int srch, int pad_l, int pad_t, int x, int y, cons… in read_4_boundary_aware_fp16() argument
65 if(row_in_bounds && (x + i >= pad_l) && (x + i < (srcw + pad_l))) in read_4_boundary_aware_fp16()
170 auto read_2_boundary_aware(int srcw, int srch, int pad_l, int pad_t, int x, int y, const T *ptr, T … in read_2_boundary_aware() argument
176 if(row_in_bounds && (x + i >= pad_l) && (x + i < (srcw + pad_l))) in read_2_boundary_aware()
/external/tensorflow/tensorflow/security/advisory/
Dtfsa-2021-077.md13 int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1;
/external/ComputeLibrary/src/cpu/kernels/pool2d/neon/
Dquantized.h329 auto load16_boundary_aware(int srcw, int srch, int pad_l, int pad_r, int pad_t, int pad_b, int x, i… in load16_boundary_aware() argument
337 if(row_in_bounds && (x + i >= pad_l) && (x + i < (srcw + pad_l))) in load16_boundary_aware()
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/
Doptimized_ops.h4902 const int pad_t, const int pad_l, const int pad_b, const int pad_r, in Col2im() argument
4906 int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1; in Col2im()
4909 int w_pad = -pad_l; in Col2im()
8150 const int pad_t, const int pad_l, const int pad_pb, const int pad_b, in Col2im() argument
8155 const int width_col = (width + pad_l + pad_r - filter_w) / stride_w + 1; in Col2im()
8160 int w_pad = -pad_l; in Col2im()