/external/tensorflow/tensorflow/lite/delegates/gpu/common/ |
D | convert.cc | 255 const int padded_size = num_pixels * num_planes * kPhwc4ChannelsInPlane; in ConvertToPHWC4() local 265 float* dest = out.data() + b * padded_size + in ConvertToPHWC4() 307 const int padded_size = num_pixels * num_planes * kPhwc4ChannelsInPlane; in ConvertToPHWC4Half() local 318 HalfBits* dest = out.data() + b * padded_size + in ConvertToPHWC4Half() 412 const int padded_size = num_pixels * num_planes * kPhwc4ChannelsInPlane; in ConvertFromPHWC4() local 416 const float* src = in.data() + b * padded_size; in ConvertFromPHWC4() 435 const float* src = in.data() + b * padded_size + in ConvertFromPHWC4() 454 const int padded_size = num_pixels * num_planes * kPhwc4ChannelsInPlane; in ConvertFromPHWC4Half() local 458 const HalfBits* src = in.data() + b * padded_size; in ConvertFromPHWC4Half() 480 const HalfBits* src = in.data() + b * padded_size + in ConvertFromPHWC4Half()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | spacetobatch_op.cc | 155 const int64 padded_size = input_size + pad_start + pad_end; in SpaceToBatchOpCompute() local 156 if (padded_size % block_shape_value != 0) { in SpaceToBatchOpCompute() 158 "]=", padded_size, in SpaceToBatchOpCompute() 163 const int64 output_size = padded_size / block_shape_value; in SpaceToBatchOpCompute()
|
/external/v8/src/base/platform/ |
D | platform-cygwin.cc | 128 size_t padded_size = size + (alignment - page_size); in Allocate() local 132 base = RandomizedVirtualAlloc(padded_size, flags, protect, hint); in Allocate() 137 CHECK(Free(base, padded_size)); in Allocate()
|
D | platform-win32.cc | 831 size_t padded_size = size + (alignment - page_size); in Allocate() local 835 base = RandomizedVirtualAlloc(padded_size, flags, protect, hint); in Allocate() 840 CHECK(Free(base, padded_size)); in Allocate()
|
/external/u-boot/tools/ |
D | rkcommon.c | 470 const char *file, int padded_size) in copy_file() argument 495 (int)sbuf.st_size, padded_size); in copy_file() 513 return pad_file(params, ifd, padded_size - size); in copy_file()
|
/external/tensorflow/tensorflow/python/tpu/ |
D | feature_column_v2.py | 810 def pad_sparse_embedding_lookup_indices(sparse_indices, padded_size): argument 833 [batch_size, padded_size]) 839 shape=(batch_size, padded_size)) 843 indices, weights, shape=(batch_size, padded_size))
|
/external/harfbuzz_ng/src/ |
D | hb-ot-glyf-table.hh | 134 | hb_map ([=, &offset] (unsigned int padded_size) in _write_loca() 136 offset += padded_size; in _write_loca() 172 | hb_map (&SubsetGlyph::padded_size) in subset() 1097 unsigned int padded_size () const { return length () + padding (); } in padded_size() function
|
/external/tensorflow/tensorflow/core/ops/ |
D | array_ops.cc | 2235 DimensionHandle padded_size; in SpaceToBatchShapeHelper() local 2237 c->Add(c->Dim(input_shape, dim + 1), pad_start, &padded_size)); in SpaceToBatchShapeHelper() 2238 TF_RETURN_IF_ERROR(c->Add(padded_size, pad_end, &padded_size)); in SpaceToBatchShapeHelper() 2239 TF_RETURN_IF_ERROR(c->Divide(padded_size, block_shape_vec[dim], in SpaceToBatchShapeHelper()
|
/external/avb/ |
D | aftltool | 1521 padded_size = avbtool.round_to_multiple(descriptor_size, padding_size) 1522 padding_needed = padded_size - descriptor_size
|
D | aftltool.py | 1521 padded_size = avbtool.round_to_multiple(descriptor_size, padding_size) 1522 padding_needed = padded_size - descriptor_size
|
D | avbtool.py | 2140 padded_size = round_to_multiple(len(vbmeta_blob), padding_size) 2141 padding_needed = padded_size - len(vbmeta_blob) 2809 padded_size = round_to_multiple(len(vbmeta_blob), padding_size) 2810 padding_needed = padded_size - len(vbmeta_blob)
|
D | avbtool | 2140 padded_size = round_to_multiple(len(vbmeta_blob), padding_size) 2141 padding_needed = padded_size - len(vbmeta_blob) 2809 padded_size = round_to_multiple(len(vbmeta_blob), padding_size) 2810 padding_needed = padded_size - len(vbmeta_blob)
|