Searched refs:output_dim_size (Results 1 – 7 of 7) sorted by relevance
34 const int output_dim_size = context->tensors[outputs->data[0]].dims->size; in PopulateSubGraph() local37 concat_axis < 0 ? concat_axis + 4 : concat_axis + 4 - output_dim_size; in PopulateSubGraph()
107 llvm::Value* output_dim_size = llvm::ConstantInt::get( in EmitDynamicUpdateSliceInPlaceImpl() local114 llvm::Value* max_bound = b->CreateSub(output_dim_size, update_dim_size); in EmitDynamicUpdateSliceInPlaceImpl()
988 const int64 output_dim_size = in HandleReshape() local991 << " output_dim_size: " << output_dim_size; in HandleReshape()993 if (input_dim_size == output_dim_size) { in HandleReshape()999 if (input_dim_size > output_dim_size) { in HandleReshape()1000 TF_RET_CHECK(input_dim_size % output_dim_size == 0) in HandleReshape()1002 const int64 divisor = input_dim_size / output_dim_size; in HandleReshape()1016 if (input_dim_size < output_dim_size) { in HandleReshape()1035 LiteralUtil::CreateR0<int32>(output_dim_size))); in HandleReshape()
1776 int64 output_dim_size = in HandleGather() local1783 std::min(operand_shape.dimensions(i) - output_dim_size, in HandleGather()
1976 int64 output_dim_size = in EmitElementalGather() local1979 operand_shape.dimensions(operand_dim) - output_dim_size; in EmitElementalGather()
2044 int64 output_dim_size = hlo->shape().dimensions(output_sharded_dim); in HandleReshape() local2048 if (input_dim_size % output_dim_size == 0) { in HandleReshape()2050 int64 split_factor = input_dim_size / output_dim_size; in HandleReshape()2088 } else if (output_dim_size % input_dim_size == 0) { in HandleReshape()2090 int64 merge_factor = output_dim_size / input_dim_size; in HandleReshape()2118 dim->set_padding_high(output_dim_size - in HandleReshape()
1216 output_dim_size = input_dim_size - kernel_dim_size + 1 (requires: input_dim_size >= kernel_dim_size…