| /third_party/mesa3d/src/amd/vulkan/ |
| D | radv_shader_args.c | 55 add_ud_arg(struct radv_shader_args *args, unsigned size, enum ac_arg_type type, struct ac_arg *arg, in add_ud_arg() argument 58 ac_add_arg(&args->ac, AC_ARG_SGPR, size, type, arg); in add_ud_arg() 60 struct radv_userdata_info *ud_info = &args->user_sgprs_locs.shader_data[ud]; in add_ud_arg() 63 ud_info->sgpr_idx = args->num_user_sgprs; in add_ud_arg() 67 args->num_user_sgprs += size; in add_ud_arg() 71 add_descriptor_set(struct radv_shader_args *args, enum ac_arg_type type, struct ac_arg *arg, uint32… in add_descriptor_set() argument 73 ac_add_arg(&args->ac, AC_ARG_SGPR, 1, type, arg); in add_descriptor_set() 75 struct radv_userdata_info *ud_info = &args->user_sgprs_locs.descriptor_sets[set]; in add_descriptor_set() 76 ud_info->sgpr_idx = args->num_user_sgprs; in add_descriptor_set() 79 args->user_sgprs_locs.descriptor_sets_enabled |= 1u << set; in add_descriptor_set() [all …]
|
| /third_party/mesa3d/src/intel/vulkan/grl/gpu/ |
| D | rebraid.grl | 43 MKRebraidArgs Args 46 dispatch init_scratch(1,1,1) args( Args.rebraid_scratch ); 47 dispatch calc_aabb(Args.num_instances,1,1) args( Args.bvh_buffer, Args.instances_buffer ); 50 …//define num_count_groups ((Args.num_instances + (COUNT_SPLITS_GROUP_SIZE-1)) / COUNT_SPLITS_GROUP… 51 …atch count_splits(num_count_groups,1,1) args( Args.bvh_buffer, Args.instances_buffer, Args.rebraid… 53 …dispatch count_splits_SG(Args.num_instances,1,1) args( Args.bvh_buffer, Args.instances_buffer, Arg… 56 define num_primref_groups ((Args.num_instances + (PRIMREF_GROUP_SIZE-1)) / PRIMREF_GROUP_SIZE); 58 …s,1,1) args( Args.global_buffer, Args.bvh_buffer, Args.instances_buffer, Args.rebraid_scratch, Arg… 61 …UG_PRINT(1,1,1) args( Args.global_buffer, Args.instances_buffer, Args.rebraid_scratch, Args.primre… 64 metakernel rebraid_indirect(MKRebraidArgs Args, qword indirectBuildRangeInfo) [all …]
|
| /third_party/mindspore/mindspore-src/source/mindspore/lite/tools/converter/micro/coder/wrapper/int8/ |
| D | conv1x1_run_int8_wrapper.c | 24 void Pre1x1Trans(Conv1x1Args *args, int8_t *src_input, int8_t *src_output) { in Pre1x1Trans() argument 25 args->output_ptr_ = src_output; in Pre1x1Trans() 26 if (args->pre_trans_input_) { in Pre1x1Trans() 27 Conv1x1InputPack(src_input, args->input_ptr_, args->conv_param_, sizeof(int8_t)); in Pre1x1Trans() 29 args->input_ptr_ = src_input; in Pre1x1Trans() 34 Conv1x1Args *args = (Conv1x1Args *)(cdata); in OcOptPre() local 35 int cur_stride = args->thread_stride_hw_ * C4NUM; in OcOptPre() 36 int res_stride = args->matmul_param_->row_ - task_id * args->thread_stride_hw_ * C4NUM; in OcOptPre() 41 …int8_t *hw_in = args->input_ptr_ + task_id * args->thread_stride_hw_ * C4NUM * args->conv_param_->… in OcOptPre() 42 …int8_t *hw_packed_in = args->packed_input_ + task_id * args->thread_stride_hw_ * C4NUM * args->mat… in OcOptPre() [all …]
|
| D | add_int8_wrapper.c | 21 AddInt8Args *args = (AddInt8Args *)(cdata); in AddBroadcastInt8Run() local 22 int stride = UP_DIV(args->out_size_, args->thread_count_); in AddBroadcastInt8Run() 23 int real_out_count = MSMIN(stride, args->out_size_ - stride * task_id); in AddBroadcastInt8Run() 31 if (args->arith_para_->in_elements_num0_ == args->arith_para_->out_elements_num_) { in AddBroadcastInt8Run() 32 cur_in0 = args->input0_data_ + task_id * stride * args->in_size_ + i * args->in_size_; in AddBroadcastInt8Run() 33 cur_in1 = args->input1_data_; in AddBroadcastInt8Run() 34 cur_out = args->output_data_ + task_id * stride * args->in_size_ + i * args->in_size_; in AddBroadcastInt8Run() 36 cur_in0 = args->input0_data_; in AddBroadcastInt8Run() 37 cur_in1 = args->input1_data_ + task_id * stride * args->in_size_ + i * args->in_size_; in AddBroadcastInt8Run() 38 cur_out = args->output_data_ + task_id * stride * args->in_size_ + i * args->in_size_; in AddBroadcastInt8Run() [all …]
|
| /third_party/mindspore/mindspore-src/source/mindspore/ccsrc/plugin/device/ascend/kernel/aicpu/aicpu_ops/cpu_kernel/ms_kernel/ |
| D | avgpool.cc | 80 uint32_t AvgPoolCpuKernel::AvgPoolProcess(CpuKernelContext &ctx, AvgPoolCalcArgs args) { in AvgPoolProcess() argument 87 if (data_num >= kParallelDataNum && args.batch_size > 1) { in AvgPoolProcess() 96 if (max_core_num > args.batch_size) { in AvgPoolProcess() 97 max_core_num = static_cast<uint32_t>(args.batch_size); in AvgPoolProcess() 101 if (args.data_format == "NCHW") { in AvgPoolProcess() 102 RealComputeNCHW<T>(start, end, args, input0, output0); in AvgPoolProcess() 104 RealComputeNHWC<T>(start, end, args, input0, output0); in AvgPoolProcess() 108 …ctx, CpuKernelUtils::ParallelFor(ctx, args.batch_size, args.batch_size / max_core_num, sharder_avg… in AvgPoolProcess() 111 if (args.data_format == "NCHW") { in AvgPoolProcess() 112 RealComputeNCHW<T>(0, args.batch_size, args, input0, output0); in AvgPoolProcess() [all …]
|
| D | adaptive_max_pool_2d_grad.cc | 52 void ComputeSingleThread(int64_t start, int64_t end, AdaptiveCalcArgs<SCALAR_T, INDICES_T> args) { in ComputeSingleThread() argument 54 SCALAR_T *grad_input_p_d = args.input_grad_data + d * args.in_size_h * args.in_size_w; in ComputeSingleThread() 55 SCALAR_T *grad_output_p_d = args.output_grad_data + d * args.out_size_h * args.out_size_w; in ComputeSingleThread() 56 INDICES_T *ind_p_d = args.indices_data + d * args.out_size_h * args.out_size_w; in ComputeSingleThread() 60 for (oh = 0; oh < args.out_size_h; oh++) { in ComputeSingleThread() 61 for (ow = 0; ow < args.out_size_w; ow++) { in ComputeSingleThread() 63 INDICES_T maxp = ind_p_d[oh * args.out_size_w + ow]; in ComputeSingleThread() 65 grad_input_p_d[maxp] += grad_output_p_d[oh * args.out_size_w + ow]; in ComputeSingleThread() 72 …iveMaxPool2dGradSingleOutFrame(CpuKernelContext &ctx, AdaptiveCalcArgs<SCALAR_T, INDICES_T> args) { in AdaptiveMaxPool2dGradSingleOutFrame() argument 80 …tive_max_pool_2d_grad = [&](int64_t start, int64_t end) { ComputeSingleThread(start, end, args); }; in AdaptiveMaxPool2dGradSingleOutFrame() [all …]
|
| D | adaptive_avg_pool_3d.cc | 71 uint32_t AdaptiveAvgPool3dOutFrame(CpuKernelContext &ctx, AdaptiveCalcArgs<SCALAR_T> args, int64_t … in AdaptiveAvgPool3dOutFrame() argument 75 for (int64_t out_t = 0; out_t < args.out_size_t; out_t++) { in AdaptiveAvgPool3dOutFrame() 76 int in_start_t = StartIndex(out_t, args.out_size_t, args.in_size_t); in AdaptiveAvgPool3dOutFrame() 77 int in_end_t = EndIndex(out_t, args.out_size_t, args.in_size_t); in AdaptiveAvgPool3dOutFrame() 80 for (int64_t out_h = 0; out_h < args.out_size_h; out_h++) { in AdaptiveAvgPool3dOutFrame() 81 int in_start_h = StartIndex(out_h, args.out_size_h, args.in_size_h); in AdaptiveAvgPool3dOutFrame() 82 int in_end_h = EndIndex(out_h, args.out_size_h, args.in_size_h); in AdaptiveAvgPool3dOutFrame() 85 for (int64_t out_w = 0; out_w < args.out_size_w; out_w++) { in AdaptiveAvgPool3dOutFrame() 86 int in_start_w = StartIndex(out_w, args.out_size_w, args.in_size_w); in AdaptiveAvgPool3dOutFrame() 87 int in_end_w = EndIndex(out_w, args.out_size_w, args.in_size_w); in AdaptiveAvgPool3dOutFrame() [all …]
|
| /third_party/mesa3d/.gitlab-ci/bare-metal/ |
| D | mkbootimg.py | 77 def get_recovery_dtbo_offset(args): argument 80 num_kernel_pages = get_number_of_pages(filesize(args.kernel), args.pagesize) 81 num_ramdisk_pages = get_number_of_pages(filesize(args.ramdisk), 82 args.pagesize) 83 num_second_pages = get_number_of_pages(filesize(args.second), args.pagesize) 84 dtbo_offset = args.pagesize * (num_header_pages + num_kernel_pages + 87 def write_header_v3_and_above(args): argument 88 if args.header_version > 3: 92 args.output.write(pack(f'{BOOT_MAGIC_SIZE}s', BOOT_MAGIC.encode())) 94 args.output.write(pack('I', filesize(args.kernel))) [all …]
|
| /third_party/mesa3d/src/amd/common/nir/ |
| D | ac_nir_lower_intrinsics_to_args.c | 13 const struct ac_shader_args *const args; member 34 nir_def *value = ac_nir_load_arg_upper_bound(&start_b, s->args, arg, upper_bound); in preload_arg() 39 nir_def *count = ac_nir_unpack_arg(&start_b, s->args, s->args->merged_wave_info, 8, 8); in preload_arg() 42 ac_nir_load_arg_upper_bound(&start_b, s->args, ls_buggy_arg, upper_bound), in preload_arg() 54 assert(s->gfx_level < GFX12 && s->args->tg_size.used); in load_subgroup_id_lowered() 57 return ac_nir_unpack_arg(b, s->args, s->args->tg_size, 20, 5); in load_subgroup_id_lowered() 63 return ac_nir_unpack_arg(b, s->args, s->args->tg_size, 6, 6); in load_subgroup_id_lowered() 66 assert(s->args->tcs_wave_id.used); in load_subgroup_id_lowered() 67 return ac_nir_unpack_arg(b, s->args, s->args->tcs_wave_id, 0, 3); in load_subgroup_id_lowered() 70 assert(s->args->merged_wave_info.used); in load_subgroup_id_lowered() [all …]
|
| /third_party/python/Objects/clinic/ |
| D | codeobject.c.h | 24 code_new(PyTypeObject *type, PyObject *args, PyObject *kwargs) in code_new() argument 51 if (!_PyArg_CheckPositional("code", PyTuple_GET_SIZE(args), 16, 18)) { in code_new() 54 argcount = _PyLong_AsInt(PyTuple_GET_ITEM(args, 0)); in code_new() 58 posonlyargcount = _PyLong_AsInt(PyTuple_GET_ITEM(args, 1)); in code_new() 62 kwonlyargcount = _PyLong_AsInt(PyTuple_GET_ITEM(args, 2)); in code_new() 66 nlocals = _PyLong_AsInt(PyTuple_GET_ITEM(args, 3)); in code_new() 70 stacksize = _PyLong_AsInt(PyTuple_GET_ITEM(args, 4)); in code_new() 74 flags = _PyLong_AsInt(PyTuple_GET_ITEM(args, 5)); in code_new() 78 if (!PyBytes_Check(PyTuple_GET_ITEM(args, 6))) { in code_new() 79 _PyArg_BadArgument("code", "argument 7", "bytes", PyTuple_GET_ITEM(args, 6)); in code_new() [all …]
|
| /third_party/alsa-lib/src/conf/cards/ |
| D | HDA-Intel.conf | 8 @args [ CARD ] 9 @args.CARD { 33 @args [ CARD ] 34 @args.CARD { 88 @args [ CARD AES0 AES1 AES2 AES3 ] 89 @args.CARD { 92 @args.AES0 { 95 @args.AES1 { 98 @args.AES2 { 101 @args.AES3 { [all …]
|
| /third_party/typescript/tests/baselines/reference/ |
| D | readonlyRestParameters.types | 15 >f1 : (...args: readonly string[]) => void 26 function f1(...args: readonly string[]) { 27 >f1 : (...args: readonly string[]) => void 28 >args : readonly string[] 30 f0(...args); // Error 31 >f0(...args) : void 33 >...args : string 34 >args : readonly string[] 38 >f1 : (...args: readonly string[]) => void 42 f1('abc', ...args); [all …]
|
| D | strictBindCallApply1.types | 22 …args: A) => R, thisArg: T, arg0: A0): (...args: A) => R; <T, A0, A1, A extends any[], R>(this: (th… 24 …args: A) => R, thisArg: T, arg0: A0): (...args: A) => R; <T, A0, A1, A extends any[], R>(this: (th… 30 …args: A) => R, thisArg: T, arg0: A0): (...args: A) => R; <T, A0, A1, A extends any[], R>(this: (th… 32 …args: A) => R, thisArg: T, arg0: A0): (...args: A) => R; <T, A0, A1, A extends any[], R>(this: (th… 39 …args: A) => R, thisArg: T, arg0: A0): (...args: A) => R; <T, A0, A1, A extends any[], R>(this: (th… 41 …args: A) => R, thisArg: T, arg0: A0): (...args: A) => R; <T, A0, A1, A extends any[], R>(this: (th… 49 …args: A) => R, thisArg: T, arg0: A0): (...args: A) => R; <T, A0, A1, A extends any[], R>(this: (th… 51 …args: A) => R, thisArg: T, arg0: A0): (...args: A) => R; <T, A0, A1, A extends any[], R>(this: (th… 59 …args: A) => R, thisArg: T, arg0: A0): (...args: A) => R; <T, A0, A1, A extends any[], R>(this: (th… 61 …args: A) => R, thisArg: T, arg0: A0): (...args: A) => R; <T, A0, A1, A extends any[], R>(this: (th… [all …]
|
| D | baseConstraintOfDecorator.types | 2 …nder<TFunction>(superClass: TFunction, _instanceModifier: (instance: any, args: any[]) => void): T… 3 >classExtender : <TFunction>(superClass: TFunction, _instanceModifier: (instance: any, args: any[])… 5 >_instanceModifier : (instance: any, args: any[]) => void 7 >args : any[] 10 … superClass { constructor(...args: any[]) { super(...args); _instance… 14 constructor(...args: any[]) { 15 >args : any[] 17 super(...args); 18 >super(...args) : void 20 >...args : any [all …]
|
| /third_party/skia/tools/calmbench/ |
| D | calmbench.py | 61 'nanobench args (example: --svgs ~/Desktop/bots/svgs --skps ' 113 # Additional args for bots 118 args = parser.parse_args() 119 if not args.basearg: 120 args.basearg = args.extraarg 122 return args 125 def nano_path(args, branch): argument 126 return args.writedir + '/nanobench_' + branch 129 def compile_branch(args, branch): argument 130 print("Compiling branch %s" % args.branch) [all …]
|
| /third_party/skia/m133/tools/calmbench/ |
| D | calmbench.py | 61 'nanobench args (example: --svgs ~/Desktop/bots/svgs --skps ' 113 # Additional args for bots 118 args = parser.parse_args() 119 if not args.basearg: 120 args.basearg = args.extraarg 122 return args 125 def nano_path(args, branch): argument 126 return args.writedir + '/nanobench_' + branch 129 def compile_branch(args, branch): argument 130 print("Compiling branch %s" % args.branch) [all …]
|
| /third_party/python/Modules/clinic/ |
| D | _testclinic.c.h | 34 objects_converter(PyObject *module, PyObject *const *args, Py_ssize_t nargs) in objects_converter() argument 43 a = args[0]; in objects_converter() 47 b = args[1]; in objects_converter() 154 bool_converter(PyObject *module, PyObject *const *args, Py_ssize_t nargs) in bool_converter() argument 167 a = PyObject_IsTrue(args[0]); in bool_converter() 174 b = PyObject_IsTrue(args[1]); in bool_converter() 181 c = _PyLong_AsInt(args[2]); in bool_converter() 208 char_converter(PyObject *module, PyObject *const *args, Py_ssize_t nargs) in char_converter() argument 232 if (PyBytes_Check(args[0]) && PyBytes_GET_SIZE(args[0]) == 1) { in char_converter() 233 a = PyBytes_AS_STRING(args[0])[0]; in char_converter() [all …]
|
| /third_party/vk-gl-cts/external/vulkan-docs/src/scripts/ |
| D | genRelease | 65 args = parser.parse_args() variable 73 deps = ApiDependencies(args.registry, args.apiname) 78 if args.internal: 80 if args.repoDir is None: 81 args.repoDir = f'{args.gitroot}/vulkan' 82 if args.outDir is None: 83 args.outDir = f'{args.gitroot}/vulkan/out' 86 if args.repoDir is None: 87 args.repoDir = f'{args.gitroot}/Vulkan-Docs' 88 if args.outDir is None: [all …]
|
| /third_party/mindspore/mindspore-src/source/tests/st/networks/ |
| D | test_mindcv_overfit.py | 51 def train(args, device_id=0, rank_id=0, device_num=1): argument 54 ms.set_context(mode=args.mode, device_id=device_id) 59 args.lr = args.lr / 8 60 args.warmup_epochs = 0 72 set_seed(args.seed) 73 set_logger(name="mindcv", output_dir=args.ckpt_save_dir, rank=rank_id, color=False) 80 num_batches = 1281168 // args.batch_size 81 train_count = args.batch_size 85 model_name=args.model, 86 num_classes=args.num_classes, [all …]
|
| /third_party/rust/rust/compiler/rustc_errors/src/ |
| D | error.rs | 10 pub enum TranslateError<'args> { 12 id: &'args Cow<'args, str>, 13 args: &'args FluentArgs<'args>, 14 kind: TranslateErrorKind<'args>, 17 primary: Box<TranslateError<'args>>, 18 fallback: Box<TranslateError<'args>>, 22 impl<'args> TranslateError<'args> { 23 pub fn message(id: &'args Cow<'args, str>, args: &'args FluentArgs<'args>) -> Self { in message() argument 24 Self::One { id, args, kind: TranslateErrorKind::MessageMissing } in message() 26 pub fn primary(id: &'args Cow<'args, str>, args: &'args FluentArgs<'args>) -> Self { in primary() argument [all …]
|
| /third_party/grpc/src/ruby/ext/grpc/ |
| D | rb_channel_args.c | 49 grpc_channel_args* args; in grpc_rb_channel_create_in_process_add_args_hash_cb() local 67 &grpc_rb_channel_args_data_type, args); in grpc_rb_channel_create_in_process_add_args_hash_cb() 68 if (args->num_args <= 0) { in grpc_rb_channel_create_in_process_add_args_hash_cb() 70 args->num_args, StringValueCStr(key)); in grpc_rb_channel_create_in_process_add_args_hash_cb() 74 args->args[args->num_args - 1].key = gpr_strdup(the_key); in grpc_rb_channel_create_in_process_add_args_hash_cb() 77 args->args[args->num_args - 1].type = GRPC_ARG_STRING; in grpc_rb_channel_create_in_process_add_args_hash_cb() 78 args->args[args->num_args - 1].value.string = in grpc_rb_channel_create_in_process_add_args_hash_cb() 80 --args->num_args; in grpc_rb_channel_create_in_process_add_args_hash_cb() 84 args->args[args->num_args - 1].type = GRPC_ARG_STRING; in grpc_rb_channel_create_in_process_add_args_hash_cb() 85 args->args[args->num_args - 1].value.string = in grpc_rb_channel_create_in_process_add_args_hash_cb() [all …]
|
| /third_party/mindspore/mindspore-src/source/mindspore/python/mindspore/nn/probability/distribution/ |
| D | distribution.py | 33 Args: 42 must be passed in through `args` or `kwargs`. `dist_spec_args` which specifies 175 def _check_param_type(self, *args): argument 179 are None, the parameters must be passed in through `args`. 186 for arg, name, default in zip(args, self.parameter_names, self.default_parameters): 193 # broadcast if the number of args > 1 213 # broadcast all args to broadcast_shape 366 def _get_dist_args(self, *args, **kwargs): argument 367 return raise_not_implemented_util('get_dist_args', self.name, *args, **kwargs) 369 def get_dist_args(self, *args, **kwargs): argument [all …]
|
| /third_party/mindspore/mindspore-src/source/mindspore/ccsrc/plugin/device/cpu/kernel/ |
| D | adaptive_avg_pool_3d_cpu_kernel.cc | 98 CTask AdaptiveAvgPool3DOutFrame(const AdaptiveCalcArgs<SCALAR_T> &args) { in AdaptiveAvgPool3DOutFrame() argument 99 auto shard_frame = [&args](int64_t start, int64_t end) { in AdaptiveAvgPool3DOutFrame() 100 MS_EXCEPTION_IF_NULL(args.input_data); in AdaptiveAvgPool3DOutFrame() 101 MS_EXCEPTION_IF_NULL(args.output_data); in AdaptiveAvgPool3DOutFrame() 104 for (int64_t out_t = 0; out_t < args.out_size_t; out_t++) { in AdaptiveAvgPool3DOutFrame() 105 int64_t in_start_t = StartIndex(out_t, args.out_size_t, args.in_size_t); in AdaptiveAvgPool3DOutFrame() 106 int64_t in_end_t = EndIndex(out_t, args.out_size_t, args.in_size_t); in AdaptiveAvgPool3DOutFrame() 108 for (int64_t out_h = 0; out_h < args.out_size_h; out_h++) { in AdaptiveAvgPool3DOutFrame() 109 int64_t in_start_h = StartIndex(out_h, args.out_size_h, args.in_size_h); in AdaptiveAvgPool3DOutFrame() 110 int64_t in_end_h = EndIndex(out_h, args.out_size_h, args.in_size_h); in AdaptiveAvgPool3DOutFrame() [all …]
|
| /third_party/mindspore/mindspore-src/source/docs/api/api_python/nn_probability/ |
| D | mindspore.nn.probability.distribution.Distribution.rst | 15 … 派生类必须重写 `_mean` 、 `_prob` 和 `_log_prob` 等操作。必填参数必须通过 `args` 或 `kwargs` 传入,如 `_prob` 的 `value` 。 24 .. py:method:: cdf(value, *args, **kwargs) 30 - **args** (list) - 位置参数列表,具体需要的参数根据子类的实现确定。 34 可以通过 `args` 或 `kwargs` 传递其 `dist_spec_args` 来选择性地将Distribution传递给函数。 39 .. py:method:: construct(name, *args, **kwargs) 49 - **args** (list) - 函数所需的位置参数列表。 55 .. py:method:: cross_entropy(dist, *args, **kwargs) 61 - **args** (list) - 位置参数列表,具体需要的参数根据子类的实现确定。 65 …Distribution b的 `dist_spec_args` 必须通过 `args` 或 `kwargs` 传递给函数。传入Distribution a的 `dist_spec_args` 是… 70 .. py:method:: entropy(*args, **kwargs) [all …]
|
| /third_party/skia/src/gpu/ops/ |
| D | TessellationPathRenderer.cpp | 88 const CanDrawPathArgs& args) const { in onCanDrawPath() 89 const GrStyledShape& shape = *args.fShape; in onCanDrawPath() 90 if (args.fAAType == GrAAType::kCoverage || in onCanDrawPath() 92 args.fViewMatrix->hasPerspective() || in onCanDrawPath() 94 !args.fProxy->canUseStencil(*args.fCaps)) { in onCanDrawPath() 102 if (args.fHasUserStencilSettings) { in onCanDrawPath() 112 bool TessellationPathRenderer::onDrawPath(const DrawPathArgs& args) { in onDrawPath() argument 113 auto sdc = args.fSurfaceDrawContext; in onDrawPath() 116 args.fShape->asPath(&path); in onDrawPath() 118 const SkRect pathDevBounds = args.fViewMatrix->mapRect(args.fShape->bounds()); in onDrawPath() [all …]
|