/external/virglrenderer/src/venus/ |
D | vkr_command_buffer.c | 14 #define VKR_CMD_CALL(cmd_name, args, ...) \ argument 17 vkr_command_buffer_from_handle(args->commandBuffer); \ 20 vn_replace_vk##cmd_name##_args_handle(args); \ 21 _vk->cmd_name(args->commandBuffer, ##__VA_ARGS__); \ 26 struct vn_command_vkCreateCommandPool *args) in vkr_dispatch_vkCreateCommandPool() argument 28 struct vkr_command_pool *pool = vkr_command_pool_create_and_add(dispatch->data, args); in vkr_dispatch_vkCreateCommandPool() 37 struct vn_command_vkDestroyCommandPool *args) in vkr_dispatch_vkDestroyCommandPool() argument 40 struct vkr_command_pool *pool = vkr_command_pool_from_handle(args->commandPool); in vkr_dispatch_vkDestroyCommandPool() 46 vkr_command_pool_destroy_and_remove(ctx, args); in vkr_dispatch_vkDestroyCommandPool() 51 struct vn_command_vkResetCommandPool *args) in vkr_dispatch_vkResetCommandPool() argument [all …]
|
D | vkr_image.c | 13 struct vn_command_vkCreateImage *args) in vkr_dispatch_vkCreateImage() argument 34 vkr_image_create_and_add(dispatch->data, args); in vkr_dispatch_vkCreateImage() 39 struct vn_command_vkDestroyImage *args) in vkr_dispatch_vkDestroyImage() argument 41 vkr_image_destroy_and_remove(dispatch->data, args); in vkr_dispatch_vkDestroyImage() 47 struct vn_command_vkGetImageMemoryRequirements *args) in vkr_dispatch_vkGetImageMemoryRequirements() argument 49 struct vkr_device *dev = vkr_device_from_handle(args->device); in vkr_dispatch_vkGetImageMemoryRequirements() 52 vn_replace_vkGetImageMemoryRequirements_args_handle(args); in vkr_dispatch_vkGetImageMemoryRequirements() 53 vk->GetImageMemoryRequirements(args->device, args->image, args->pMemoryRequirements); in vkr_dispatch_vkGetImageMemoryRequirements() 59 struct vn_command_vkGetImageMemoryRequirements2 *args) in vkr_dispatch_vkGetImageMemoryRequirements2() argument 61 struct vkr_device *dev = vkr_device_from_handle(args->device); in vkr_dispatch_vkGetImageMemoryRequirements2() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/llvm_ir/ |
D | ir_builder_mixin.h | 37 template <class... Args> 38 llvm::Value* Add(Args&&... args) { in Add() argument 39 return mixin_builder()->CreateAdd(std::forward<Args>(args)...); in Add() 42 template <class... Args> 43 llvm::LoadInst* AlignedLoad(Args&&... args) { in AlignedLoad() argument 44 return mixin_builder()->CreateAlignedLoad(std::forward<Args>(args)...); in AlignedLoad() 47 template <class... Args> 48 llvm::StoreInst* AlignedStore(Args&&... args) { in AlignedStore() argument 49 return mixin_builder()->CreateAlignedStore(std::forward<Args>(args)...); in AlignedStore() 52 template <class... Args> [all …]
|
/external/mesa3d/src/amd/vulkan/ |
D | radv_shader_args.c | 55 add_ud_arg(struct radv_shader_args *args, unsigned size, enum ac_arg_type type, struct ac_arg *arg, in add_ud_arg() argument 58 ac_add_arg(&args->ac, AC_ARG_SGPR, size, type, arg); in add_ud_arg() 60 struct radv_userdata_info *ud_info = &args->user_sgprs_locs.shader_data[ud]; in add_ud_arg() 63 ud_info->sgpr_idx = args->num_user_sgprs; in add_ud_arg() 67 args->num_user_sgprs += size; in add_ud_arg() 71 add_descriptor_set(struct radv_shader_args *args, enum ac_arg_type type, struct ac_arg *arg, uint32… in add_descriptor_set() argument 73 ac_add_arg(&args->ac, AC_ARG_SGPR, 1, type, arg); in add_descriptor_set() 75 struct radv_userdata_info *ud_info = &args->user_sgprs_locs.descriptor_sets[set]; in add_descriptor_set() 76 ud_info->sgpr_idx = args->num_user_sgprs; in add_descriptor_set() 79 args->user_sgprs_locs.descriptor_sets_enabled |= 1u << set; in add_descriptor_set() [all …]
|
/external/mesa3d/src/intel/vulkan/grl/gpu/ |
D | rebraid.grl | 43 MKRebraidArgs Args 46 dispatch init_scratch(1,1,1) args( Args.rebraid_scratch ); 47 dispatch calc_aabb(Args.num_instances,1,1) args( Args.bvh_buffer, Args.instances_buffer ); 50 …//define num_count_groups ((Args.num_instances + (COUNT_SPLITS_GROUP_SIZE-1)) / COUNT_SPLITS_GROUP… 51 …atch count_splits(num_count_groups,1,1) args( Args.bvh_buffer, Args.instances_buffer, Args.rebraid… 53 …dispatch count_splits_SG(Args.num_instances,1,1) args( Args.bvh_buffer, Args.instances_buffer, Arg… 56 define num_primref_groups ((Args.num_instances + (PRIMREF_GROUP_SIZE-1)) / PRIMREF_GROUP_SIZE); 58 …s,1,1) args( Args.global_buffer, Args.bvh_buffer, Args.instances_buffer, Args.rebraid_scratch, Arg… 61 …UG_PRINT(1,1,1) args( Args.global_buffer, Args.instances_buffer, Args.rebraid_scratch, Args.primre… 64 metakernel rebraid_indirect(MKRebraidArgs Args, qword indirectBuildRangeInfo) [all …]
|
/external/virglrenderer/src/venus/venus-protocol/ |
D | vn_protocol_renderer_command_buffer.h | 2574 …mmandBuffers_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkAllocateCommandBuffers *args) in vn_decode_vkAllocateCommandBuffers_args_temp() argument 2576 vn_decode_VkDevice_lookup(dec, &args->device); in vn_decode_vkAllocateCommandBuffers_args_temp() 2578 args->pAllocateInfo = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pAllocateInfo)); in vn_decode_vkAllocateCommandBuffers_args_temp() 2579 if (!args->pAllocateInfo) return; in vn_decode_vkAllocateCommandBuffers_args_temp() 2580 …vn_decode_VkCommandBufferAllocateInfo_temp(dec, (VkCommandBufferAllocateInfo *)args->pAllocateInfo… in vn_decode_vkAllocateCommandBuffers_args_temp() 2582 args->pAllocateInfo = NULL; in vn_decode_vkAllocateCommandBuffers_args_temp() 2586 …const uint32_t iter_count = vn_decode_array_size(dec, (args->pAllocateInfo ? args->pAllocateInfo->… in vn_decode_vkAllocateCommandBuffers_args_temp() 2587 …args->pCommandBuffers = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pCommandBuffers) * iter_count); in vn_decode_vkAllocateCommandBuffers_args_temp() 2588 if (!args->pCommandBuffers) return; in vn_decode_vkAllocateCommandBuffers_args_temp() 2590 vn_decode_VkCommandBuffer_temp(dec, &args->pCommandBuffers[i]); in vn_decode_vkAllocateCommandBuffers_args_temp() [all …]
|
D | vn_protocol_renderer_transport.h | 130 …treamMESA_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkSetReplyCommandStreamMESA *args) in vn_decode_vkSetReplyCommandStreamMESA_args_temp() argument 133 args->pStream = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pStream)); in vn_decode_vkSetReplyCommandStreamMESA_args_temp() 134 if (!args->pStream) return; in vn_decode_vkSetReplyCommandStreamMESA_args_temp() 135 …vn_decode_VkCommandStreamDescriptionMESA_temp(dec, (VkCommandStreamDescriptionMESA *)args->pStream… in vn_decode_vkSetReplyCommandStreamMESA_args_temp() 137 args->pStream = NULL; in vn_decode_vkSetReplyCommandStreamMESA_args_temp() 142 …eplace_vkSetReplyCommandStreamMESA_args_handle(struct vn_command_vkSetReplyCommandStreamMESA *args) in vn_replace_vkSetReplyCommandStreamMESA_args_handle() argument 144 if (args->pStream) in vn_replace_vkSetReplyCommandStreamMESA_args_handle() 145 … vn_replace_VkCommandStreamDescriptionMESA_handle((VkCommandStreamDescriptionMESA *)args->pStream); in vn_replace_vkSetReplyCommandStreamMESA_args_handle() 148 …eamMESA_reply(struct vn_cs_encoder *enc, const struct vn_command_vkSetReplyCommandStreamMESA *args) in vn_encode_vkSetReplyCommandStreamMESA_reply() argument 152 /* skip args->pStream */ in vn_encode_vkSetReplyCommandStreamMESA_reply() [all …]
|
D | vn_protocol_renderer_pipeline_cache.h | 86 …tePipelineCache_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkCreatePipelineCache *args) in vn_decode_vkCreatePipelineCache_args_temp() argument 88 vn_decode_VkDevice_lookup(dec, &args->device); in vn_decode_vkCreatePipelineCache_args_temp() 90 args->pCreateInfo = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pCreateInfo)); in vn_decode_vkCreatePipelineCache_args_temp() 91 if (!args->pCreateInfo) return; in vn_decode_vkCreatePipelineCache_args_temp() 92 … vn_decode_VkPipelineCacheCreateInfo_temp(dec, (VkPipelineCacheCreateInfo *)args->pCreateInfo); in vn_decode_vkCreatePipelineCache_args_temp() 94 args->pCreateInfo = NULL; in vn_decode_vkCreatePipelineCache_args_temp() 100 args->pAllocator = NULL; in vn_decode_vkCreatePipelineCache_args_temp() 103 args->pPipelineCache = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pPipelineCache)); in vn_decode_vkCreatePipelineCache_args_temp() 104 if (!args->pPipelineCache) return; in vn_decode_vkCreatePipelineCache_args_temp() 105 vn_decode_VkPipelineCache(dec, args->pPipelineCache); in vn_decode_vkCreatePipelineCache_args_temp() [all …]
|
D | vn_protocol_renderer_query_pool.h | 80 …e_vkCreateQueryPool_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkCreateQueryPool *args) in vn_decode_vkCreateQueryPool_args_temp() argument 82 vn_decode_VkDevice_lookup(dec, &args->device); in vn_decode_vkCreateQueryPool_args_temp() 84 args->pCreateInfo = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pCreateInfo)); in vn_decode_vkCreateQueryPool_args_temp() 85 if (!args->pCreateInfo) return; in vn_decode_vkCreateQueryPool_args_temp() 86 vn_decode_VkQueryPoolCreateInfo_temp(dec, (VkQueryPoolCreateInfo *)args->pCreateInfo); in vn_decode_vkCreateQueryPool_args_temp() 88 args->pCreateInfo = NULL; in vn_decode_vkCreateQueryPool_args_temp() 94 args->pAllocator = NULL; in vn_decode_vkCreateQueryPool_args_temp() 97 args->pQueryPool = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pQueryPool)); in vn_decode_vkCreateQueryPool_args_temp() 98 if (!args->pQueryPool) return; in vn_decode_vkCreateQueryPool_args_temp() 99 vn_decode_VkQueryPool(dec, args->pQueryPool); in vn_decode_vkCreateQueryPool_args_temp() [all …]
|
D | vn_protocol_renderer_fence.h | 161 …vn_decode_vkCreateFence_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkCreateFence *args) in vn_decode_vkCreateFence_args_temp() argument 163 vn_decode_VkDevice_lookup(dec, &args->device); in vn_decode_vkCreateFence_args_temp() 165 args->pCreateInfo = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pCreateInfo)); in vn_decode_vkCreateFence_args_temp() 166 if (!args->pCreateInfo) return; in vn_decode_vkCreateFence_args_temp() 167 vn_decode_VkFenceCreateInfo_temp(dec, (VkFenceCreateInfo *)args->pCreateInfo); in vn_decode_vkCreateFence_args_temp() 169 args->pCreateInfo = NULL; in vn_decode_vkCreateFence_args_temp() 175 args->pAllocator = NULL; in vn_decode_vkCreateFence_args_temp() 178 args->pFence = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pFence)); in vn_decode_vkCreateFence_args_temp() 179 if (!args->pFence) return; in vn_decode_vkCreateFence_args_temp() 180 vn_decode_VkFence(dec, args->pFence); in vn_decode_vkCreateFence_args_temp() [all …]
|
D | vn_protocol_renderer_private_data_slot.h | 74 …ivateDataSlot_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkCreatePrivateDataSlot *args) in vn_decode_vkCreatePrivateDataSlot_args_temp() argument 76 vn_decode_VkDevice_lookup(dec, &args->device); in vn_decode_vkCreatePrivateDataSlot_args_temp() 78 args->pCreateInfo = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pCreateInfo)); in vn_decode_vkCreatePrivateDataSlot_args_temp() 79 if (!args->pCreateInfo) return; in vn_decode_vkCreatePrivateDataSlot_args_temp() 80 … vn_decode_VkPrivateDataSlotCreateInfo_temp(dec, (VkPrivateDataSlotCreateInfo *)args->pCreateInfo); in vn_decode_vkCreatePrivateDataSlot_args_temp() 82 args->pCreateInfo = NULL; in vn_decode_vkCreatePrivateDataSlot_args_temp() 88 args->pAllocator = NULL; in vn_decode_vkCreatePrivateDataSlot_args_temp() 91 args->pPrivateDataSlot = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pPrivateDataSlot)); in vn_decode_vkCreatePrivateDataSlot_args_temp() 92 if (!args->pPrivateDataSlot) return; in vn_decode_vkCreatePrivateDataSlot_args_temp() 93 vn_decode_VkPrivateDataSlot(dec, args->pPrivateDataSlot); in vn_decode_vkCreatePrivateDataSlot_args_temp() [all …]
|
D | vn_protocol_renderer_event.h | 74 …vn_decode_vkCreateEvent_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkCreateEvent *args) in vn_decode_vkCreateEvent_args_temp() argument 76 vn_decode_VkDevice_lookup(dec, &args->device); in vn_decode_vkCreateEvent_args_temp() 78 args->pCreateInfo = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pCreateInfo)); in vn_decode_vkCreateEvent_args_temp() 79 if (!args->pCreateInfo) return; in vn_decode_vkCreateEvent_args_temp() 80 vn_decode_VkEventCreateInfo_temp(dec, (VkEventCreateInfo *)args->pCreateInfo); in vn_decode_vkCreateEvent_args_temp() 82 args->pCreateInfo = NULL; in vn_decode_vkCreateEvent_args_temp() 88 args->pAllocator = NULL; in vn_decode_vkCreateEvent_args_temp() 91 args->pEvent = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pEvent)); in vn_decode_vkCreateEvent_args_temp() 92 if (!args->pEvent) return; in vn_decode_vkCreateEvent_args_temp() 93 vn_decode_VkEvent(dec, args->pEvent); in vn_decode_vkCreateEvent_args_temp() [all …]
|
/external/ComputeLibrary/src/core/NEON/kernels/arm_gemm/ |
D | gemm_fp32.cpp | 87 … [](const GemmArgs &args) { return args._Msize==1 && args._nbatches>1 && !args._indirect_input; }, in __anon79dafa9e0102() 89 [](const GemmArgs &args) { return new GemvBatched<float, float>(args); } in __anon79dafa9e0202() 97 [](const GemmArgs &args) { return args._fast_mode && args._ci->has_bf16(); }, in __anon79dafa9e0302() 98 …[](const GemmArgs &args) { return GemmInterleaved<cls_a64_interleaved_bf16fp32_mmla_8x12, float, f… in __anon79dafa9e0402() 99 …[](const GemmArgs &args) { return new GemmInterleaved<cls_a64_interleaved_bf16fp32_mmla_8x12, floa… in __anon79dafa9e0502() 105 [](const GemmArgs &args) { return args._fast_mode && args._ci->has_bf16(); }, in __anon79dafa9e0602() 106 …[](const GemmArgs &args) { return GemmHybridIndirect<cls_a64_hybrid_fp32bf16fp32_mmla_6x16, float,… in __anon79dafa9e0702() 107 …[](const GemmArgs &args) { return new GemmHybridIndirect<cls_a64_hybrid_fp32bf16fp32_mmla_6x16, fl… in __anon79dafa9e0802() 112 [](const GemmArgs &args) { return args._fast_mode && args._ci->has_bf16(); }, in __anon79dafa9e0902() 113 …[](const GemmArgs &args) { return GemmHybridIndirect<cls_a64_hybrid_fp32bf16fp32_mmla_4x24, float,… in __anon79dafa9e0a02() [all …]
|
D | gemm_qint8.cpp | 77 …rgs &args, const Requantize32 &qp) { return args._ci->has_sme2() && quant_hybrid_asymmetric(qp) &&… in __anon98fe7f570102() 79 …nst GemmArgs &args, const Requantize32 &qp) { return new GemvPretransposed<cls_sme2_gemv_s8qa_dot_… in __anon98fe7f570202() 84 …[](const GemmArgs &args, const Requantize32 &qp) { return args._ci->has_sme2() && args._maxthreads… in __anon98fe7f570302() 85 …[](const GemmArgs &args, const Requantize32 &) { const auto VL = sme::get_vector_length<int32_t>(); in __anon98fe7f570402() 86 … return args._Msize <= VL || (2*VL < args._Msize && args._Msize <= 3*VL); }, in __anon98fe7f570402() 87 …args, const Requantize32 &qp) { return new GemmInterleavedPretransposedNoMergeQuantizedInline<cls_… in __anon98fe7f570502() 92 …[](const GemmArgs &args, const Requantize32 &qp) { return args._ci->has_sme2() && args._maxthreads… in __anon98fe7f570602() 93 …[](const GemmArgs &args, const Requantize32 &) { const auto VL = sme::get_vector_length<int32_t>(); in __anon98fe7f570702() 94 … return args._Nsize <= VL || (2*VL < args._Nsize && args._Nsize <= 3*VL); }, in __anon98fe7f570702() 95 …args, const Requantize32 &qp) { return new GemmInterleavedPretransposedNoMergeQuantizedInline<cls_… in __anon98fe7f570802() [all …]
|
D | gemm_int8.cpp | 65 [](const GemmArgs &args) { return args._ci->has_sme2(); }, in __anon3cb7e3a60102() 66 [](const GemmArgs &args) { const auto VL = sme::get_vector_length<int32_t>(); in __anon3cb7e3a60202() 67 … return args._Msize <= VL || (2*VL < args._Msize && args._Msize <= 3*VL); }, in __anon3cb7e3a60202() 68 …[](const GemmArgs &args) { return new GemmInterleavedNoMerge<cls_sme2_interleaved_nomerge_s8s32_mo… in __anon3cb7e3a60302() 73 [](const GemmArgs &args) { return args._ci->has_sme2(); }, in __anon3cb7e3a60402() 74 [](const GemmArgs &args) { const auto VL = sme::get_vector_length<int32_t>(); in __anon3cb7e3a60502() 75 … return args._Nsize <= VL || (2*VL < args._Nsize && args._Nsize <= 3*VL); }, in __anon3cb7e3a60502() 76 …[](const GemmArgs &args) { return new GemmInterleavedNoMerge<cls_sme2_interleaved_nomerge_s8s32_mo… in __anon3cb7e3a60602() 81 [](const GemmArgs &args) { return args._ci->has_sme2(); }, in __anon3cb7e3a60702() 83 …[](const GemmArgs &args) { return new GemmInterleavedNoMerge<cls_sme2_interleaved_nomerge_s8s32_mo… in __anon3cb7e3a60802() [all …]
|
D | gemm_quint8.cpp | 73 …rgs &args, const Requantize32 &qp) { return args._ci->has_sme2() && quant_hybrid_asymmetric(qp) &&… in __anon500a552c0102() 75 …st GemmArgs &args, const Requantize32 &qp) { return new GemvPretransposed<cls_sme2_gemv_u8qa_dot_1… in __anon500a552c0202() 80 …[](const GemmArgs &args, const Requantize32 &qp) { return args._ci->has_sme2() && args._maxthreads… in __anon500a552c0302() 81 …[](const GemmArgs &args, const Requantize32 &) { const auto VL = sme::get_vector_length<uint32_t>(… in __anon500a552c0402() 82 … return args._Msize <= VL || (2*VL < args._Msize && args._Msize <= 3*VL); }, in __anon500a552c0402() 83 …args, const Requantize32 &qp) { return new GemmInterleavedPretransposedNoMergeQuantizedInline<cls_… in __anon500a552c0502() 88 …[](const GemmArgs &args, const Requantize32 &qp) { return args._ci->has_sme2() && args._maxthreads… in __anon500a552c0602() 89 …[](const GemmArgs &args, const Requantize32 &) { const auto VL = sme::get_vector_length<int32_t>(); in __anon500a552c0702() 90 … return args._Nsize <= VL || (2*VL < args._Nsize && args._Nsize <= 3*VL); }, in __anon500a552c0702() 91 …args, const Requantize32 &qp) { return new GemmInterleavedPretransposedNoMergeQuantizedInline<cls_… in __anon500a552c0802() [all …]
|
D | gemm_bf16.cpp | 78 …[](const GemmArgs &args) { return args._ci->has_sme2() && args._Msize==1 && args._nbatches==1 && !… in __anonc2ab58520102() 80 …[](const GemmArgs &args) { return new GemvPretransposed<cls_sme2_gemv_bf16fp32_dot_16VL, bfloat16,… in __anonc2ab58520202() 85 [](const GemmArgs &args) { return args._ci->has_sme2(); }, in __anonc2ab58520302() 86 [](const GemmArgs &args) { const auto VL = sme::get_vector_length<float>(); in __anonc2ab58520402() 87 … return args._Msize <= VL || (2*VL < args._Msize && args._Msize <= 3*VL); }, in __anonc2ab58520402() 88 …[](const GemmArgs &args) { return new GemmInterleavedNoMerge<cls_sme2_interleaved_nomerge_bf16fp32… in __anonc2ab58520502() 93 [](const GemmArgs &args) { return args._ci->has_sme2(); }, in __anonc2ab58520602() 94 [](const GemmArgs &args) { const auto VL = sme::get_vector_length<float>(); in __anonc2ab58520702() 95 … return args._Nsize <= VL || (2*VL < args._Nsize && args._Nsize <= 3*VL); }, in __anonc2ab58520702() 96 …[](const GemmArgs &args) { return new GemmInterleavedNoMerge<cls_sme2_interleaved_nomerge_bf16fp32… in __anonc2ab58520802() [all …]
|
D | gemm_uint8.cpp | 55 [](const GemmArgs &args) { return args._ci->has_svei8mm(); }, in __anon6af0435b0102() 56 …](const GemmArgs &args) { return GemmHybridIndirect<cls_sve_hybrid_u8u32_mmla_6x4VL, uint8_t, uint… in __anon6af0435b0202() 57 …[](const GemmArgs &args) { return new GemmHybridIndirect<cls_sve_hybrid_u8u32_mmla_6x4VL, uint8_t,… in __anon6af0435b0302() 62 [](const GemmArgs &args) { return args._ci->has_svei8mm() && (args._Ksize>8); }, in __anon6af0435b0402() 63 …(const GemmArgs &args) { return GemmInterleaved<cls_sve_interleaved_u8u32_mmla_8x3VL, uint8_t, uin… in __anon6af0435b0502() 64 …[](const GemmArgs &args) { return new GemmInterleaved<cls_sve_interleaved_u8u32_mmla_8x3VL, uint8_… in __anon6af0435b0602() 69 …[](const GemmArgs &args) { return args._ci->has_sve() && args._Ksize<=64 && !args._indirect_input;… in __anon6af0435b0702() 70 [](const GemmArgs &args) { return !(args._ci->has_svei8mm() || args._ci->has_i8mm()); }, in __anon6af0435b0802() 71 …[](const GemmArgs &args) { return new GemmHybrid<cls_sve_smallK_hybrid_u8u32_dot_8x1VL, uint8_t, u… in __anon6af0435b0902() 76 [](const GemmArgs &args) { return args._ci->has_sve(); }, in __anon6af0435b0a02() [all …]
|
/external/pytorch/benchmarks/distributed/rpc/parameter_server/ |
D | launcher.py | 29 def get_name(rank, args): argument 33 Args: 35 args (parser): benchmark configurations 37 t_count = args.ntrainer + args.ncudatrainer 38 s_count = args.nserver + args.ncudaserver 47 def get_server_rank(args, rank): argument 51 Args: 52 args (parser): benchmark configurations 55 s_offset = args.ntrainer + args.ncudatrainer 56 tps = args.ntrainer // args.nserver [all …]
|
/external/zstd/contrib/pzstd/test/ |
D | OptionsTest.cpp | 83 template <typename... Args> 84 std::array<const char *, sizeof...(Args) + 1> makeArray(Args... args) { in makeArray() argument 85 return {{nullptr, args...}}; in makeArray() 91 auto args = makeArray("--processes", "5", "-o", "x", "y", "-f"); in TEST() local 92 EXPECT_SUCCESS(options.parse(args.size(), args.data())); in TEST() 99 auto args = makeArray("-p", "1", "input", "-19"); in TEST() local 100 EXPECT_SUCCESS(options.parse(args.size(), args.data())); in TEST() 107 auto args = in TEST() local 109 EXPECT_SUCCESS(options.parse(args.size(), args.data())); in TEST() 116 auto args = makeArray("--processes", "100", "hello.zst", "--decompress", in TEST() local [all …]
|
/external/mesa3d/.gitlab-ci/bare-metal/ |
D | mkbootimg.py | 77 def get_recovery_dtbo_offset(args): argument 80 num_kernel_pages = get_number_of_pages(filesize(args.kernel), args.pagesize) 81 num_ramdisk_pages = get_number_of_pages(filesize(args.ramdisk), 82 args.pagesize) 83 num_second_pages = get_number_of_pages(filesize(args.second), args.pagesize) 84 dtbo_offset = args.pagesize * (num_header_pages + num_kernel_pages + 87 def write_header_v3_and_above(args): argument 88 if args.header_version > 3: 92 args.output.write(pack(f'{BOOT_MAGIC_SIZE}s', BOOT_MAGIC.encode())) 94 args.output.write(pack('I', filesize(args.kernel))) [all …]
|
/external/sdv/vsomeip/third_party/boost/type_traits/include/boost/type_traits/detail/ |
D | is_function_cxx_11.hpp | 36 template <class Ret, class...Args BOOST_TT_NOEXCEPT_PARAM> 37 struct is_function<Ret BOOST_TT_DEF_CALL(Args...)BOOST_TT_NOEXCEPT_DECL> : public true_type {}; 39 template <class Ret, class ...Args BOOST_TT_NOEXCEPT_PARAM> 40 struct is_function<Ret(Args..., ...)BOOST_TT_NOEXCEPT_DECL> : public true_type {}; 44 template <class Ret, class...Args BOOST_TT_NOEXCEPT_PARAM> 45 …struct is_function<Ret BOOST_TT_DEF_CALL(Args...)const BOOST_TT_NOEXCEPT_DECL> : public true_type … 47 template <class Ret, class ...Args BOOST_TT_NOEXCEPT_PARAM> 48 struct is_function<Ret(Args..., ...)const BOOST_TT_NOEXCEPT_DECL> : public true_type {}; 52 template <class Ret, class...Args BOOST_TT_NOEXCEPT_PARAM> 53 …struct is_function<Ret BOOST_TT_DEF_CALL(Args...)volatile BOOST_TT_NOEXCEPT_DECL> : public true_ty… [all …]
|
D | is_member_function_pointer_cxx_11.hpp | 47 template <class Ret, class C, class...Args BOOST_TT_NOEXCEPT_PARAM> 48 …struct is_member_function_pointer<Ret (BOOST_TT_DEF_CALL C::*)(Args...)BOOST_TT_NOEXCEPT_DECL> : p… 49 template <class Ret, class C, class ...Args BOOST_TT_NOEXCEPT_PARAM> 50 …struct is_member_function_pointer<Ret (C::*)(Args..., ...)BOOST_TT_NOEXCEPT_DECL> : public true_ty… 52 template <class Ret, class C, class...Args BOOST_TT_NOEXCEPT_PARAM> 53 …struct is_member_function_pointer<Ret (BOOST_TT_DEF_CALL C::*)(Args...)const BOOST_TT_NOEXCEPT_DEC… 54 template <class Ret, class C, class ...Args BOOST_TT_NOEXCEPT_PARAM> 55 …struct is_member_function_pointer<Ret(C::*)(Args..., ...)const BOOST_TT_NOEXCEPT_DECL> : public tr… 57 template <class Ret, class C, class...Args BOOST_TT_NOEXCEPT_PARAM> 58 …struct is_member_function_pointer<Ret (BOOST_TT_DEF_CALL C::*)(Args...)volatile BOOST_TT_NOEXCEPT_… [all …]
|
/external/XNNPACK/bench/ |
D | spmm.h | 31 b->Args({56 * 56, 36, 24}); in ShuffleNetV1G1SpmmArguments() 32 b->Args({28 * 28, 120, 36}); in ShuffleNetV1G1SpmmArguments() 33 b->Args({28 * 28, 36, 144}); in ShuffleNetV1G1SpmmArguments() 34 b->Args({28 * 28, 144, 36}); in ShuffleNetV1G1SpmmArguments() 35 b->Args({28 * 28, 72, 144}); in ShuffleNetV1G1SpmmArguments() 36 b->Args({14 * 14, 144, 72}); in ShuffleNetV1G1SpmmArguments() 37 b->Args({14 * 14, 72, 288}); in ShuffleNetV1G1SpmmArguments() 38 b->Args({14 * 14, 288, 72}); in ShuffleNetV1G1SpmmArguments() 39 b->Args({14 * 14, 144, 288}); in ShuffleNetV1G1SpmmArguments() 40 b->Args({ 7 * 7, 288, 144}); in ShuffleNetV1G1SpmmArguments() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/pjrt/c/ |
D | pjrt_c_api_wrapper_impl.cc | 60 void PJRT_Error_Destroy(PJRT_Error_Destroy_Args* args) { in PJRT_Error_Destroy() argument 63 args->struct_size); in PJRT_Error_Destroy() 67 if (args->struct_size >= PJRT_STRUCT_SIZE(PJRT_Error_Destroy_Args, error)) { in PJRT_Error_Destroy() 68 delete args->error; in PJRT_Error_Destroy() 72 void PJRT_Error_Message(PJRT_Error_Message_Args* args) { in PJRT_Error_Message() argument 75 args->struct_size); in PJRT_Error_Message() 79 if (args->struct_size >= PJRT_STRUCT_SIZE(PJRT_Error_Destroy_Args, error)) { in PJRT_Error_Message() 80 const xla::Status* status = &args->error->status; in PJRT_Error_Message() 81 args->message = status->error_message().data(); in PJRT_Error_Message() 82 args->message_size = status->error_message().size(); in PJRT_Error_Message() [all …]
|