| /external/virglrenderer/src/venus/ |
| D | vkr_command_buffer.c | 14 #define VKR_CMD_CALL(cmd_name, args, ...) \ argument 17 vkr_command_buffer_from_handle(args->commandBuffer); \ 20 vn_replace_vk##cmd_name##_args_handle(args); \ 21 _vk->cmd_name(args->commandBuffer, ##__VA_ARGS__); \ 26 struct vn_command_vkCreateCommandPool *args) in vkr_dispatch_vkCreateCommandPool() argument 28 struct vkr_command_pool *pool = vkr_command_pool_create_and_add(dispatch->data, args); in vkr_dispatch_vkCreateCommandPool() 37 struct vn_command_vkDestroyCommandPool *args) in vkr_dispatch_vkDestroyCommandPool() argument 40 struct vkr_command_pool *pool = vkr_command_pool_from_handle(args->commandPool); in vkr_dispatch_vkDestroyCommandPool() 46 vkr_command_pool_destroy_and_remove(ctx, args); in vkr_dispatch_vkDestroyCommandPool() 51 struct vn_command_vkResetCommandPool *args) in vkr_dispatch_vkResetCommandPool() argument [all …]
|
| D | vkr_image.c | 13 struct vn_command_vkCreateImage *args) in vkr_dispatch_vkCreateImage() argument 34 vkr_image_create_and_add(dispatch->data, args); in vkr_dispatch_vkCreateImage() 39 struct vn_command_vkDestroyImage *args) in vkr_dispatch_vkDestroyImage() argument 41 vkr_image_destroy_and_remove(dispatch->data, args); in vkr_dispatch_vkDestroyImage() 47 struct vn_command_vkGetImageMemoryRequirements *args) in vkr_dispatch_vkGetImageMemoryRequirements() argument 49 struct vkr_device *dev = vkr_device_from_handle(args->device); in vkr_dispatch_vkGetImageMemoryRequirements() 52 vn_replace_vkGetImageMemoryRequirements_args_handle(args); in vkr_dispatch_vkGetImageMemoryRequirements() 53 vk->GetImageMemoryRequirements(args->device, args->image, args->pMemoryRequirements); in vkr_dispatch_vkGetImageMemoryRequirements() 59 struct vn_command_vkGetImageMemoryRequirements2 *args) in vkr_dispatch_vkGetImageMemoryRequirements2() argument 61 struct vkr_device *dev = vkr_device_from_handle(args->device); in vkr_dispatch_vkGetImageMemoryRequirements2() [all …]
|
| /external/tensorflow/tensorflow/compiler/xla/service/llvm_ir/ |
| D | ir_builder_mixin.h | 37 template <class... Args> 38 llvm::Value* Add(Args&&... args) { in Add() argument 39 return mixin_builder()->CreateAdd(std::forward<Args>(args)...); in Add() 42 template <class... Args> 43 llvm::LoadInst* AlignedLoad(Args&&... args) { in AlignedLoad() argument 44 return mixin_builder()->CreateAlignedLoad(std::forward<Args>(args)...); in AlignedLoad() 47 template <class... Args> 48 llvm::StoreInst* AlignedStore(Args&&... args) { in AlignedStore() argument 49 return mixin_builder()->CreateAlignedStore(std::forward<Args>(args)...); in AlignedStore() 52 template <class... Args> [all …]
|
| /external/mesa3d/src/amd/vulkan/ |
| D | radv_shader_args.c | 43 set_loc_shader(struct radv_shader_args *args, int idx, uint8_t *sgpr_idx, in set_loc_shader() argument 47 &args->shader_info->user_sgprs_locs.shader_data[idx]; in set_loc_shader() 54 set_loc_shader_ptr(struct radv_shader_args *args, int idx, uint8_t *sgpr_idx) in set_loc_shader_ptr() argument 58 set_loc_shader(args, idx, sgpr_idx, use_32bit_pointers ? 1 : 2); in set_loc_shader_ptr() 62 set_loc_desc(struct radv_shader_args *args, int idx, uint8_t *sgpr_idx) in set_loc_desc() argument 65 &args->shader_info->user_sgprs_locs; in set_loc_desc() 79 static bool needs_view_index_sgpr(struct radv_shader_args *args, in needs_view_index_sgpr() argument 84 if (args->shader_info->needs_multiview_view_index || in needs_view_index_sgpr() 85 …(!args->options->key.vs_common_out.as_es && !args->options->key.vs_common_out.as_ls && args->optio… in needs_view_index_sgpr() 89 …if (args->shader_info->needs_multiview_view_index || (!args->options->key.vs_common_out.as_es && a… in needs_view_index_sgpr() [all …]
|
| /external/virglrenderer/src/venus/venus-protocol/ |
| D | vn_protocol_renderer_command_buffer.h | 2574 …mmandBuffers_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkAllocateCommandBuffers *args) in vn_decode_vkAllocateCommandBuffers_args_temp() argument 2576 vn_decode_VkDevice_lookup(dec, &args->device); in vn_decode_vkAllocateCommandBuffers_args_temp() 2578 args->pAllocateInfo = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pAllocateInfo)); in vn_decode_vkAllocateCommandBuffers_args_temp() 2579 if (!args->pAllocateInfo) return; in vn_decode_vkAllocateCommandBuffers_args_temp() 2580 …vn_decode_VkCommandBufferAllocateInfo_temp(dec, (VkCommandBufferAllocateInfo *)args->pAllocateInfo… in vn_decode_vkAllocateCommandBuffers_args_temp() 2582 args->pAllocateInfo = NULL; in vn_decode_vkAllocateCommandBuffers_args_temp() 2586 …const uint32_t iter_count = vn_decode_array_size(dec, (args->pAllocateInfo ? args->pAllocateInfo->… in vn_decode_vkAllocateCommandBuffers_args_temp() 2587 …args->pCommandBuffers = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pCommandBuffers) * iter_count); in vn_decode_vkAllocateCommandBuffers_args_temp() 2588 if (!args->pCommandBuffers) return; in vn_decode_vkAllocateCommandBuffers_args_temp() 2590 vn_decode_VkCommandBuffer_temp(dec, &args->pCommandBuffers[i]); in vn_decode_vkAllocateCommandBuffers_args_temp() [all …]
|
| D | vn_protocol_renderer_transport.h | 130 …treamMESA_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkSetReplyCommandStreamMESA *args) in vn_decode_vkSetReplyCommandStreamMESA_args_temp() argument 133 args->pStream = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pStream)); in vn_decode_vkSetReplyCommandStreamMESA_args_temp() 134 if (!args->pStream) return; in vn_decode_vkSetReplyCommandStreamMESA_args_temp() 135 …vn_decode_VkCommandStreamDescriptionMESA_temp(dec, (VkCommandStreamDescriptionMESA *)args->pStream… in vn_decode_vkSetReplyCommandStreamMESA_args_temp() 137 args->pStream = NULL; in vn_decode_vkSetReplyCommandStreamMESA_args_temp() 142 …eplace_vkSetReplyCommandStreamMESA_args_handle(struct vn_command_vkSetReplyCommandStreamMESA *args) in vn_replace_vkSetReplyCommandStreamMESA_args_handle() argument 144 if (args->pStream) in vn_replace_vkSetReplyCommandStreamMESA_args_handle() 145 … vn_replace_VkCommandStreamDescriptionMESA_handle((VkCommandStreamDescriptionMESA *)args->pStream); in vn_replace_vkSetReplyCommandStreamMESA_args_handle() 148 …eamMESA_reply(struct vn_cs_encoder *enc, const struct vn_command_vkSetReplyCommandStreamMESA *args) in vn_encode_vkSetReplyCommandStreamMESA_reply() argument 152 /* skip args->pStream */ in vn_encode_vkSetReplyCommandStreamMESA_reply() [all …]
|
| D | vn_protocol_renderer_pipeline_cache.h | 86 …tePipelineCache_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkCreatePipelineCache *args) in vn_decode_vkCreatePipelineCache_args_temp() argument 88 vn_decode_VkDevice_lookup(dec, &args->device); in vn_decode_vkCreatePipelineCache_args_temp() 90 args->pCreateInfo = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pCreateInfo)); in vn_decode_vkCreatePipelineCache_args_temp() 91 if (!args->pCreateInfo) return; in vn_decode_vkCreatePipelineCache_args_temp() 92 … vn_decode_VkPipelineCacheCreateInfo_temp(dec, (VkPipelineCacheCreateInfo *)args->pCreateInfo); in vn_decode_vkCreatePipelineCache_args_temp() 94 args->pCreateInfo = NULL; in vn_decode_vkCreatePipelineCache_args_temp() 100 args->pAllocator = NULL; in vn_decode_vkCreatePipelineCache_args_temp() 103 args->pPipelineCache = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pPipelineCache)); in vn_decode_vkCreatePipelineCache_args_temp() 104 if (!args->pPipelineCache) return; in vn_decode_vkCreatePipelineCache_args_temp() 105 vn_decode_VkPipelineCache(dec, args->pPipelineCache); in vn_decode_vkCreatePipelineCache_args_temp() [all …]
|
| D | vn_protocol_renderer_query_pool.h | 80 …e_vkCreateQueryPool_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkCreateQueryPool *args) in vn_decode_vkCreateQueryPool_args_temp() argument 82 vn_decode_VkDevice_lookup(dec, &args->device); in vn_decode_vkCreateQueryPool_args_temp() 84 args->pCreateInfo = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pCreateInfo)); in vn_decode_vkCreateQueryPool_args_temp() 85 if (!args->pCreateInfo) return; in vn_decode_vkCreateQueryPool_args_temp() 86 vn_decode_VkQueryPoolCreateInfo_temp(dec, (VkQueryPoolCreateInfo *)args->pCreateInfo); in vn_decode_vkCreateQueryPool_args_temp() 88 args->pCreateInfo = NULL; in vn_decode_vkCreateQueryPool_args_temp() 94 args->pAllocator = NULL; in vn_decode_vkCreateQueryPool_args_temp() 97 args->pQueryPool = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pQueryPool)); in vn_decode_vkCreateQueryPool_args_temp() 98 if (!args->pQueryPool) return; in vn_decode_vkCreateQueryPool_args_temp() 99 vn_decode_VkQueryPool(dec, args->pQueryPool); in vn_decode_vkCreateQueryPool_args_temp() [all …]
|
| D | vn_protocol_renderer_fence.h | 161 …vn_decode_vkCreateFence_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkCreateFence *args) in vn_decode_vkCreateFence_args_temp() argument 163 vn_decode_VkDevice_lookup(dec, &args->device); in vn_decode_vkCreateFence_args_temp() 165 args->pCreateInfo = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pCreateInfo)); in vn_decode_vkCreateFence_args_temp() 166 if (!args->pCreateInfo) return; in vn_decode_vkCreateFence_args_temp() 167 vn_decode_VkFenceCreateInfo_temp(dec, (VkFenceCreateInfo *)args->pCreateInfo); in vn_decode_vkCreateFence_args_temp() 169 args->pCreateInfo = NULL; in vn_decode_vkCreateFence_args_temp() 175 args->pAllocator = NULL; in vn_decode_vkCreateFence_args_temp() 178 args->pFence = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pFence)); in vn_decode_vkCreateFence_args_temp() 179 if (!args->pFence) return; in vn_decode_vkCreateFence_args_temp() 180 vn_decode_VkFence(dec, args->pFence); in vn_decode_vkCreateFence_args_temp() [all …]
|
| D | vn_protocol_renderer_private_data_slot.h | 74 …ivateDataSlot_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkCreatePrivateDataSlot *args) in vn_decode_vkCreatePrivateDataSlot_args_temp() argument 76 vn_decode_VkDevice_lookup(dec, &args->device); in vn_decode_vkCreatePrivateDataSlot_args_temp() 78 args->pCreateInfo = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pCreateInfo)); in vn_decode_vkCreatePrivateDataSlot_args_temp() 79 if (!args->pCreateInfo) return; in vn_decode_vkCreatePrivateDataSlot_args_temp() 80 … vn_decode_VkPrivateDataSlotCreateInfo_temp(dec, (VkPrivateDataSlotCreateInfo *)args->pCreateInfo); in vn_decode_vkCreatePrivateDataSlot_args_temp() 82 args->pCreateInfo = NULL; in vn_decode_vkCreatePrivateDataSlot_args_temp() 88 args->pAllocator = NULL; in vn_decode_vkCreatePrivateDataSlot_args_temp() 91 args->pPrivateDataSlot = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pPrivateDataSlot)); in vn_decode_vkCreatePrivateDataSlot_args_temp() 92 if (!args->pPrivateDataSlot) return; in vn_decode_vkCreatePrivateDataSlot_args_temp() 93 vn_decode_VkPrivateDataSlot(dec, args->pPrivateDataSlot); in vn_decode_vkCreatePrivateDataSlot_args_temp() [all …]
|
| D | vn_protocol_renderer_event.h | 74 …vn_decode_vkCreateEvent_args_temp(struct vn_cs_decoder *dec, struct vn_command_vkCreateEvent *args) in vn_decode_vkCreateEvent_args_temp() argument 76 vn_decode_VkDevice_lookup(dec, &args->device); in vn_decode_vkCreateEvent_args_temp() 78 args->pCreateInfo = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pCreateInfo)); in vn_decode_vkCreateEvent_args_temp() 79 if (!args->pCreateInfo) return; in vn_decode_vkCreateEvent_args_temp() 80 vn_decode_VkEventCreateInfo_temp(dec, (VkEventCreateInfo *)args->pCreateInfo); in vn_decode_vkCreateEvent_args_temp() 82 args->pCreateInfo = NULL; in vn_decode_vkCreateEvent_args_temp() 88 args->pAllocator = NULL; in vn_decode_vkCreateEvent_args_temp() 91 args->pEvent = vn_cs_decoder_alloc_temp(dec, sizeof(*args->pEvent)); in vn_decode_vkCreateEvent_args_temp() 92 if (!args->pEvent) return; in vn_decode_vkCreateEvent_args_temp() 93 vn_decode_VkEvent(dec, args->pEvent); in vn_decode_vkCreateEvent_args_temp() [all …]
|
| /external/python/cpython2/Lib/bsddb/ |
| D | dbobj.py | 39 def __init__(self, *args, **kwargs): argument 40 self._cobj = db.DBEnv(*args, **kwargs) 42 def close(self, *args, **kwargs): argument 43 return self._cobj.close(*args, **kwargs) 44 def open(self, *args, **kwargs): argument 45 return self._cobj.open(*args, **kwargs) 46 def remove(self, *args, **kwargs): argument 47 return self._cobj.remove(*args, **kwargs) 48 def set_shm_key(self, *args, **kwargs): argument 49 return self._cobj.set_shm_key(*args, **kwargs) [all …]
|
| /external/ComputeLibrary/src/core/NEON/kernels/arm_gemm/ |
| D | gemm_fp32.cpp | 87 … [](const GemmArgs &args) { return args._Msize==1 && args._nbatches>1 && !args._indirect_input; }, in __anonf5bf1eed0102() 89 [](const GemmArgs &args) { return new GemvBatched<float, float>(args); } in __anonf5bf1eed0202() 97 [](const GemmArgs &args) { return args._fast_mode && args._ci->has_bf16(); }, in __anonf5bf1eed0302() 98 …[](const GemmArgs &args) { return GemmInterleaved<cls_a64_interleaved_bf16fp32_mmla_8x12, float, f… in __anonf5bf1eed0402() 99 …[](const GemmArgs &args) { return new GemmInterleaved<cls_a64_interleaved_bf16fp32_mmla_8x12, floa… in __anonf5bf1eed0502() 105 [](const GemmArgs &args) { return args._fast_mode && args._ci->has_bf16(); }, in __anonf5bf1eed0602() 106 …[](const GemmArgs &args) { return GemmHybridIndirect<cls_a64_hybrid_fp32bf16fp32_mmla_6x16, float,… in __anonf5bf1eed0702() 107 …[](const GemmArgs &args) { return new GemmHybridIndirect<cls_a64_hybrid_fp32bf16fp32_mmla_6x16, fl… in __anonf5bf1eed0802() 112 [](const GemmArgs &args) { return args._fast_mode && args._ci->has_bf16(); }, in __anonf5bf1eed0902() 113 …[](const GemmArgs &args) { return GemmHybridIndirect<cls_a64_hybrid_fp32bf16fp32_mmla_4x24, float,… in __anonf5bf1eed0a02() [all …]
|
| D | gemm_qint8.cpp | 77 …rgs &args, const Requantize32 &qp) { return args._ci->has_sme2() && quant_hybrid_asymmetric(qp) &&… in __anon91672d860102() 79 …nst GemmArgs &args, const Requantize32 &qp) { return new GemvPretransposed<cls_sme2_gemv_s8qa_dot_… in __anon91672d860202() 84 …[](const GemmArgs &args, const Requantize32 &qp) { return args._ci->has_sme2() && args._maxthreads… in __anon91672d860302() 85 …[](const GemmArgs &args, const Requantize32 &) { const auto VL = sme::get_vector_length<int32_t>(); in __anon91672d860402() 86 … return args._Msize <= VL || (2*VL < args._Msize && args._Msize <= 3*VL); }, in __anon91672d860402() 87 …args, const Requantize32 &qp) { return new GemmInterleavedPretransposedNoMergeQuantizedInline<cls_… in __anon91672d860502() 92 …[](const GemmArgs &args, const Requantize32 &qp) { return args._ci->has_sme2() && args._maxthreads… in __anon91672d860602() 93 …[](const GemmArgs &args, const Requantize32 &) { const auto VL = sme::get_vector_length<int32_t>(); in __anon91672d860702() 94 … return args._Nsize <= VL || (2*VL < args._Nsize && args._Nsize <= 3*VL); }, in __anon91672d860702() 95 …args, const Requantize32 &qp) { return new GemmInterleavedPretransposedNoMergeQuantizedInline<cls_… in __anon91672d860802() [all …]
|
| D | gemm_quint8.cpp | 73 …rgs &args, const Requantize32 &qp) { return args._ci->has_sme2() && quant_hybrid_asymmetric(qp) &&… in __anon5588c93b0102() 75 …st GemmArgs &args, const Requantize32 &qp) { return new GemvPretransposed<cls_sme2_gemv_u8qa_dot_1… in __anon5588c93b0202() 80 …[](const GemmArgs &args, const Requantize32 &qp) { return args._ci->has_sme2() && args._maxthreads… in __anon5588c93b0302() 81 …[](const GemmArgs &args, const Requantize32 &) { const auto VL = sme::get_vector_length<uint32_t>(… in __anon5588c93b0402() 82 … return args._Msize <= VL || (2*VL < args._Msize && args._Msize <= 3*VL); }, in __anon5588c93b0402() 83 …args, const Requantize32 &qp) { return new GemmInterleavedPretransposedNoMergeQuantizedInline<cls_… in __anon5588c93b0502() 88 …[](const GemmArgs &args, const Requantize32 &qp) { return args._ci->has_sme2() && args._maxthreads… in __anon5588c93b0602() 89 …[](const GemmArgs &args, const Requantize32 &) { const auto VL = sme::get_vector_length<int32_t>(); in __anon5588c93b0702() 90 … return args._Nsize <= VL || (2*VL < args._Nsize && args._Nsize <= 3*VL); }, in __anon5588c93b0702() 91 …args, const Requantize32 &qp) { return new GemmInterleavedPretransposedNoMergeQuantizedInline<cls_… in __anon5588c93b0802() [all …]
|
| D | gemm_int8.cpp | 65 [](const GemmArgs &args) { return args._ci->has_sme2(); }, in __anonb89c07f50102() 66 [](const GemmArgs &args) { const auto VL = sme::get_vector_length<int32_t>(); in __anonb89c07f50202() 67 … return args._Msize <= VL || (2*VL < args._Msize && args._Msize <= 3*VL); }, in __anonb89c07f50202() 68 …[](const GemmArgs &args) { return new GemmInterleavedNoMerge<cls_sme2_interleaved_nomerge_s8s32_mo… in __anonb89c07f50302() 73 [](const GemmArgs &args) { return args._ci->has_sme2(); }, in __anonb89c07f50402() 74 [](const GemmArgs &args) { const auto VL = sme::get_vector_length<int32_t>(); in __anonb89c07f50502() 75 … return args._Nsize <= VL || (2*VL < args._Nsize && args._Nsize <= 3*VL); }, in __anonb89c07f50502() 76 …[](const GemmArgs &args) { return new GemmInterleavedNoMerge<cls_sme2_interleaved_nomerge_s8s32_mo… in __anonb89c07f50602() 81 [](const GemmArgs &args) { return args._ci->has_sme2(); }, in __anonb89c07f50702() 83 …[](const GemmArgs &args) { return new GemmInterleavedNoMerge<cls_sme2_interleaved_nomerge_s8s32_mo… in __anonb89c07f50802() [all …]
|
| D | gemm_bf16.cpp | 78 …[](const GemmArgs &args) { return args._ci->has_sme2() && args._Msize==1 && args._nbatches==1 && !… in __anon3e8f7ca10102() 80 …[](const GemmArgs &args) { return new GemvPretransposed<cls_sme2_gemv_bf16fp32_dot_16VL, bfloat16,… in __anon3e8f7ca10202() 85 [](const GemmArgs &args) { return args._ci->has_sme2(); }, in __anon3e8f7ca10302() 86 [](const GemmArgs &args) { const auto VL = sme::get_vector_length<float>(); in __anon3e8f7ca10402() 87 … return args._Msize <= VL || (2*VL < args._Msize && args._Msize <= 3*VL); }, in __anon3e8f7ca10402() 88 …[](const GemmArgs &args) { return new GemmInterleavedNoMerge<cls_sme2_interleaved_nomerge_bf16fp32… in __anon3e8f7ca10502() 93 [](const GemmArgs &args) { return args._ci->has_sme2(); }, in __anon3e8f7ca10602() 94 [](const GemmArgs &args) { const auto VL = sme::get_vector_length<float>(); in __anon3e8f7ca10702() 95 … return args._Nsize <= VL || (2*VL < args._Nsize && args._Nsize <= 3*VL); }, in __anon3e8f7ca10702() 96 …[](const GemmArgs &args) { return new GemmInterleavedNoMerge<cls_sme2_interleaved_nomerge_bf16fp32… in __anon3e8f7ca10802() [all …]
|
| D | gemm_uint8.cpp | 55 [](const GemmArgs &args) { return args._ci->has_svei8mm(); }, in __anon6358f18a0102() 56 …](const GemmArgs &args) { return GemmHybridIndirect<cls_sve_hybrid_u8u32_mmla_6x4VL, uint8_t, uint… in __anon6358f18a0202() 57 …[](const GemmArgs &args) { return new GemmHybridIndirect<cls_sve_hybrid_u8u32_mmla_6x4VL, uint8_t,… in __anon6358f18a0302() 62 [](const GemmArgs &args) { return args._ci->has_svei8mm() && (args._Ksize>8); }, in __anon6358f18a0402() 63 …(const GemmArgs &args) { return GemmInterleaved<cls_sve_interleaved_u8u32_mmla_8x3VL, uint8_t, uin… in __anon6358f18a0502() 64 …[](const GemmArgs &args) { return new GemmInterleaved<cls_sve_interleaved_u8u32_mmla_8x3VL, uint8_… in __anon6358f18a0602() 69 …[](const GemmArgs &args) { return args._ci->has_sve() && args._Ksize<=64 && !args._indirect_input;… in __anon6358f18a0702() 70 [](const GemmArgs &args) { return !(args._ci->has_svei8mm() || args._ci->has_i8mm()); }, in __anon6358f18a0802() 71 …[](const GemmArgs &args) { return new GemmHybrid<cls_sve_smallK_hybrid_u8u32_dot_8x1VL, uint8_t, u… in __anon6358f18a0902() 76 [](const GemmArgs &args) { return args._ci->has_sve(); }, in __anon6358f18a0a02() [all …]
|
| /external/zstd/contrib/pzstd/test/ |
| D | OptionsTest.cpp | 83 template <typename... Args> 84 std::array<const char *, sizeof...(Args) + 1> makeArray(Args... args) { in makeArray() argument 85 return {{nullptr, args...}}; in makeArray() 91 auto args = makeArray("--processes", "5", "-o", "x", "y", "-f"); in TEST() local 92 EXPECT_SUCCESS(options.parse(args.size(), args.data())); in TEST() 99 auto args = makeArray("-p", "1", "input", "-19"); in TEST() local 100 EXPECT_SUCCESS(options.parse(args.size(), args.data())); in TEST() 107 auto args = in TEST() local 109 EXPECT_SUCCESS(options.parse(args.size(), args.data())); in TEST() 116 auto args = makeArray("--processes", "100", "hello.zst", "--decompress", in TEST() local [all …]
|
| /external/XNNPACK/bench/ |
| D | spmm.h | 31 b->Args({56 * 56, 36, 24}); in ShuffleNetV1G1SpmmArguments() 32 b->Args({28 * 28, 120, 36}); in ShuffleNetV1G1SpmmArguments() 33 b->Args({28 * 28, 36, 144}); in ShuffleNetV1G1SpmmArguments() 34 b->Args({28 * 28, 144, 36}); in ShuffleNetV1G1SpmmArguments() 35 b->Args({28 * 28, 72, 144}); in ShuffleNetV1G1SpmmArguments() 36 b->Args({14 * 14, 144, 72}); in ShuffleNetV1G1SpmmArguments() 37 b->Args({14 * 14, 72, 288}); in ShuffleNetV1G1SpmmArguments() 38 b->Args({14 * 14, 288, 72}); in ShuffleNetV1G1SpmmArguments() 39 b->Args({14 * 14, 144, 288}); in ShuffleNetV1G1SpmmArguments() 40 b->Args({ 7 * 7, 288, 144}); in ShuffleNetV1G1SpmmArguments() [all …]
|
| /external/perfetto/ui/src/common/ |
| D | actions.ts | 148 openTraceFromFile(state: StateDraft, args: {file: File}): void { 154 source: {type: 'FILE', file: args.file}, 158 openTraceFromBuffer(state: StateDraft, args: PostedTrace): void { 164 source: {type: 'ARRAY_BUFFER', ...args}, 168 openTraceFromUrl(state: StateDraft, args: {url: string}): void { 174 source: {type: 'URL', url: args.url}, 188 setTraceUuid(state: StateDraft, args: {traceUuid: string}) { 189 state.traceUuid = args.traceUuid; 221 addTracks(state: StateDraft, args: {tracks: AddTrackArgs[]}) { 222 args.tracks.forEach((track) => { [all …]
|
| /external/tensorflow/tensorflow/compiler/xla/pjrt/c/ |
| D | pjrt_c_api_wrapper_impl.cc | 60 void PJRT_Error_Destroy(PJRT_Error_Destroy_Args* args) { in PJRT_Error_Destroy() argument 63 args->struct_size); in PJRT_Error_Destroy() 67 if (args->struct_size >= PJRT_STRUCT_SIZE(PJRT_Error_Destroy_Args, error)) { in PJRT_Error_Destroy() 68 delete args->error; in PJRT_Error_Destroy() 72 void PJRT_Error_Message(PJRT_Error_Message_Args* args) { in PJRT_Error_Message() argument 75 args->struct_size); in PJRT_Error_Message() 79 if (args->struct_size >= PJRT_STRUCT_SIZE(PJRT_Error_Destroy_Args, error)) { in PJRT_Error_Message() 80 const xla::Status* status = &args->error->status; in PJRT_Error_Message() 81 args->message = status->error_message().data(); in PJRT_Error_Message() 82 args->message_size = status->error_message().size(); in PJRT_Error_Message() [all …]
|
| /external/fonttools/Lib/fontTools/cffLib/ |
| D | specializer.py | 79 # replace the blend op args on the stack with a single list 80 # containing all the blend op args. 101 # lenBlendStack has the number of args represented by the last blend 102 # arg and all the preceding args. We need to now add the number of 103 # args following the last blend arg. 124 def _flattenBlendArgs(args): argument 126 for arg in args: 138 for op,args in commands: 139 if any(isinstance(arg, list) for arg in args): 140 args = _flattenBlendArgs(args) [all …]
|
| /external/cronet/buildtools/third_party/libc++/trunk/test/std/experimental/memory/memory.polymorphic.allocator.class/memory.polymorphic.allocator.mem/ |
| D | construct_types.pass.cpp | 20 // template <class U, class ...Args> 21 // void polymorphic_allocator<T>::construct(U *, Args &&...) 47 template <class ...Args> 48 void construct(Args&&... args) { in construct() 49 A.construct(ptr, std::forward<Args>(args)...); in construct() 59 template <class T, class ...Args> 60 bool doTest(UsesAllocatorType UAExpect, Args&&... args) in doTest() argument 64 TH.construct(std::forward<Args>(args)...); in doTest() 65 return checkConstruct<Args&&...>(*TH.ptr, UAExpect, &TH.R); in doTest() 70 template <class T, class ...Args> [all …]
|
| /external/libcxx/test/std/experimental/memory/memory.polymorphic.allocator.class/memory.polymorphic.allocator.mem/ |
| D | construct_types.pass.cpp | 17 // template <class U, class ...Args> 18 // void polymorphic_allocator<T>::construct(U *, Args &&...) 42 template <class ...Args> 43 void construct(Args&&... args) { in construct() 44 A.construct(ptr, std::forward<Args>(args)...); in construct() 54 template <class T, class ...Args> 55 bool doTest(UsesAllocatorType UAExpect, Args&&... args) in doTest() argument 59 TH.construct(std::forward<Args>(args)...); in doTest() 60 return checkConstruct<Args&&...>(*TH.ptr, UAExpect, &TH.R); in doTest() 65 template <class T, class ...Args> [all …]
|