Home
last modified time | relevance | path

Searched defs:batched (Results 1 – 25 of 60) sorted by relevance

123

/external/pytorch/aten/src/ATen/test/
Dlegacy_vmap_test.cpp105 auto* batched = maybeGetBatchedImpl(tensor); in TEST() local
127 auto* batched = maybeGetBatchedImpl(tensor); in TEST() local
137 auto* batched = maybeGetBatchedImpl(tensor); in TEST() local
145 auto* batched = maybeGetBatchedImpl(tensor); in TEST() local
155 auto* batched = maybeGetBatchedImpl(tensor); in TEST() local
164 auto* batched = maybeGetBatchedImpl(tensor); in TEST() local
178 auto batched = addBatchDim(tensor, /*lvl*/1, /*dim*/0); in TEST() local
199 auto batched = makeBatched(tensor, bdims); in TEST() local
208 auto batched = makeBatched(tensor, bdims); in TEST() local
218 auto batched = makeBatched(tensor, bdims); in TEST() local
[all …]
/external/pytorch/aten/src/ATen/functorch/
DLegacyVmapTransforms.cpp17 static Tensor permuteBatchDimsToFront(const BatchedTensorImpl* batched) { in permuteBatchDimsToFront()
38 auto* batched = maybeGetBatchedImpl(logical_tensor); in logicalToPhysical() local
126 auto* batched = maybeGetBatchedImpl(logical_tensor); in logicalToPhysical() local
142 auto* batched = maybeGetBatchedImpl(logical_tensor); in logicalToPhysical() local
178 auto* batched = maybeGetBatchedImpl(logical_tensor); in logicalToPhysical() local
197 auto* batched = maybeGetBatchedImpl(logical_tensor); in logicalToPhysical() local
DLegacyBatchingRegistrations.cpp107 auto* batched = maybeGetBatchedImpl(self); in squeeze_dims__batching_rule() local
153 auto* batched = maybeGetBatchedImpl(self); in squeeze__batching_rule() local
190 auto* batched = maybeGetBatchedImpl(self); in unsqueeze__batching_rule() local
210 auto* batched = maybeGetBatchedImpl(self); in transpose__batching_rule() local
550 const auto* batched = unsafeGetBatchedImpl(tensor); in cat_batching_rule() local
565 const auto* batched = unsafeGetBatchedImpl(tensor); in cat_batching_rule() local
DBatchedFallback.cpp127 const auto* batched = maybeGetBatchedImpl(tensor); in batchedTensorInplaceForLoopFallback() local
302 const auto* batched = maybeGetBatchedImpl(tensor); in batchedTensorForLoopFallback() local
443 const auto* batched = maybeGetBatchedImpl(tensor); in batchedNestedTensorForLoopFallback() local
/external/pytorch/torch/csrc/functorch/
Dinit.cpp31 const auto* batched = maybeGetBatchedImpl(self); in has_level() local
96 const BatchedTensorImpl* batched, in remove_existing_batch_dim()
161 const auto* batched = maybeGetBatchedImpl(self); in _remove_batch_dim() local
304 auto* batched = maybeGetBatchedImpl(tensor); in is_batchedtensor() local
323 auto* batched = maybeGetBatchedImpl(tensor); in get_unwrapped() local
340 auto* batched = maybeGetBatchedImpl(tensor); in maybe_get_level() local
362 auto* batched = maybeGetBatchedImpl(tensor); in maybe_get_bdim() local
435 auto* batched = maybeGetBatchedImpl(tensor); in unwrapBatched() local
/external/pytorch/aten/src/ATen/
DLegacyVmapTransforms.cpp20 static Tensor permuteBatchDimsToFront(BatchedTensorImpl* batched) { in permuteBatchDimsToFront()
44 auto* batched = maybeGetBatchedImpl(logical_tensor); in logicalToPhysical() local
108 auto* batched = maybeGetBatchedImpl(self); in getPhysicalTensorAndLevels() local
194 auto* batched = maybeGetBatchedImpl(logical_tensor); in logicalToPhysical() local
244 auto* batched = maybeGetBatchedImpl(tensor); in getLevelsAndLargestLogicalDim() local
/external/pytorch/aten/src/ATen/native/
DLegacyBatching.cpp19 const auto* batched = maybeGetBatchedImpl(self); in has_level() local
49 const BatchedTensorImpl* batched, int64_t level) { in remove_existing_batch_dim()
115 const auto* batched = maybeGetBatchedImpl(self); in _remove_batch_dim() local
/external/webrtc/logging/rtc_event_log/events/
Drtc_event_rtcp_packet_outgoing.h50 bool batched, in Parse()
Drtc_event_rtcp_packet_incoming.h50 bool batched, in Parse()
Drtc_event_alr_state.cc33 bool batched, in Parse()
Drtc_event_end_log.cc34 bool batched, in Parse()
Drtc_event_dtls_transport_state.h58 bool batched, in Parse()
Drtc_event_begin_log.cc41 bool batched, in Parse()
Drtc_event_audio_network_adaptation.h64 bool batched, in Parse()
Drtc_event_audio_send_stream_config.h60 bool batched, in Parse()
Drtc_event_video_receive_stream_config.h61 bool batched, in Parse()
Drtc_event_dtls_writable_state.h58 bool batched, in Parse()
Drtc_event_audio_receive_stream_config.h61 bool batched, in Parse()
Drtc_event_video_send_stream_config.h61 bool batched, in Parse()
Drtc_event_remote_estimate.h57 bool batched, in Parse()
Drtc_event_rtp_packet_incoming.h75 bool batched, in Parse()
Drtc_event_route_change.h61 bool batched, in Parse()
Drtc_event_probe_result_success.h65 bool batched, in Parse()
Drtc_event_audio_playout.h61 bool batched, in Parse()
/external/tensorflow/tensorflow/core/kernels/
Dragged_tensor_to_variant_op_test.h47 const std::vector<VALUE_TYPE>& ragged_values, const bool batched) { in BuildEncodeRaggedTensorGraph()

123