| /external/tensorflow/tensorflow/python/keras/ |
| D | callbacks.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 15 # pylint: disable=g-import-not-at-top 16 # pylint: disable=g-classes-have-attributes 83 mode=ModeKeys.TRAIN): argument 90 batch_size: Number of samples per batch. 95 count_mode: One of 'steps' or 'samples'. Per-batch or per-sample count. 96 mode: String. One of ModeKeys.TRAIN, ModeKeys.TEST, or ModeKeys.PREDICT. 97 Which loop mode to configure callbacks for. 110 if mode == ModeKeys.TRAIN: 118 callback_model = model._get_callback_model() # pylint: disable=protected-access [all …]
|
| /external/mesa3d/src/gallium/drivers/zink/ |
| D | zink_draw.c | 19 struct zink_batch *batch, in allocate_descriptor_set() argument 22 assert(batch->descs_left >= prog->num_descriptors); in allocate_descriptor_set() 27 dsai.descriptorPool = batch->descpool; in allocate_descriptor_set() 29 dsai.pSetLayouts = &prog->dsl; in allocate_descriptor_set() 32 if (vkAllocateDescriptorSets(screen->dev, &dsai, &desc_set) != VK_SUCCESS) { in allocate_descriptor_set() 37 batch->descs_left -= prog->num_descriptors; in allocate_descriptor_set() 50 * - from VK_EXT_transform_feedback spec in zink_emit_xfb_counter_barrier() 55 for (unsigned i = 0; i < ctx->num_so_targets; i++) { in zink_emit_xfb_counter_barrier() 56 struct zink_so_target *t = zink_so_target(ctx->so_targets[i]); in zink_emit_xfb_counter_barrier() 57 if (t->counter_buffer_valid) { in zink_emit_xfb_counter_barrier() [all …]
|
| /external/tensorflow/tensorflow/core/kernels/image/ |
| D | image_ops.h | 7 http://www.apache.org/licenses/LICENSE-2.0 33 enum Mode { FILL_REFLECT, FILL_WRAP, FILL_CONSTANT, FILL_NEAREST }; enum 40 template <typename Device, Mode M> 46 struct MapCoordinate<Device, Mode::FILL_REFLECT> { 57 in_coord = sz2 * static_cast<DenseIndex>(-in_coord / sz2) + in_coord; 59 in_coord = (in_coord < -len) ? in_coord + sz2 : -in_coord - 1; 61 } else if (in_coord > len - 1) { 66 in_coord -= sz2 * static_cast<DenseIndex>(in_coord / sz2); 68 in_coord = sz2 - in_coord - 1; 74 return Eigen::internal::scalar_clamp_op<float>(0.0f, len - 1)(in_coord); [all …]
|
| /external/tensorflow/tensorflow/python/distribute/ |
| D | custom_training_loop_input_test.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 80 mode=["eager"] 101 mode=["eager"] 123 mode=["eager"])) 125 dataset = get_dataset_from_tensor_slices([5., 6., 7., 8.]).batch(2) 140 distribution=strategy_combinations.all_strategies, mode=["eager"])) 143 dataset = get_dataset_from_tensor_slices(data).batch(2) 160 distribution=strategy_combinations.all_strategies, mode=["eager"])) 165 8, output_type=dtypes.int32).batch(global_batch_size) 185 distribution=strategy_combinations.tpu_strategies, mode=["eager"])) [all …]
|
| D | input_lib_test.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 61 # The passed input_context is to create a sharded dataset in between-graph 73 # MultiWorkerMirroredStrategy. It doesn't apply to in-graph case where 85 # Note: `input_workers.num_workers` is always 1 in between-graph 227 # After re-initializing the iterator, should be able to iterate again. 258 # re-initializing the iterator 286 mode=["eager"], 312 mode=["graph", "eager"], 338 mode=["eager"], 361 mode=["graph", "eager"], [all …]
|
| D | input_lib_type_spec_test.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 49 mode=["eager"], 61 dataset = dataset_ops.DatasetV2.range(10).batch(2) 81 mode=["eager"], 94 dataset = dataset_ops.DatasetV2.range(10).batch(2) 114 mode=["eager"], 138 dataset = dataset_ops.DatasetV2.range(10).batch( 159 mode=["eager"], 180 np.ones([9, 12]).astype(np.float32)).batch( 199 mode=["eager"], [all …]
|
| /external/mesa3d/src/gallium/drivers/iris/ |
| D | iris_draw.c | 46 /* We don't need to worry about adjacency - it can only be used with in prim_is_points_or_lines() 49 return draw->mode == PIPE_PRIM_POINTS || in prim_is_points_or_lines() 50 draw->mode == PIPE_PRIM_LINES || in prim_is_points_or_lines() 51 draw->mode == PIPE_PRIM_LINE_LOOP || in prim_is_points_or_lines() 52 draw->mode == PIPE_PRIM_LINE_STRIP; in prim_is_points_or_lines() 56 * Record the current primitive mode and restart information, flagging 66 struct iris_screen *screen = (struct iris_screen *)ice->ctx.screen; in iris_update_draw_info() 67 const struct brw_compiler *compiler = screen->compiler; in iris_update_draw_info() 69 if (ice->state.prim_mode != info->mode) { in iris_update_draw_info() 70 ice->state.prim_mode = info->mode; in iris_update_draw_info() [all …]
|
| /external/tensorflow/tensorflow/python/keras/distribute/ |
| D | distributed_training_utils_v1.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 16 # pylint:disable=protected-access 136 def unwrap_output_dict(strategy, grouped_outputs, mode): argument 138 if mode == ModeKeys.PREDICT: 156 # We only do this in eager mode for now since this function is used in 157 # both graph and eager mode and in the graph case we currently don't use 204 # We only do this in eager mode for now since this function is used in 205 # both graph and eager mode and in the graph case we currently don't use 229 # pylint: disable=g-complex-comprehension 377 all_variables = backend._get_variables(backend.get_graph()) # pylint: disable=protected-access [all …]
|
| /external/tensorflow/tensorflow/python/compiler/tensorrt/test/ |
| D | trt_mode_test.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 15 """Model script to test TF-TensorRT integration.""" 17 from unittest import SkipTest # pylint: disable=g-importing-member 27 """Test squeeze on batch dim and some unary operations in TF-TRT.""" 43 "non-calibration") 50 In explicit batch mode, TensorRT can convert the whole graph. In this mode 51 it is possible to manipulate the batch dimension using the squeeze op. 53 In implicit batch mode TensorRT cannot convert the whole graph. We are not 54 allowed to manipulate (squeeze) the first dimension in implicit batch mode. 64 # The first dimension of the input is squeezed and the batch size for the [all …]
|
| /external/mesa3d/src/gallium/drivers/freedreno/ |
| D | freedreno_draw.c | 46 resource_read(struct fd_batch *batch, struct pipe_resource *prsc) in resource_read() argument 50 fd_batch_resource_read(batch, fd_resource(prsc)); in resource_read() 54 resource_written(struct fd_batch *batch, struct pipe_resource *prsc) in resource_written() argument 58 fd_batch_resource_write(batch, fd_resource(prsc)); in resource_written() 62 batch_draw_tracking(struct fd_batch *batch, const struct pipe_draw_info *info) in batch_draw_tracking() argument 64 struct fd_context *ctx = batch->ctx; in batch_draw_tracking() 65 struct pipe_framebuffer_state *pfb = &batch->framebuffer; in batch_draw_tracking() 68 /* NOTE: needs to be before resource_written(batch->query_buf), otherwise in batch_draw_tracking() 71 fd_batch_set_stage(batch, FD_STAGE_DRAW); in batch_draw_tracking() 77 fd_screen_lock(ctx->screen); in batch_draw_tracking() [all …]
|
| /external/cronet/net/third_party/quiche/src/quiche/quic/core/ |
| D | quic_packet_writer.h | 2 // Use of this source code is governed by a BSD-style license that can be 24 // Returns a heap-allocated copy of |this|. 45 // - PassThrough mode. This is the default mode. Caller calls WritePacket with 46 // caller-allocated packet buffer. Unless the writer is blocked, each call to 49 // - Batch mode. In this mode, a call to WritePacket may not cause a packet to 56 // In Batch mode, a writer manages an internal buffer, which is large enough to 58 // caller-allocated packet buffer, the writer will memcpy the buffer into the 67 // PassThrough mode: 68 // Sends the packet out to the peer, with some optional per-packet options. 73 // Batch mode: [all …]
|
| /external/tensorflow/tensorflow/python/keras/engine/ |
| D | training_arrays_v1.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 16 # pylint: disable=protected-access 37 from scipy.sparse import issparse # pylint: disable=g-import-not-at-top 58 mode=ModeKeys.TRAIN, 70 batch_size: Integer batch size or None if unknown. 72 verbose: 0, 1, or 2. Verbosity mode. 99 mode: One of ModeKeys.TRAIN/ModeKeys.TEST/ModeKeys.PREDICT. 114 - In TRAIN mode: `History` object. 115 - In TEST mode: Evaluation metrics. 116 - In PREDICT mode: Outputs of the Model called on inputs. [all …]
|
| /external/igt-gpu-tools/tests/i915/ |
| D | gem_streaming_writes.c | 24 * Chris Wilson <chris@chris-wilson.co.uk> 55 #define BATCH 2 macro 62 static void test_streaming(int fd, int mode, int sync) in test_streaming() argument 75 } *batch; in test_streaming() local 82 switch (mode) { in test_streaming() 137 exec[BATCH].relocation_count = 2; in test_streaming() 143 batch = malloc(sizeof(*batch) * (OBJECT_SIZE / CHUNK_SIZE / 64)); in test_streaming() 147 batch[i].handle = gem_create(fd, 4096); in test_streaming() 148 batch[i].offset = 0; in test_streaming() 150 base = gem_mmap__cpu(fd, batch[i].handle, 0, 4096, PROT_WRITE); in test_streaming() [all …]
|
| D | gem_mocs_settings.c | 35 #define GEN9_NUM_MOCS_ENTRIES 62 /* 62 out of 64 - 63 & 64 are reserved. */ 36 #define GEN11_NUM_MOCS_ENTRIES 64 /* 63-64 are reserved, but configured. */ 147 [3 ... GEN9_NUM_MOCS_ENTRIES - 1] = MOCS_PTE, 151 [0 ... GEN9_NUM_MOCS_ENTRIES - 1] = { 0x00003FFF, 0x003F, 0x1 }, 158 [3 ... GEN9_NUM_MOCS_ENTRIES - 1] = MOCS_PTE, 162 [0 ... GEN9_NUM_MOCS_ENTRIES - 1] = { 0x00007FFF, 0x003F, 0x1 }, 166 [0 ... GEN9_NUM_MOCS_ENTRIES - 1] = 0xFFFFFFFF, 181 table->size = ARRAY_SIZE(dirty_skylake_mocs_table); in get_mocs_settings() 182 table->table = dirty_skylake_mocs_table; in get_mocs_settings() 184 table->size = ARRAY_SIZE(skylake_mocs_table); in get_mocs_settings() [all …]
|
| D | gem_exec_parallel.c | 78 uint32_t batch[16]; in thread() local 81 pthread_mutex_lock(t->mutex); in thread() 82 while (*t->go == 0) in thread() 83 pthread_cond_wait(t->cond, t->mutex); in thread() 84 pthread_mutex_unlock(t->mutex); in thread() 86 if (t->flags & FDS) in thread() 89 fd = t->fd; in thread() 92 batch[i] = MI_STORE_DWORD_IMM | (t->gen < 6 ? 1 << 22 : 0); in thread() 93 if (t->gen >= 8) { in thread() 94 batch[++i] = 4*t->id; in thread() [all …]
|
| D | gem_exec_whisper.c | 65 gem_read(fd, handle, reloc->offset, &target, 8); in verify_reloc() 67 gem_read(fd, handle, reloc->offset, &target, 4); in verify_reloc() 69 reloc->presumed_offset + reloc->delta); in verify_reloc() 93 uint32_t *batch; in init_hang() local 96 h->fd = drm_open_driver(DRIVER_INTEL); in init_hang() 97 igt_allow_hang(h->fd, 0, 0); in init_hang() 99 gen = intel_gen(intel_get_drm_devid(h->fd)); in init_hang() 101 memset(&h->execbuf, 0, sizeof(h->execbuf)); in init_hang() 102 h->execbuf.buffers_ptr = to_user_pointer(&h->obj); in init_hang() 103 h->execbuf.buffer_count = 1; in init_hang() [all …]
|
| /external/mesa3d/src/gallium/drivers/freedreno/a6xx/ |
| D | fd6_vsc.c | 35 * https://github.com/freedreno/freedreno/wiki/Visibility-Stream-Format 45 * https://github.com/freedreno/freedreno/wiki/Visibility-Stream-Format#numbers 52 return n + (n - 1); in number_size_bits() 57 * https://github.com/freedreno/freedreno/wiki/Visibility-Stream-Format#bitfields 69 unsigned vtx_per_prim = (info->mode == PIPE_PRIM_MAX) ? 2 : in prim_count() 70 u_vertices_per_prim(info->mode); in prim_count() 71 return (info->count * info->instance_count) / vtx_per_prim; in prim_count() 75 * The primitive stream uses a run-length encoding, where each packet contains a 79 * - The (compressed) bitfield of bins covered 80 * - The number of primitives with this bitset [all …]
|
| /external/angle/src/tests/test_utils/runner/ |
| D | README.md | 9 * multi-process execution 11 ## Command-Line Arguments 14 following additional command-line arguments: 16 * `--batch-size` limits the number of tests to run in each batch 17 * `--batch-timeout` limits the amount of time spent in each batch 18 * `--bot-mode` enables multi-process execution and test batching 19 * `--debug-test-groups` dumps the test config categories when using `bot-mode` 20 * `--filter-file` allows passing a larger `gtest_filter` via a file 21 * `--histogram-json-file` outputs a [formatted JSON file][HistogramSet] for perf dashboards 22 * `--max-processes` limits the number of simuntaneous processes [all …]
|
| /external/tensorflow/tensorflow/python/ops/ |
| D | batch_norm_benchmark.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 15 """End-to-end benchmark for batch normalization.""" 32 import tensorflow.python.ops.nn_grad # pylint: disable=unused-import 37 """Fused kernel for batch normalization.""" 40 # pylint: disable=protected-access 43 # pylint: enable=protected-access 47 # batch_norm = (tensor - mean) * tf.math.rsqrt(variance + 0.001) 52 """Python implementation of batch normalization.""" 58 batch_norm = (tensor - mean) * math_ops.rsqrt(variance + 0.001) 64 def build_graph(device, input_shape, axes, num_layers, mode, scale, train): argument [all …]
|
| /external/igt-gpu-tools/tests/ |
| D | kms_psr.c | 73 drmModeModeInfo *mode; member 84 fb_id = igt_create_fb(data->drm_fd, 64, 64, in create_cursor_fb() 86 &data->fb_white); in create_cursor_fb() 89 cr = igt_get_cairo_ctx(data->drm_fd, &data->fb_white); in create_cursor_fb() 91 igt_put_cairo_ctx(data->drm_fd, &data->fb_white, cr); in create_cursor_fb() 96 igt_display_t *display = &data->display; in setup_output() 101 drmModeConnectorPtr c = output->config.connector; in setup_output() 103 if (c->connector_type != DRM_MODE_CONNECTOR_eDP) in setup_output() 107 data->crtc_id = output->config.crtc->crtc_id; in setup_output() 108 data->output = output; in setup_output() [all …]
|
| D | kms_fence_pin_leak.c | 47 struct intel_batchbuffer *batch; in exec_nop() local 49 dst = gem_handle_to_libdrm_bo(data->bufmgr, data->drm_fd, "", handle); in exec_nop() 52 batch = intel_batchbuffer_alloc(data->bufmgr, data->devid); in exec_nop() 53 igt_assert(batch); in exec_nop() 63 intel_batchbuffer_flush_with_context(batch, context); in exec_nop() 64 intel_batchbuffer_free(batch); in exec_nop() 73 for (i = 0; i < ARRAY_SIZE(data->bos); i++) { in alloc_fence_objs() 76 bo = drm_intel_bo_alloc(data->bufmgr, "fence bo", 4096, 4096); in alloc_fence_objs() 78 gem_set_tiling(data->drm_fd, bo->handle, I915_TILING_X, 512); in alloc_fence_objs() 80 data->bos[i] = bo; in alloc_fence_objs() [all …]
|
| D | prime_busy.c | 32 pfd->events = excl ? POLLOUT : POLLIN; in prime_busy() 47 #define BATCH 1 in busy() macro 52 uint32_t *batch, *bbe; in busy() local 67 obj[BATCH].handle = gem_create(fd, size); in busy() 68 obj[BATCH].relocs_ptr = (uintptr_t)store; in busy() 69 obj[BATCH].relocation_count = ARRAY_SIZE(store); in busy() 75 pfd[BATCH].fd = prime_handle_to_fd(fd, obj[BATCH].handle); in busy() 78 batch = gem_mmap__wc(fd, obj[BATCH].handle, 0, size, PROT_WRITE); in busy() 79 gem_set_domain(fd, obj[BATCH].handle, in busy() 85 store[count].presumed_offset = -1; in busy() [all …]
|
| /external/mesa3d/src/gallium/drivers/panfrost/ |
| D | pan_context.c | 2 * Copyright (C) 2019-2020 Collabora, Ltd. 4 * Copyright © 2014-2017 Broadcom 34 #include "panfrost-quirks.h" 62 panfrost_emit_midg_tiler(struct panfrost_batch *batch, in panfrost_emit_midg_tiler() argument 66 struct panfrost_device *device = pan_device(batch->ctx->base.screen); in panfrost_emit_midg_tiler() 67 bool hierarchy = !(device->quirks & MIDGARD_NO_HIER_TILING); in panfrost_emit_midg_tiler() 68 unsigned height = batch->key.height; in panfrost_emit_midg_tiler() 69 unsigned width = batch->key.width; in panfrost_emit_midg_tiler() 88 panfrost_batch_get_polygon_list(batch, in panfrost_emit_midg_tiler() 92 t.heap_start = device->tiler_heap->ptr.gpu; in panfrost_emit_midg_tiler() [all …]
|
| /external/mesa3d/src/mesa/drivers/dri/i965/ |
| D | brw_misc_state.c | 49 * Upload pointers to the per-stage state. 57 const struct gen_device_info *devinfo = &brw->screen->devinfo; in upload_pipelined_state_pointers() 59 if (devinfo->gen == 5) { in upload_pipelined_state_pointers() 67 OUT_BATCH(_3DSTATE_PIPELINED_POINTERS << 16 | (7 - 2)); in upload_pipelined_state_pointers() 68 OUT_RELOC(brw->batch.state.bo, 0, brw->vs.base.state_offset); in upload_pipelined_state_pointers() 69 if (brw->ff_gs.prog_active) in upload_pipelined_state_pointers() 70 OUT_RELOC(brw->batch.state.bo, 0, brw->ff_gs.state_offset | 1); in upload_pipelined_state_pointers() 73 OUT_RELOC(brw->batch.state.bo, 0, brw->clip.state_offset | 1); in upload_pipelined_state_pointers() 74 OUT_RELOC(brw->batch.state.bo, 0, brw->sf.state_offset); in upload_pipelined_state_pointers() 75 OUT_RELOC(brw->batch.state.bo, 0, brw->wm.base.state_offset); in upload_pipelined_state_pointers() [all …]
|
| /external/gson/.github/workflows/ |
| D | check-api-compatibility.yml | 6 check-api-compatibility: 7 runs-on: ubuntu-latest 10 - name: Checkout old version 14 path: 'gson-old-japicmp' 16 - name: Set up JDK 11 17 uses: actions/setup-java@v3 20 java-version: '11' 23 - name: Build old version 25 cd gson-old-japicmp 27 …mvn --batch-mode --no-transfer-progress org.codehaus.mojo:versions-maven-plugin:2.11.0:set -DnewVe… [all …]
|