Home
last modified time | relevance | path

Searched +full:- +full:- +full:batch +full:- +full:mode (Results 1 – 25 of 1013) sorted by relevance

12345678910>>...41

/external/tensorflow/tensorflow/python/keras/
Dcallbacks.py7 # http://www.apache.org/licenses/LICENSE-2.0
15 # pylint: disable=g-import-not-at-top
16 # pylint: disable=g-classes-have-attributes
83 mode=ModeKeys.TRAIN): argument
90 batch_size: Number of samples per batch.
95 count_mode: One of 'steps' or 'samples'. Per-batch or per-sample count.
96 mode: String. One of ModeKeys.TRAIN, ModeKeys.TEST, or ModeKeys.PREDICT.
97 Which loop mode to configure callbacks for.
110 if mode == ModeKeys.TRAIN:
118 callback_model = model._get_callback_model() # pylint: disable=protected-access
[all …]
/external/mesa3d/src/gallium/drivers/zink/
Dzink_draw.c19 struct zink_batch *batch, in allocate_descriptor_set() argument
22 assert(batch->descs_left >= prog->num_descriptors); in allocate_descriptor_set()
27 dsai.descriptorPool = batch->descpool; in allocate_descriptor_set()
29 dsai.pSetLayouts = &prog->dsl; in allocate_descriptor_set()
32 if (vkAllocateDescriptorSets(screen->dev, &dsai, &desc_set) != VK_SUCCESS) { in allocate_descriptor_set()
37 batch->descs_left -= prog->num_descriptors; in allocate_descriptor_set()
50 * - from VK_EXT_transform_feedback spec in zink_emit_xfb_counter_barrier()
55 for (unsigned i = 0; i < ctx->num_so_targets; i++) { in zink_emit_xfb_counter_barrier()
56 struct zink_so_target *t = zink_so_target(ctx->so_targets[i]); in zink_emit_xfb_counter_barrier()
57 if (t->counter_buffer_valid) { in zink_emit_xfb_counter_barrier()
[all …]
/external/tensorflow/tensorflow/core/kernels/image/
Dimage_ops.h7 http://www.apache.org/licenses/LICENSE-2.0
33 enum Mode { FILL_REFLECT, FILL_WRAP, FILL_CONSTANT, FILL_NEAREST }; enum
40 template <typename Device, Mode M>
46 struct MapCoordinate<Device, Mode::FILL_REFLECT> {
57 in_coord = sz2 * static_cast<DenseIndex>(-in_coord / sz2) + in_coord;
59 in_coord = (in_coord < -len) ? in_coord + sz2 : -in_coord - 1;
61 } else if (in_coord > len - 1) {
66 in_coord -= sz2 * static_cast<DenseIndex>(in_coord / sz2);
68 in_coord = sz2 - in_coord - 1;
74 return Eigen::internal::scalar_clamp_op<float>(0.0f, len - 1)(in_coord);
[all …]
/external/tensorflow/tensorflow/python/distribute/
Dcustom_training_loop_input_test.py7 # http://www.apache.org/licenses/LICENSE-2.0
80 mode=["eager"]
101 mode=["eager"]
123 mode=["eager"]))
125 dataset = get_dataset_from_tensor_slices([5., 6., 7., 8.]).batch(2)
140 distribution=strategy_combinations.all_strategies, mode=["eager"]))
143 dataset = get_dataset_from_tensor_slices(data).batch(2)
160 distribution=strategy_combinations.all_strategies, mode=["eager"]))
165 8, output_type=dtypes.int32).batch(global_batch_size)
185 distribution=strategy_combinations.tpu_strategies, mode=["eager"]))
[all …]
Dinput_lib_test.py7 # http://www.apache.org/licenses/LICENSE-2.0
61 # The passed input_context is to create a sharded dataset in between-graph
73 # MultiWorkerMirroredStrategy. It doesn't apply to in-graph case where
85 # Note: `input_workers.num_workers` is always 1 in between-graph
227 # After re-initializing the iterator, should be able to iterate again.
258 # re-initializing the iterator
286 mode=["eager"],
312 mode=["graph", "eager"],
338 mode=["eager"],
361 mode=["graph", "eager"],
[all …]
Dinput_lib_type_spec_test.py7 # http://www.apache.org/licenses/LICENSE-2.0
49 mode=["eager"],
61 dataset = dataset_ops.DatasetV2.range(10).batch(2)
81 mode=["eager"],
94 dataset = dataset_ops.DatasetV2.range(10).batch(2)
114 mode=["eager"],
138 dataset = dataset_ops.DatasetV2.range(10).batch(
159 mode=["eager"],
180 np.ones([9, 12]).astype(np.float32)).batch(
199 mode=["eager"],
[all …]
/external/mesa3d/src/gallium/drivers/iris/
Diris_draw.c46 /* We don't need to worry about adjacency - it can only be used with in prim_is_points_or_lines()
49 return draw->mode == PIPE_PRIM_POINTS || in prim_is_points_or_lines()
50 draw->mode == PIPE_PRIM_LINES || in prim_is_points_or_lines()
51 draw->mode == PIPE_PRIM_LINE_LOOP || in prim_is_points_or_lines()
52 draw->mode == PIPE_PRIM_LINE_STRIP; in prim_is_points_or_lines()
56 * Record the current primitive mode and restart information, flagging
66 struct iris_screen *screen = (struct iris_screen *)ice->ctx.screen; in iris_update_draw_info()
67 const struct brw_compiler *compiler = screen->compiler; in iris_update_draw_info()
69 if (ice->state.prim_mode != info->mode) { in iris_update_draw_info()
70 ice->state.prim_mode = info->mode; in iris_update_draw_info()
[all …]
/external/tensorflow/tensorflow/python/keras/distribute/
Ddistributed_training_utils_v1.py7 # http://www.apache.org/licenses/LICENSE-2.0
16 # pylint:disable=protected-access
136 def unwrap_output_dict(strategy, grouped_outputs, mode): argument
138 if mode == ModeKeys.PREDICT:
156 # We only do this in eager mode for now since this function is used in
157 # both graph and eager mode and in the graph case we currently don't use
204 # We only do this in eager mode for now since this function is used in
205 # both graph and eager mode and in the graph case we currently don't use
229 # pylint: disable=g-complex-comprehension
377 all_variables = backend._get_variables(backend.get_graph()) # pylint: disable=protected-access
[all …]
/external/tensorflow/tensorflow/python/compiler/tensorrt/test/
Dtrt_mode_test.py7 # http://www.apache.org/licenses/LICENSE-2.0
15 """Model script to test TF-TensorRT integration."""
17 from unittest import SkipTest # pylint: disable=g-importing-member
27 """Test squeeze on batch dim and some unary operations in TF-TRT."""
43 "non-calibration")
50 In explicit batch mode, TensorRT can convert the whole graph. In this mode
51 it is possible to manipulate the batch dimension using the squeeze op.
53 In implicit batch mode TensorRT cannot convert the whole graph. We are not
54 allowed to manipulate (squeeze) the first dimension in implicit batch mode.
64 # The first dimension of the input is squeezed and the batch size for the
[all …]
/external/mesa3d/src/gallium/drivers/freedreno/
Dfreedreno_draw.c46 resource_read(struct fd_batch *batch, struct pipe_resource *prsc) in resource_read() argument
50 fd_batch_resource_read(batch, fd_resource(prsc)); in resource_read()
54 resource_written(struct fd_batch *batch, struct pipe_resource *prsc) in resource_written() argument
58 fd_batch_resource_write(batch, fd_resource(prsc)); in resource_written()
62 batch_draw_tracking(struct fd_batch *batch, const struct pipe_draw_info *info) in batch_draw_tracking() argument
64 struct fd_context *ctx = batch->ctx; in batch_draw_tracking()
65 struct pipe_framebuffer_state *pfb = &batch->framebuffer; in batch_draw_tracking()
68 /* NOTE: needs to be before resource_written(batch->query_buf), otherwise in batch_draw_tracking()
71 fd_batch_set_stage(batch, FD_STAGE_DRAW); in batch_draw_tracking()
77 fd_screen_lock(ctx->screen); in batch_draw_tracking()
[all …]
/external/cronet/net/third_party/quiche/src/quiche/quic/core/
Dquic_packet_writer.h2 // Use of this source code is governed by a BSD-style license that can be
24 // Returns a heap-allocated copy of |this|.
45 // - PassThrough mode. This is the default mode. Caller calls WritePacket with
46 // caller-allocated packet buffer. Unless the writer is blocked, each call to
49 // - Batch mode. In this mode, a call to WritePacket may not cause a packet to
56 // In Batch mode, a writer manages an internal buffer, which is large enough to
58 // caller-allocated packet buffer, the writer will memcpy the buffer into the
67 // PassThrough mode:
68 // Sends the packet out to the peer, with some optional per-packet options.
73 // Batch mode:
[all …]
/external/tensorflow/tensorflow/python/keras/engine/
Dtraining_arrays_v1.py7 # http://www.apache.org/licenses/LICENSE-2.0
16 # pylint: disable=protected-access
37 from scipy.sparse import issparse # pylint: disable=g-import-not-at-top
58 mode=ModeKeys.TRAIN,
70 batch_size: Integer batch size or None if unknown.
72 verbose: 0, 1, or 2. Verbosity mode.
99 mode: One of ModeKeys.TRAIN/ModeKeys.TEST/ModeKeys.PREDICT.
114 - In TRAIN mode: `History` object.
115 - In TEST mode: Evaluation metrics.
116 - In PREDICT mode: Outputs of the Model called on inputs.
[all …]
/external/igt-gpu-tools/tests/i915/
Dgem_streaming_writes.c24 * Chris Wilson <chris@chris-wilson.co.uk>
55 #define BATCH 2 macro
62 static void test_streaming(int fd, int mode, int sync) in test_streaming() argument
75 } *batch; in test_streaming() local
82 switch (mode) { in test_streaming()
137 exec[BATCH].relocation_count = 2; in test_streaming()
143 batch = malloc(sizeof(*batch) * (OBJECT_SIZE / CHUNK_SIZE / 64)); in test_streaming()
147 batch[i].handle = gem_create(fd, 4096); in test_streaming()
148 batch[i].offset = 0; in test_streaming()
150 base = gem_mmap__cpu(fd, batch[i].handle, 0, 4096, PROT_WRITE); in test_streaming()
[all …]
Dgem_mocs_settings.c35 #define GEN9_NUM_MOCS_ENTRIES 62 /* 62 out of 64 - 63 & 64 are reserved. */
36 #define GEN11_NUM_MOCS_ENTRIES 64 /* 63-64 are reserved, but configured. */
147 [3 ... GEN9_NUM_MOCS_ENTRIES - 1] = MOCS_PTE,
151 [0 ... GEN9_NUM_MOCS_ENTRIES - 1] = { 0x00003FFF, 0x003F, 0x1 },
158 [3 ... GEN9_NUM_MOCS_ENTRIES - 1] = MOCS_PTE,
162 [0 ... GEN9_NUM_MOCS_ENTRIES - 1] = { 0x00007FFF, 0x003F, 0x1 },
166 [0 ... GEN9_NUM_MOCS_ENTRIES - 1] = 0xFFFFFFFF,
181 table->size = ARRAY_SIZE(dirty_skylake_mocs_table); in get_mocs_settings()
182 table->table = dirty_skylake_mocs_table; in get_mocs_settings()
184 table->size = ARRAY_SIZE(skylake_mocs_table); in get_mocs_settings()
[all …]
Dgem_exec_parallel.c78 uint32_t batch[16]; in thread() local
81 pthread_mutex_lock(t->mutex); in thread()
82 while (*t->go == 0) in thread()
83 pthread_cond_wait(t->cond, t->mutex); in thread()
84 pthread_mutex_unlock(t->mutex); in thread()
86 if (t->flags & FDS) in thread()
89 fd = t->fd; in thread()
92 batch[i] = MI_STORE_DWORD_IMM | (t->gen < 6 ? 1 << 22 : 0); in thread()
93 if (t->gen >= 8) { in thread()
94 batch[++i] = 4*t->id; in thread()
[all …]
Dgem_exec_whisper.c65 gem_read(fd, handle, reloc->offset, &target, 8); in verify_reloc()
67 gem_read(fd, handle, reloc->offset, &target, 4); in verify_reloc()
69 reloc->presumed_offset + reloc->delta); in verify_reloc()
93 uint32_t *batch; in init_hang() local
96 h->fd = drm_open_driver(DRIVER_INTEL); in init_hang()
97 igt_allow_hang(h->fd, 0, 0); in init_hang()
99 gen = intel_gen(intel_get_drm_devid(h->fd)); in init_hang()
101 memset(&h->execbuf, 0, sizeof(h->execbuf)); in init_hang()
102 h->execbuf.buffers_ptr = to_user_pointer(&h->obj); in init_hang()
103 h->execbuf.buffer_count = 1; in init_hang()
[all …]
/external/mesa3d/src/gallium/drivers/freedreno/a6xx/
Dfd6_vsc.c35 * https://github.com/freedreno/freedreno/wiki/Visibility-Stream-Format
45 * https://github.com/freedreno/freedreno/wiki/Visibility-Stream-Format#numbers
52 return n + (n - 1); in number_size_bits()
57 * https://github.com/freedreno/freedreno/wiki/Visibility-Stream-Format#bitfields
69 unsigned vtx_per_prim = (info->mode == PIPE_PRIM_MAX) ? 2 : in prim_count()
70 u_vertices_per_prim(info->mode); in prim_count()
71 return (info->count * info->instance_count) / vtx_per_prim; in prim_count()
75 * The primitive stream uses a run-length encoding, where each packet contains a
79 * - The (compressed) bitfield of bins covered
80 * - The number of primitives with this bitset
[all …]
/external/angle/src/tests/test_utils/runner/
DREADME.md9 * multi-process execution
11 ## Command-Line Arguments
14 following additional command-line arguments:
16 * `--batch-size` limits the number of tests to run in each batch
17 * `--batch-timeout` limits the amount of time spent in each batch
18 * `--bot-mode` enables multi-process execution and test batching
19 * `--debug-test-groups` dumps the test config categories when using `bot-mode`
20 * `--filter-file` allows passing a larger `gtest_filter` via a file
21 * `--histogram-json-file` outputs a [formatted JSON file][HistogramSet] for perf dashboards
22 * `--max-processes` limits the number of simuntaneous processes
[all …]
/external/tensorflow/tensorflow/python/ops/
Dbatch_norm_benchmark.py7 # http://www.apache.org/licenses/LICENSE-2.0
15 """End-to-end benchmark for batch normalization."""
32 import tensorflow.python.ops.nn_grad # pylint: disable=unused-import
37 """Fused kernel for batch normalization."""
40 # pylint: disable=protected-access
43 # pylint: enable=protected-access
47 # batch_norm = (tensor - mean) * tf.math.rsqrt(variance + 0.001)
52 """Python implementation of batch normalization."""
58 batch_norm = (tensor - mean) * math_ops.rsqrt(variance + 0.001)
64 def build_graph(device, input_shape, axes, num_layers, mode, scale, train): argument
[all …]
/external/igt-gpu-tools/tests/
Dkms_psr.c73 drmModeModeInfo *mode; member
84 fb_id = igt_create_fb(data->drm_fd, 64, 64, in create_cursor_fb()
86 &data->fb_white); in create_cursor_fb()
89 cr = igt_get_cairo_ctx(data->drm_fd, &data->fb_white); in create_cursor_fb()
91 igt_put_cairo_ctx(data->drm_fd, &data->fb_white, cr); in create_cursor_fb()
96 igt_display_t *display = &data->display; in setup_output()
101 drmModeConnectorPtr c = output->config.connector; in setup_output()
103 if (c->connector_type != DRM_MODE_CONNECTOR_eDP) in setup_output()
107 data->crtc_id = output->config.crtc->crtc_id; in setup_output()
108 data->output = output; in setup_output()
[all …]
Dkms_fence_pin_leak.c47 struct intel_batchbuffer *batch; in exec_nop() local
49 dst = gem_handle_to_libdrm_bo(data->bufmgr, data->drm_fd, "", handle); in exec_nop()
52 batch = intel_batchbuffer_alloc(data->bufmgr, data->devid); in exec_nop()
53 igt_assert(batch); in exec_nop()
63 intel_batchbuffer_flush_with_context(batch, context); in exec_nop()
64 intel_batchbuffer_free(batch); in exec_nop()
73 for (i = 0; i < ARRAY_SIZE(data->bos); i++) { in alloc_fence_objs()
76 bo = drm_intel_bo_alloc(data->bufmgr, "fence bo", 4096, 4096); in alloc_fence_objs()
78 gem_set_tiling(data->drm_fd, bo->handle, I915_TILING_X, 512); in alloc_fence_objs()
80 data->bos[i] = bo; in alloc_fence_objs()
[all …]
Dprime_busy.c32 pfd->events = excl ? POLLOUT : POLLIN; in prime_busy()
47 #define BATCH 1 in busy() macro
52 uint32_t *batch, *bbe; in busy() local
67 obj[BATCH].handle = gem_create(fd, size); in busy()
68 obj[BATCH].relocs_ptr = (uintptr_t)store; in busy()
69 obj[BATCH].relocation_count = ARRAY_SIZE(store); in busy()
75 pfd[BATCH].fd = prime_handle_to_fd(fd, obj[BATCH].handle); in busy()
78 batch = gem_mmap__wc(fd, obj[BATCH].handle, 0, size, PROT_WRITE); in busy()
79 gem_set_domain(fd, obj[BATCH].handle, in busy()
85 store[count].presumed_offset = -1; in busy()
[all …]
/external/mesa3d/src/gallium/drivers/panfrost/
Dpan_context.c2 * Copyright (C) 2019-2020 Collabora, Ltd.
4 * Copyright © 2014-2017 Broadcom
34 #include "panfrost-quirks.h"
62 panfrost_emit_midg_tiler(struct panfrost_batch *batch, in panfrost_emit_midg_tiler() argument
66 struct panfrost_device *device = pan_device(batch->ctx->base.screen); in panfrost_emit_midg_tiler()
67 bool hierarchy = !(device->quirks & MIDGARD_NO_HIER_TILING); in panfrost_emit_midg_tiler()
68 unsigned height = batch->key.height; in panfrost_emit_midg_tiler()
69 unsigned width = batch->key.width; in panfrost_emit_midg_tiler()
88 panfrost_batch_get_polygon_list(batch, in panfrost_emit_midg_tiler()
92 t.heap_start = device->tiler_heap->ptr.gpu; in panfrost_emit_midg_tiler()
[all …]
/external/mesa3d/src/mesa/drivers/dri/i965/
Dbrw_misc_state.c49 * Upload pointers to the per-stage state.
57 const struct gen_device_info *devinfo = &brw->screen->devinfo; in upload_pipelined_state_pointers()
59 if (devinfo->gen == 5) { in upload_pipelined_state_pointers()
67 OUT_BATCH(_3DSTATE_PIPELINED_POINTERS << 16 | (7 - 2)); in upload_pipelined_state_pointers()
68 OUT_RELOC(brw->batch.state.bo, 0, brw->vs.base.state_offset); in upload_pipelined_state_pointers()
69 if (brw->ff_gs.prog_active) in upload_pipelined_state_pointers()
70 OUT_RELOC(brw->batch.state.bo, 0, brw->ff_gs.state_offset | 1); in upload_pipelined_state_pointers()
73 OUT_RELOC(brw->batch.state.bo, 0, brw->clip.state_offset | 1); in upload_pipelined_state_pointers()
74 OUT_RELOC(brw->batch.state.bo, 0, brw->sf.state_offset); in upload_pipelined_state_pointers()
75 OUT_RELOC(brw->batch.state.bo, 0, brw->wm.base.state_offset); in upload_pipelined_state_pointers()
[all …]
/external/gson/.github/workflows/
Dcheck-api-compatibility.yml6 check-api-compatibility:
7 runs-on: ubuntu-latest
10 - name: Checkout old version
14 path: 'gson-old-japicmp'
16 - name: Set up JDK 11
17 uses: actions/setup-java@v3
20 java-version: '11'
23 - name: Build old version
25 cd gson-old-japicmp
27 …mvn --batch-mode --no-transfer-progress org.codehaus.mojo:versions-maven-plugin:2.11.0:set -DnewVe…
[all …]

12345678910>>...41