HomeSort by relevance Sort by last modified time
    Searched refs:batch (Results 1 - 25 of 709) sorted by null

1 2 3 4 5 6 7 8 91011>>

  /external/mesa3d/src/gallium/drivers/freedreno/
freedreno_batch.c 38 batch_init(struct fd_batch *batch)
40 struct fd_context *ctx = batch->ctx;
44 util_queue_fence_init(&batch->flush_fence);
56 batch->draw = fd_ringbuffer_new(ctx->screen->pipe, size);
57 batch->binning = fd_ringbuffer_new(ctx->screen->pipe, size);
58 batch->gmem = fd_ringbuffer_new(ctx->screen->pipe, size);
60 fd_ringbuffer_set_parent(batch->gmem, NULL);
61 fd_ringbuffer_set_parent(batch->draw, batch->gmem);
62 fd_ringbuffer_set_parent(batch->binning, batch->gmem)
92 struct fd_batch *batch = CALLOC_STRUCT(fd_batch); local
239 struct fd_batch *batch = job; local
248 struct fd_batch *batch = job; local
    [all...]
freedreno_batch.h 62 /* A batch tracks everything about a cmdstream batch/submit, including the
119 unsigned num_draws; /* number of draws in current batch */
122 * batch. Used at the tile rendering step (fd_gmem_render_tiles(),
161 /* next sample offset.. incremented for each sample in the batch/
171 /* which sample providers were active in the current batch: */
179 /* list of samples in current batch: */
188 /* Set of resources used by currently-unsubmitted batch (read or
193 /** key in batch-cache (if not null): */
203 void fd_batch_reset(struct fd_batch *batch);
    [all...]
freedreno_gmem.c 107 calculate_tiles(struct fd_batch *batch)
109 struct fd_context *ctx = batch->ctx;
111 struct pipe_scissor_state *scissor = &batch->max_scissor;
112 struct pipe_framebuffer_state *pfb = &batch->framebuffer;
123 bool has_zs = !!(batch->resolve & (FD_BUFFER_DEPTH | FD_BUFFER_STENCIL));
310 render_tiles(struct fd_batch *batch)
312 struct fd_context *ctx = batch->ctx;
316 ctx->emit_tile_init(batch);
318 if (batch->restore)
327 ctx->emit_tile_prep(batch, tile)
    [all...]
freedreno_query_hw.c 64 get_sample(struct fd_batch *batch, struct fd_ringbuffer *ring,
67 struct fd_context *ctx = batch->ctx;
73 if (!batch->sample_cache[idx]) {
75 ctx->sample_providers[idx]->get_sample(batch, ring);
76 fd_hw_sample_reference(ctx, &batch->sample_cache[idx], new_samp);
77 util_dynarray_append(&batch->samples, struct fd_hw_sample *, new_samp);
78 batch->needs_flush = true;
81 fd_hw_sample_reference(ctx, &samp, batch->sample_cache[idx]);
87 clear_sample_cache(struct fd_batch *batch)
91 for (i = 0; i < ARRAY_SIZE(batch->sample_cache); i++
160 struct fd_batch *batch = ctx->batch; local
186 struct fd_batch *batch = ctx->batch; local
    [all...]
freedreno_draw.c 43 resource_read(struct fd_batch *batch, struct pipe_resource *prsc)
47 fd_batch_resource_used(batch, fd_resource(prsc), false);
51 resource_written(struct fd_batch *batch, struct pipe_resource *prsc)
55 fd_batch_resource_used(batch, fd_resource(prsc), true);
62 struct fd_batch *batch = ctx->batch; local
63 struct pipe_framebuffer_state *pfb = &batch->framebuffer;
88 fd_batch_reset(batch);
92 batch->blit = ctx->in_blit;
93 batch->back_blit = ctx->in_shadow
284 struct fd_batch *batch = ctx->batch; local
    [all...]
freedreno_batch_cache.c 39 * The batch cache provides lookup for mapping pipe_framebuffer_state
40 * to a batch.
45 * Batch Cache hashtable key:
52 * Batch:
54 * Each batch needs to hold a reference to each resource it depends on (ie.
59 * When a resource is destroyed, we need to remove entries in the batch
64 * When a batch has weak reference to no more resources (ie. all the
65 * surfaces it rendered to are destroyed) the batch can be destroyed.
68 * surfaces are destroyed before the batch is submitted.
70 * If (for example), batch writes to zsbuf but that surface is destroye
136 struct fd_batch *batch = NULL; local
159 struct fd_batch *batch; local
208 struct fd_batch *batch; local
233 struct fd_batch *batch; local
315 struct fd_batch *batch = NULL; local
    [all...]
freedreno_query_hw.h 80 * at least once before the first ->get_sample() in a batch.
87 struct fd_hw_sample *(*get_sample)(struct fd_batch *batch,
106 * the sample is constructed, but when the batch is submitted.
130 struct list_head list; /* list-node in batch->active_queries */
143 struct fd_hw_sample * fd_hw_sample_init(struct fd_batch *batch, uint32_t size);
146 void fd_hw_query_prepare(struct fd_batch *batch, uint32_t num_tiles);
147 void fd_hw_query_prepare_tile(struct fd_batch *batch, uint32_t n,
149 void fd_hw_query_set_stage(struct fd_batch *batch,
151 void fd_hw_query_enable(struct fd_batch *batch, struct fd_ringbuffer *ring);
  /external/mesa3d/src/gallium/drivers/i915/
i915_batchbuffer.h 37 i915_winsys_batchbuffer_space(struct i915_winsys_batchbuffer *batch)
39 return batch->size - (batch->ptr - batch->map);
43 i915_winsys_batchbuffer_check(struct i915_winsys_batchbuffer *batch,
46 return dwords * 4 <= i915_winsys_batchbuffer_space(batch);
50 i915_winsys_batchbuffer_dword_unchecked(struct i915_winsys_batchbuffer *batch,
53 *(unsigned *)batch->ptr = dword;
54 batch->ptr += 4;
58 i915_winsys_batchbuffer_float(struct i915_winsys_batchbuffer *batch,
    [all...]
i915_batch.h 36 (i915_winsys_batchbuffer_check(i915->batch, dwords))
39 i915_winsys_batchbuffer_dword(i915->batch, dword)
42 i915_winsys_batchbuffer_float(i915->batch, f)
45 i915_winsys_batchbuffer_reloc(i915->batch, buf, usage, offset, false)
48 i915_winsys_batchbuffer_reloc(i915->batch, buf, usage, offset, true)
i915_flush.c 48 if (!i915->batch)
54 if (!fence && (i915->batch->map == i915->batch->ptr)) {
79 struct i915_winsys_batchbuffer *batch = i915->batch; local
81 batch->iws->batchbuffer_flush(batch, fence, flags);
  /external/tensorflow/tensorflow/core/kernels/batching_util/
batch_scheduler_test.cc 41 Batch<FakeTask> batch; local
43 EXPECT_EQ(0, batch.num_tasks());
44 EXPECT_TRUE(batch.empty());
45 EXPECT_EQ(0, batch.size());
46 EXPECT_FALSE(batch.IsClosed());
49 batch.AddTask(std::unique_ptr<FakeTask>(task0));
51 EXPECT_EQ(1, batch.num_tasks());
52 EXPECT_FALSE(batch.empty());
53 EXPECT_EQ(task0->size(), batch.size())
84 Batch<FakeTask> batch; local
98 Batch<FakeTask>* batch = new Batch<FakeTask>; local
    [all...]
  /external/mesa3d/src/mesa/drivers/dri/i965/
intel_batchbuffer.h 14 * Number of bytes to reserve for commands necessary to complete a batch.
18 * - Optional MI_NOOP for ensuring the batch length is qword aligned (4 bytes)
42 void intel_batchbuffer_init(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr,
44 void intel_batchbuffer_free(struct intel_batchbuffer *batch);
66 uint32_t intel_batchbuffer_reloc(struct intel_batchbuffer *batch,
72 uint64_t intel_batchbuffer_reloc64(struct intel_batchbuffer *batch,
79 #define USED_BATCH(batch) ((uintptr_t)((batch).map_next - (batch).map))
98 intel_batchbuffer_space(struct intel_batchbuffer *batch)
136 struct intel_batchbuffer *batch = &brw->batch; local
    [all...]
genX_blorp_exec.c 38 blorp_emit_dwords(struct blorp_batch *batch, unsigned n)
40 assert(batch->blorp->driver_ctx == batch->driver_batch);
41 struct brw_context *brw = batch->driver_batch;
44 uint32_t *map = brw->batch.map_next;
45 brw->batch.map_next += n;
51 blorp_emit_reloc(struct blorp_batch *batch,
54 assert(batch->blorp->driver_ctx == batch->driver_batch);
55 struct brw_context *brw = batch->driver_batch
    [all...]
intel_batchbuffer.c 39 intel_batchbuffer_reset(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr,
43 intel_batchbuffer_init(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr,
46 intel_batchbuffer_reset(batch, bufmgr, has_llc);
49 batch->cpu_map = malloc(BATCH_SZ);
50 batch->map = batch->cpu_map;
51 batch->map_next = batch->cpu_map;
56 intel_batchbuffer_reset(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr,
59 if (batch->last_bo != NULL)
146 struct intel_batchbuffer *batch = &brw->batch; local
324 struct intel_batchbuffer *batch = &brw->batch; local
    [all...]
brw_state_batch.c 44 struct intel_batchbuffer *batch = &brw->batch; local
52 batch->bo->size / 32);
76 * Generate a set of aub file annotations for the current batch buffer, and
79 * The "used" section of the batch buffer (the portion containing batch
81 * batch buffer (which contains data structures pointed to by batch commands)
91 4 * USED_BATCH(brw->batch));
102 drm_intel_bufmgr_gem_set_aub_annotations(brw->batch.bo, annotations
129 struct intel_batchbuffer *batch = &brw->batch; local
    [all...]
brw_blorp.h 74 void gen6_blorp_exec(struct blorp_batch *batch,
76 void gen7_blorp_exec(struct blorp_batch *batch,
78 void gen75_blorp_exec(struct blorp_batch *batch,
80 void gen8_blorp_exec(struct blorp_batch *batch,
82 void gen9_blorp_exec(struct blorp_batch *batch,
  /external/mesa3d/src/gallium/winsys/i915/drm/
i915_drm_batchbuffer.c 30 i915_drm_batchbuffer(struct i915_winsys_batchbuffer *batch)
32 return (struct i915_drm_batchbuffer *)batch;
36 i915_drm_batchbuffer_reset(struct i915_drm_batchbuffer *batch)
38 struct i915_drm_winsys *idws = i915_drm_winsys(batch->base.iws);
40 if (batch->bo)
41 drm_intel_bo_unreference(batch->bo);
42 batch->bo = drm_intel_bo_alloc(idws->gem_manager,
44 batch->actual_size,
47 memset(batch->base.map, 0, batch->actual_size)
57 struct i915_drm_batchbuffer *batch = CALLOC_STRUCT(i915_drm_batchbuffer); local
100 struct i915_drm_batchbuffer *batch = i915_drm_batchbuffer(ibatch); local
165 struct i915_drm_batchbuffer *batch = i915_drm_batchbuffer(ibatch); local
221 struct i915_drm_batchbuffer *batch = i915_drm_batchbuffer(ibatch); local
    [all...]
  /external/mesa3d/src/mesa/drivers/dri/i915/
intel_batchbuffer.c 43 intel->batch.cpu_map = malloc(intel->maxBatchSize);
44 intel->batch.map = intel->batch.cpu_map;
50 if (intel->batch.last_bo != NULL) {
51 drm_intel_bo_unreference(intel->batch.last_bo);
52 intel->batch.last_bo = NULL;
54 intel->batch.last_bo = intel->batch.bo;
56 intel->batch.bo = drm_intel_bo_alloc(intel->bufmgr, "batchbuffer",
59 intel->batch.reserved_space = BATCH_RESERVED
75 struct intel_batchbuffer *batch = &intel->batch; local
116 struct intel_batchbuffer *batch = &intel->batch; local
    [all...]
intel_batchbuffer.h 15 * Number of bytes to reserve for commands necessary to complete a batch.
19 * - Optional MI_NOOP for ensuring the batch length is qword aligned (4 bytes)
76 return (intel->batch.bo->size - intel->batch.reserved_space)
77 - intel->batch.used*4;
87 intel->batch.map[intel->batch.used++] = dword;
112 intel->batch.emit = intel->batch.used;
114 intel->batch.total = n
122 struct intel_batchbuffer *batch = &intel->batch; local
    [all...]
  /external/mesa3d/src/intel/vulkan/
genX_blorp_exec.c 38 blorp_emit_dwords(struct blorp_batch *batch, unsigned n)
40 struct anv_cmd_buffer *cmd_buffer = batch->driver_batch;
41 return anv_batch_emit_dwords(&cmd_buffer->batch, n);
45 blorp_emit_reloc(struct blorp_batch *batch,
48 struct anv_cmd_buffer *cmd_buffer = batch->driver_batch;
49 assert(cmd_buffer->batch.start <= location &&
50 location < cmd_buffer->batch.end);
51 return anv_batch_emit_reloc(&cmd_buffer->batch, location,
56 blorp_surface_reloc(struct blorp_batch *batch, uint32_t ss_offset,
59 struct anv_cmd_buffer *cmd_buffer = batch->driver_batch
    [all...]
  /external/jsoncpp/src/lib_json/
json_batchallocator.h 18 * This memory allocator allocates memory for a batch of object (specified by
44 for (BatchInfo* batch = batches_; batch;) {
45 BatchInfo* nextBatch = batch->next_;
46 free(batch);
47 batch = nextBatch;
66 if (!currentBatch_) // no free batch found, allocate a new one
103 BatchInfo* batch = static_cast<BatchInfo*>(malloc(mallocSize)); local
104 batch->next_ = 0;
105 batch->used_ = batch->buffer_
    [all...]
  /external/mesa3d/src/intel/blorp/
blorp_genX_exec.h 48 blorp_emit_dwords(struct blorp_batch *batch, unsigned n);
51 blorp_emit_reloc(struct blorp_batch *batch,
55 blorp_alloc_dynamic_state(struct blorp_batch *batch,
61 blorp_alloc_vertex_buffer(struct blorp_batch *batch, uint32_t size,
65 blorp_alloc_binding_table(struct blorp_batch *batch, unsigned num_entries,
71 blorp_flush_range(struct blorp_batch *batch, void *start, size_t size);
74 blorp_surface_reloc(struct blorp_batch *batch, uint32_t ss_offset,
78 blorp_emit_urb_config(struct blorp_batch *batch, unsigned vs_entry_size);
85 _blorp_combine_address(struct blorp_batch *batch, void *location,
91 return blorp_emit_reloc(batch, location, address, delta)
    [all...]
  /external/tensorflow/tensorflow/core/kernels/
eigen_activations_test.cc 30 const ptrdiff_t batch = 10; local
34 Tensor<float, 4> input(depth, rows, cols, batch);
37 Tensor<float, 4> result(depth, rows, cols, batch);
40 for (int b = 0; b < batch; ++b) {
54 const ptrdiff_t batch = 10; local
58 Tensor<float, 4> input(depth, rows, cols, batch);
61 Tensor<float, 4> result(depth, rows, cols, batch);
64 for (int b = 0; b < batch; ++b) {
78 const ptrdiff_t batch = 10; local
82 Tensor<float, 4> input(depth, rows, cols, batch);
    [all...]
  /external/skia/tests/
PathOpsThreadedCommon.cpp 19 SkTaskGroup().batch(fRunnables.count(), [&](int i) {
  /external/skqp/tests/
PathOpsThreadedCommon.cpp 19 SkTaskGroup().batch(fRunnables.count(), [&](int i) {

Completed in 253 milliseconds

1 2 3 4 5 6 7 8 91011>>