/external/mesa3d/src/gallium/drivers/freedreno/ |
freedreno_batch.c | 38 batch_init(struct fd_batch *batch) 40 struct fd_context *ctx = batch->ctx; 44 util_queue_fence_init(&batch->flush_fence); 56 batch->draw = fd_ringbuffer_new(ctx->screen->pipe, size); 57 batch->binning = fd_ringbuffer_new(ctx->screen->pipe, size); 58 batch->gmem = fd_ringbuffer_new(ctx->screen->pipe, size); 60 fd_ringbuffer_set_parent(batch->gmem, NULL); 61 fd_ringbuffer_set_parent(batch->draw, batch->gmem); 62 fd_ringbuffer_set_parent(batch->binning, batch->gmem) 92 struct fd_batch *batch = CALLOC_STRUCT(fd_batch); local 239 struct fd_batch *batch = job; local 248 struct fd_batch *batch = job; local [all...] |
freedreno_batch.h | 62 /* A batch tracks everything about a cmdstream batch/submit, including the 119 unsigned num_draws; /* number of draws in current batch */ 122 * batch. Used at the tile rendering step (fd_gmem_render_tiles(), 161 /* next sample offset.. incremented for each sample in the batch/ 171 /* which sample providers were active in the current batch: */ 179 /* list of samples in current batch: */ 188 /* Set of resources used by currently-unsubmitted batch (read or 193 /** key in batch-cache (if not null): */ 203 void fd_batch_reset(struct fd_batch *batch); [all...] |
freedreno_gmem.c | 107 calculate_tiles(struct fd_batch *batch) 109 struct fd_context *ctx = batch->ctx; 111 struct pipe_scissor_state *scissor = &batch->max_scissor; 112 struct pipe_framebuffer_state *pfb = &batch->framebuffer; 123 bool has_zs = !!(batch->resolve & (FD_BUFFER_DEPTH | FD_BUFFER_STENCIL)); 310 render_tiles(struct fd_batch *batch) 312 struct fd_context *ctx = batch->ctx; 316 ctx->emit_tile_init(batch); 318 if (batch->restore) 327 ctx->emit_tile_prep(batch, tile) [all...] |
freedreno_query_hw.c | 64 get_sample(struct fd_batch *batch, struct fd_ringbuffer *ring, 67 struct fd_context *ctx = batch->ctx; 73 if (!batch->sample_cache[idx]) { 75 ctx->sample_providers[idx]->get_sample(batch, ring); 76 fd_hw_sample_reference(ctx, &batch->sample_cache[idx], new_samp); 77 util_dynarray_append(&batch->samples, struct fd_hw_sample *, new_samp); 78 batch->needs_flush = true; 81 fd_hw_sample_reference(ctx, &samp, batch->sample_cache[idx]); 87 clear_sample_cache(struct fd_batch *batch) 91 for (i = 0; i < ARRAY_SIZE(batch->sample_cache); i++ 160 struct fd_batch *batch = ctx->batch; local 186 struct fd_batch *batch = ctx->batch; local [all...] |
freedreno_draw.c | 43 resource_read(struct fd_batch *batch, struct pipe_resource *prsc) 47 fd_batch_resource_used(batch, fd_resource(prsc), false); 51 resource_written(struct fd_batch *batch, struct pipe_resource *prsc) 55 fd_batch_resource_used(batch, fd_resource(prsc), true); 62 struct fd_batch *batch = ctx->batch; local 63 struct pipe_framebuffer_state *pfb = &batch->framebuffer; 88 fd_batch_reset(batch); 92 batch->blit = ctx->in_blit; 93 batch->back_blit = ctx->in_shadow 284 struct fd_batch *batch = ctx->batch; local [all...] |
freedreno_batch_cache.c | 39 * The batch cache provides lookup for mapping pipe_framebuffer_state 40 * to a batch. 45 * Batch Cache hashtable key: 52 * Batch: 54 * Each batch needs to hold a reference to each resource it depends on (ie. 59 * When a resource is destroyed, we need to remove entries in the batch 64 * When a batch has weak reference to no more resources (ie. all the 65 * surfaces it rendered to are destroyed) the batch can be destroyed. 68 * surfaces are destroyed before the batch is submitted. 70 * If (for example), batch writes to zsbuf but that surface is destroye 136 struct fd_batch *batch = NULL; local 159 struct fd_batch *batch; local 208 struct fd_batch *batch; local 233 struct fd_batch *batch; local 315 struct fd_batch *batch = NULL; local [all...] |
freedreno_query_hw.h | 80 * at least once before the first ->get_sample() in a batch. 87 struct fd_hw_sample *(*get_sample)(struct fd_batch *batch, 106 * the sample is constructed, but when the batch is submitted. 130 struct list_head list; /* list-node in batch->active_queries */ 143 struct fd_hw_sample * fd_hw_sample_init(struct fd_batch *batch, uint32_t size); 146 void fd_hw_query_prepare(struct fd_batch *batch, uint32_t num_tiles); 147 void fd_hw_query_prepare_tile(struct fd_batch *batch, uint32_t n, 149 void fd_hw_query_set_stage(struct fd_batch *batch, 151 void fd_hw_query_enable(struct fd_batch *batch, struct fd_ringbuffer *ring);
|
/external/mesa3d/src/gallium/drivers/i915/ |
i915_batchbuffer.h | 37 i915_winsys_batchbuffer_space(struct i915_winsys_batchbuffer *batch) 39 return batch->size - (batch->ptr - batch->map); 43 i915_winsys_batchbuffer_check(struct i915_winsys_batchbuffer *batch, 46 return dwords * 4 <= i915_winsys_batchbuffer_space(batch); 50 i915_winsys_batchbuffer_dword_unchecked(struct i915_winsys_batchbuffer *batch, 53 *(unsigned *)batch->ptr = dword; 54 batch->ptr += 4; 58 i915_winsys_batchbuffer_float(struct i915_winsys_batchbuffer *batch, [all...] |
i915_batch.h | 36 (i915_winsys_batchbuffer_check(i915->batch, dwords)) 39 i915_winsys_batchbuffer_dword(i915->batch, dword) 42 i915_winsys_batchbuffer_float(i915->batch, f) 45 i915_winsys_batchbuffer_reloc(i915->batch, buf, usage, offset, false) 48 i915_winsys_batchbuffer_reloc(i915->batch, buf, usage, offset, true)
|
i915_flush.c | 48 if (!i915->batch) 54 if (!fence && (i915->batch->map == i915->batch->ptr)) { 79 struct i915_winsys_batchbuffer *batch = i915->batch; local 81 batch->iws->batchbuffer_flush(batch, fence, flags);
|
/external/tensorflow/tensorflow/core/kernels/batching_util/ |
batch_scheduler_test.cc | 41 Batch<FakeTask> batch; local 43 EXPECT_EQ(0, batch.num_tasks()); 44 EXPECT_TRUE(batch.empty()); 45 EXPECT_EQ(0, batch.size()); 46 EXPECT_FALSE(batch.IsClosed()); 49 batch.AddTask(std::unique_ptr<FakeTask>(task0)); 51 EXPECT_EQ(1, batch.num_tasks()); 52 EXPECT_FALSE(batch.empty()); 53 EXPECT_EQ(task0->size(), batch.size()) 84 Batch<FakeTask> batch; local 98 Batch<FakeTask>* batch = new Batch<FakeTask>; local [all...] |
/external/mesa3d/src/mesa/drivers/dri/i965/ |
intel_batchbuffer.h | 14 * Number of bytes to reserve for commands necessary to complete a batch. 18 * - Optional MI_NOOP for ensuring the batch length is qword aligned (4 bytes) 42 void intel_batchbuffer_init(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr, 44 void intel_batchbuffer_free(struct intel_batchbuffer *batch); 66 uint32_t intel_batchbuffer_reloc(struct intel_batchbuffer *batch, 72 uint64_t intel_batchbuffer_reloc64(struct intel_batchbuffer *batch, 79 #define USED_BATCH(batch) ((uintptr_t)((batch).map_next - (batch).map)) 98 intel_batchbuffer_space(struct intel_batchbuffer *batch) 136 struct intel_batchbuffer *batch = &brw->batch; local [all...] |
genX_blorp_exec.c | 38 blorp_emit_dwords(struct blorp_batch *batch, unsigned n) 40 assert(batch->blorp->driver_ctx == batch->driver_batch); 41 struct brw_context *brw = batch->driver_batch; 44 uint32_t *map = brw->batch.map_next; 45 brw->batch.map_next += n; 51 blorp_emit_reloc(struct blorp_batch *batch, 54 assert(batch->blorp->driver_ctx == batch->driver_batch); 55 struct brw_context *brw = batch->driver_batch [all...] |
intel_batchbuffer.c | 39 intel_batchbuffer_reset(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr, 43 intel_batchbuffer_init(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr, 46 intel_batchbuffer_reset(batch, bufmgr, has_llc); 49 batch->cpu_map = malloc(BATCH_SZ); 50 batch->map = batch->cpu_map; 51 batch->map_next = batch->cpu_map; 56 intel_batchbuffer_reset(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr, 59 if (batch->last_bo != NULL) 146 struct intel_batchbuffer *batch = &brw->batch; local 324 struct intel_batchbuffer *batch = &brw->batch; local [all...] |
brw_state_batch.c | 44 struct intel_batchbuffer *batch = &brw->batch; local 52 batch->bo->size / 32); 76 * Generate a set of aub file annotations for the current batch buffer, and 79 * The "used" section of the batch buffer (the portion containing batch 81 * batch buffer (which contains data structures pointed to by batch commands) 91 4 * USED_BATCH(brw->batch)); 102 drm_intel_bufmgr_gem_set_aub_annotations(brw->batch.bo, annotations 129 struct intel_batchbuffer *batch = &brw->batch; local [all...] |
brw_blorp.h | 74 void gen6_blorp_exec(struct blorp_batch *batch, 76 void gen7_blorp_exec(struct blorp_batch *batch, 78 void gen75_blorp_exec(struct blorp_batch *batch, 80 void gen8_blorp_exec(struct blorp_batch *batch, 82 void gen9_blorp_exec(struct blorp_batch *batch,
|
/external/mesa3d/src/gallium/winsys/i915/drm/ |
i915_drm_batchbuffer.c | 30 i915_drm_batchbuffer(struct i915_winsys_batchbuffer *batch) 32 return (struct i915_drm_batchbuffer *)batch; 36 i915_drm_batchbuffer_reset(struct i915_drm_batchbuffer *batch) 38 struct i915_drm_winsys *idws = i915_drm_winsys(batch->base.iws); 40 if (batch->bo) 41 drm_intel_bo_unreference(batch->bo); 42 batch->bo = drm_intel_bo_alloc(idws->gem_manager, 44 batch->actual_size, 47 memset(batch->base.map, 0, batch->actual_size) 57 struct i915_drm_batchbuffer *batch = CALLOC_STRUCT(i915_drm_batchbuffer); local 100 struct i915_drm_batchbuffer *batch = i915_drm_batchbuffer(ibatch); local 165 struct i915_drm_batchbuffer *batch = i915_drm_batchbuffer(ibatch); local 221 struct i915_drm_batchbuffer *batch = i915_drm_batchbuffer(ibatch); local [all...] |
/external/mesa3d/src/mesa/drivers/dri/i915/ |
intel_batchbuffer.c | 43 intel->batch.cpu_map = malloc(intel->maxBatchSize); 44 intel->batch.map = intel->batch.cpu_map; 50 if (intel->batch.last_bo != NULL) { 51 drm_intel_bo_unreference(intel->batch.last_bo); 52 intel->batch.last_bo = NULL; 54 intel->batch.last_bo = intel->batch.bo; 56 intel->batch.bo = drm_intel_bo_alloc(intel->bufmgr, "batchbuffer", 59 intel->batch.reserved_space = BATCH_RESERVED 75 struct intel_batchbuffer *batch = &intel->batch; local 116 struct intel_batchbuffer *batch = &intel->batch; local [all...] |
intel_batchbuffer.h | 15 * Number of bytes to reserve for commands necessary to complete a batch. 19 * - Optional MI_NOOP for ensuring the batch length is qword aligned (4 bytes) 76 return (intel->batch.bo->size - intel->batch.reserved_space) 77 - intel->batch.used*4; 87 intel->batch.map[intel->batch.used++] = dword; 112 intel->batch.emit = intel->batch.used; 114 intel->batch.total = n 122 struct intel_batchbuffer *batch = &intel->batch; local [all...] |
/external/mesa3d/src/intel/vulkan/ |
genX_blorp_exec.c | 38 blorp_emit_dwords(struct blorp_batch *batch, unsigned n) 40 struct anv_cmd_buffer *cmd_buffer = batch->driver_batch; 41 return anv_batch_emit_dwords(&cmd_buffer->batch, n); 45 blorp_emit_reloc(struct blorp_batch *batch, 48 struct anv_cmd_buffer *cmd_buffer = batch->driver_batch; 49 assert(cmd_buffer->batch.start <= location && 50 location < cmd_buffer->batch.end); 51 return anv_batch_emit_reloc(&cmd_buffer->batch, location, 56 blorp_surface_reloc(struct blorp_batch *batch, uint32_t ss_offset, 59 struct anv_cmd_buffer *cmd_buffer = batch->driver_batch [all...] |
/external/jsoncpp/src/lib_json/ |
json_batchallocator.h | 18 * This memory allocator allocates memory for a batch of object (specified by 44 for (BatchInfo* batch = batches_; batch;) { 45 BatchInfo* nextBatch = batch->next_; 46 free(batch); 47 batch = nextBatch; 66 if (!currentBatch_) // no free batch found, allocate a new one 103 BatchInfo* batch = static_cast<BatchInfo*>(malloc(mallocSize)); local 104 batch->next_ = 0; 105 batch->used_ = batch->buffer_ [all...] |
/external/mesa3d/src/intel/blorp/ |
blorp_genX_exec.h | 48 blorp_emit_dwords(struct blorp_batch *batch, unsigned n); 51 blorp_emit_reloc(struct blorp_batch *batch, 55 blorp_alloc_dynamic_state(struct blorp_batch *batch, 61 blorp_alloc_vertex_buffer(struct blorp_batch *batch, uint32_t size, 65 blorp_alloc_binding_table(struct blorp_batch *batch, unsigned num_entries, 71 blorp_flush_range(struct blorp_batch *batch, void *start, size_t size); 74 blorp_surface_reloc(struct blorp_batch *batch, uint32_t ss_offset, 78 blorp_emit_urb_config(struct blorp_batch *batch, unsigned vs_entry_size); 85 _blorp_combine_address(struct blorp_batch *batch, void *location, 91 return blorp_emit_reloc(batch, location, address, delta) [all...] |
/external/tensorflow/tensorflow/core/kernels/ |
eigen_activations_test.cc | 30 const ptrdiff_t batch = 10; local 34 Tensor<float, 4> input(depth, rows, cols, batch); 37 Tensor<float, 4> result(depth, rows, cols, batch); 40 for (int b = 0; b < batch; ++b) { 54 const ptrdiff_t batch = 10; local 58 Tensor<float, 4> input(depth, rows, cols, batch); 61 Tensor<float, 4> result(depth, rows, cols, batch); 64 for (int b = 0; b < batch; ++b) { 78 const ptrdiff_t batch = 10; local 82 Tensor<float, 4> input(depth, rows, cols, batch); [all...] |
/external/skia/tests/ |
PathOpsThreadedCommon.cpp | 19 SkTaskGroup().batch(fRunnables.count(), [&](int i) {
|
/external/skqp/tests/ |
PathOpsThreadedCommon.cpp | 19 SkTaskGroup().batch(fRunnables.count(), [&](int i) {
|