HomeSort by relevance Sort by last modified time
    Searched defs:bo (Results 101 - 125 of 245) sorted by null

1 2 3 45 6 7 8 910

  /external/libdrm/etnaviv/
etnaviv_drmif.h 70 /* bo flags: */
78 /* bo access flags: (keep aligned to ETNA_PREP_x) */
119 struct etna_bo *etna_bo_ref(struct etna_bo *bo);
120 void etna_bo_del(struct etna_bo *bo);
121 int etna_bo_get_name(struct etna_bo *bo, uint32_t *name);
122 uint32_t etna_bo_handle(struct etna_bo *bo);
123 int etna_bo_dmabuf(struct etna_bo *bo);
124 uint32_t etna_bo_size(struct etna_bo *bo);
125 void * etna_bo_map(struct etna_bo *bo);
126 int etna_bo_cpu_prep(struct etna_bo *bo, uint32_t op)
182 struct etna_bo *bo; member in struct:etna_reloc
    [all...]
  /external/libdrm/omap/
omap_drm.c 60 /* The handle_table is used to track GEM bo handles associated w/
66 * to an invalid handle (and possible a GEM bo that is already
167 struct omap_bo *bo = NULL; local
168 if (!drmHashLookup(dev->handle_table, handle, (void **)&bo)) {
170 bo = omap_bo_ref(bo);
172 return bo;
179 struct omap_bo *bo = calloc(sizeof(*bo), 1); local
180 if (!bo) {
200 struct omap_bo *bo = NULL; local
291 struct omap_bo *bo = NULL; local
325 struct omap_bo *bo = NULL; local
    [all...]
  /external/libdrm/radeon/
radeon_cs_gem.c 176 struct radeon_bo *bo,
181 struct radeon_bo_int *boi = (struct radeon_bo_int *)bo;
191 /* in one CS a bo can only be in read or write domain but not
203 if this bo is for sure not in this cs.*/
204 if ((atomic_read((atomic_t *)radeon_gem_get_reloc_in_cs(bo)) & cs->id)) {
205 /* check if bo is already referenced.
207 * it often relocates same shared dma bo again. */
212 if (reloc->handle == bo->handle) {
265 reloc->handle = bo->handle;
270 radeon_bo_ref(bo);
344 bof_t *bcs, *blob, *array, *bo, *size, *handle, *device_id, *root; local
    [all...]
radeon_cs_int.h 6 struct radeon_bo_int *bo; member in struct:radeon_cs_space_check
39 struct radeon_bo *bo,
  /external/libdrm/tests/exynos/
exynos_fimg2d_event.c 97 * potentially remove a BO which the engine still operates on.
100 * Also any subsequent BO allocations fail then with:
210 struct exynos_bo *bo; local
285 bo = exynos_bo_create(dev, bufw * bufh * 4, 0);
286 if (bo == NULL) {
287 fprintf(stderr, "error: failed to create bo\n");
299 img.bo[0] = bo->handle;
313 exynos_bo_destroy(bo);
  /external/libxcam/xcore/
drm_display.cpp 390 const drm_intel_bo* bo = bo_buf->get_bo (); local
391 XCAM_ASSERT (bo);
392 bo_handle = bo->handle;
511 drm_intel_bo *bo = NULL; local
526 XCAM_LOG_DEBUG ("DrmDisplay only support dma buffer conversion to drm bo by now");
530 bo = drm_intel_bo_gem_create_from_prime (_buf_manager, dma_fd, video_info.size);
531 if (bo == NULL) {
532 XCAM_LOG_WARNING ("convert dma fd to drm bo failed");
535 bo_data = new DrmBoData (self, bo);
550 drm_intel_bo *bo = drm_intel_bo_alloc local
560 drm_intel_bo *bo = NULL; local
    [all...]
  /external/mesa3d/src/egl/drivers/dri2/
platform_drm.c 49 struct gbm_bo *bo; local
56 bo = dri2_surf->current->bo;
63 return bo;
67 release_buffer(struct gbm_surface *_surf, struct gbm_bo *bo)
74 if (dri2_surf->color_buffers[i].bo == bo) {
197 if (dri2_surf->color_buffers[i].bo)
198 gbm_bo_destroy(dri2_surf->color_buffers[i].bo);
233 if (dri2_surf->back->bo == NULL
270 struct gbm_dri_bo *bo; local
392 struct gbm_dri_bo *bo; local
526 struct gbm_dri_bo *bo; local
560 struct gbm_dri_bo *bo; local
    [all...]
  /external/mesa3d/src/gallium/auxiliary/pipebuffer/
pb_bufmgr_slab.c 92 struct pb_buffer *bo; member in struct:pb_slab
219 pb_reference(&slab->bo, NULL);
259 return pb_validate(buf->slab->bo, vl, flags);
268 pb_fence(buf->slab->bo, fence);
278 pb_get_base_buffer(buf->slab->bo, base_buf, offset);
312 slab->bo = mgr->provider->create_buffer(mgr->provider, mgr->slabSize, &mgr->desc);
313 if(!slab->bo) {
320 slab->virtual = pb_map(slab->bo,
327 pb_unmap(slab->bo);
329 numBuffers = slab->bo->size / mgr->bufSize
    [all...]
  /external/mesa3d/src/gallium/drivers/etnaviv/
etnaviv_resource.c 207 struct etna_bo *bo = etna_bo_new(screen->dev, size, flags); local
208 if (unlikely(bo == NULL)) {
213 rsc->bo = bo;
220 void *map = etna_bo_map(bo);
285 if (rsc->bo)
286 etna_bo_del(rsc->bo);
327 rsc->bo = etna_screen_bo_from_handle(pscreen, handle, &level->stride);
328 if (!rsc->bo)
342 /* The DDX must give us a BO which conforms to our padding size
    [all...]
  /external/mesa3d/src/gallium/drivers/ilo/
ilo_resource.c 295 struct intel_bo *bo; local
297 bo = intel_winsys_alloc_bo(is->dev.winsys, name,
301 if (bo && (tex->image.tiling == GEN6_TILING_X ||
306 if (intel_bo_set_tiling(bo, tiling, tex->image.bo_stride)) {
307 intel_bo_unref(bo);
308 bo = NULL;
311 if (!bo)
314 intel_bo_unref(tex->vma.bo);
315 ilo_vma_set_bo(&tex->vma, &is->dev, bo, 0);
353 struct intel_bo *bo; local
385 struct intel_bo *bo; local
444 struct intel_bo *bo; local
600 struct intel_bo *bo; local
    [all...]
  /external/mesa3d/src/gallium/drivers/nouveau/
nouveau_mm.c 44 struct nouveau_bo *bo; member in struct:mm_slab
106 /* size of bo allocation for slab with chunks of (1 << chunk_order) bytes */
136 slab->bo = NULL;
139 &slab->bo);
162 /* @return token to identify slab or NULL if we just allocated a new bo */
165 uint32_t size, struct nouveau_bo **bo, uint32_t *offset)
175 bo);
202 nouveau_bo_ref(slab->bo, bo);
273 nouveau_bo_ref(NULL, &slab->bo);
    [all...]
  /external/mesa3d/src/gallium/drivers/nouveau/nv30/
nv30_state_validate.c 123 rbo = nv30_miptree(rsf->base.texture)->base.bo;
124 zbo = nv30_miptree(zsf->base.texture)->base.bo;
143 struct nouveau_bo *bo = nv30_miptree(sf->base.texture)->base.bo; local
146 PUSH_MTHDl(push, NV30_3D(COLOR1_OFFSET), BUFCTX_FB, bo, sf->offset,
153 struct nouveau_bo *bo = nv30_miptree(sf->base.texture)->base.bo; local
156 PUSH_MTHDl(push, NV40_3D(COLOR2_OFFSET), BUFCTX_FB, bo, sf->offset,
164 struct nouveau_bo *bo = nv30_miptree(sf->base.texture)->base.bo; local
    [all...]
nv30_winsys.h 23 PUSH_RELOC(struct nouveau_pushbuf *push, struct nouveau_bo *bo, uint32_t offset,
26 nouveau_pushbuf_reloc(push, bo, offset, flags, vor, tor);
44 struct nouveau_bo *bo, uint32_t access)
46 nouveau_bufctx_refn(bufctx(push), bin, bo, access); local
51 struct nouveau_bo *bo, uint32_t offset, uint32_t access)
54 bo, offset, access | NOUVEAU_BO_LOW, 0, 0)->priv = NULL; local
55 PUSH_DATA(push, bo->offset + offset);
60 struct nouveau_bo *bo, uint32_t access, uint32_t vor, uint32_t tor)
63 bo, 0, access | NOUVEAU_BO_OR, vor, tor)->priv = NULL; local
64 if (bo->flags & NOUVEAU_BO_VRAM
76 bo, data, access | NOUVEAU_BO_OR, vor, tor)->priv = NULL; local
    [all...]
  /external/mesa3d/src/gallium/drivers/nouveau/nv50/
nv50_state_validate.c 39 struct nouveau_bo *bo; local
48 bo = mt->base.bo;
61 if (likely(nouveau_bo_memtype(bo))) {
nv50_vbo.c 221 struct nouveau_bo *bo; local
231 &bo);
234 NOUVEAU_BO_RD, bo);
266 struct nouveau_bo *bo; local
270 base, size, &bo);
272 BCTX_REFN_bo(nv50->bufctx_3d, 3D_VERTEX_TMP, bo_flags, bo);
640 PUSH_REFN(push, buf->bo, NOUVEAU_BO_RD | buf->domain);
645 nouveau_pushbuf_data(push, buf->bo, base + start * 4, count * 4);
654 nouveau_pushbuf_data(push, buf->bo, base + pb_start, pb_bytes);
666 nouveau_pushbuf_data(push, buf->bo, base + pb_start, pb_bytes)
    [all...]
  /external/mesa3d/src/gallium/drivers/nouveau/nvc0/
nve4_compute.c 368 struct nouveau_bo *bo = nvc0->screen->uniform_bo; local
375 PUSH_DATAh(push, bo->offset + base);
376 PUSH_DATA (push, bo->offset + base);
519 PUSH_REFN(push, res->bo, NOUVEAU_BO_RD | res->domain);
524 nouveau_pushbuf_data(push, res->bo, offset,
659 PUSH_REFN(push, res->bo, NOUVEAU_BO_RD | res->domain);
663 nouveau_pushbuf_data(push, res->bo, offset,
676 nouveau_pushbuf_data(push, res->bo, offset + 8,
841 struct nouveau_bo *bo = screen->parm; local
846 ret = nouveau_bo_map(bo, NOUVEAU_BO_RDWR, nvc0->base.client)
    [all...]
  /external/mesa3d/src/gallium/drivers/vc4/kernel/
vc4_drv.h 39 struct drm_gem_cma_object **bo; member in struct:vc4_exec_info
47 /* Current unvalidated indices into @bo loaded by the non-hardware
52 /* This is the BO where we store the validated command lists, shader
121 * width/height/stride/etc. from p1 and p2/p3 doesn't sample outside the BO.
140 * record's uniform BO pointer can provide, and we need to apply relocations
  /external/mesa3d/src/gallium/drivers/vc4/
vc4_bufmgr.c 82 vc4_bo_remove_from_cache(struct vc4_bo_cache *cache, struct vc4_bo *bo)
84 list_del(&bo->time_list);
85 list_del(&bo->size_list);
87 cache->bo_size -= bo->size;
99 struct vc4_bo *bo = NULL; local
102 bo = LIST_ENTRY(struct vc4_bo, cache->size_list[page_index].next,
105 /* Check that the BO has gone idle. If not, then we want to
109 if (!vc4_bo_wait(bo, 0, NULL)) {
114 pipe_reference_init(&bo->reference, 1);
115 vc4_bo_remove_from_cache(cache, bo);
126 struct vc4_bo *bo; local
321 struct vc4_bo *bo; local
415 struct vc4_bo *bo; local
    [all...]
  /external/mesa3d/src/gallium/winsys/amdgpu/drm/
amdgpu_bo.c 51 struct amdgpu_winsys_bo *bo = amdgpu_winsys_bo(_buf); local
52 struct amdgpu_winsys *ws = bo->ws;
56 if (p_atomic_read(&bo->num_active_ioctls))
63 if (!os_wait_until_zero_abs_timeout(&bo->num_active_ioctls, abs_timeout))
67 if (bo->is_shared) {
75 r = amdgpu_bo_wait_for_idle(bo->bo, timeout, &buffer_busy);
88 for (idle_fences = 0; idle_fences < bo->num_fences; ++idle_fences) {
89 if (!amdgpu_fence_wait(bo->fences[idle_fences], 0, false))
95 amdgpu_fence_reference(&bo->fences[i], NULL)
159 struct amdgpu_winsys_bo *bo = amdgpu_winsys_bo(_buf); local
191 struct amdgpu_winsys_bo *bo = amdgpu_winsys_bo(_buf); local
205 struct amdgpu_winsys_bo *bo = (struct amdgpu_winsys_bo*)buf; local
320 struct amdgpu_winsys_bo *bo = (struct amdgpu_winsys_bo*)buf; local
366 struct amdgpu_winsys_bo *bo; local
447 struct amdgpu_winsys_bo *bo = amdgpu_winsys_bo(_buf); local
458 struct amdgpu_winsys_bo *bo = NULL; \/* fix container_of *\/ local
466 struct amdgpu_winsys_bo *bo = amdgpu_winsys_bo(_buf); local
528 struct amdgpu_winsys_bo *bo = &slab->entries[i]; local
598 struct amdgpu_winsys_bo *bo = amdgpu_winsys_bo(_buf); local
634 struct amdgpu_winsys_bo *bo = amdgpu_winsys_bo(_buf); local
675 struct amdgpu_winsys_bo *bo; local
783 struct amdgpu_winsys_bo *bo; local
875 struct amdgpu_winsys_bo *bo = amdgpu_winsys_bo(buffer); local
916 struct amdgpu_winsys_bo *bo; local
    [all...]
  /external/mesa3d/src/gallium/winsys/intel/drm/
intel_drm_winsys.c 65 gem_bo(const struct intel_bo *bo)
67 return (drm_intel_bo *) bo;
94 drm_intel_bo *bo; local
98 bo = drm_intel_bo_alloc_tiled(winsys->bufmgr,
100 if (bo) {
101 drm_intel_bo_get_tiling(bo, &tiling, &swizzle);
102 drm_intel_bo_unreference(bo);
127 * uploaded to, say, the beginning of a bo, we want the first vertex in the
128 * bo to be fetched. One way to do this is to set the base address of the
131 * bo->offset64 + (vb->buffer_offset - vb->stride * N)
282 drm_intel_bo *bo; local
313 drm_intel_bo *bo; local
    [all...]
  /external/mesa3d/src/gallium/winsys/radeon/drm/
radeon_drm_cs.c 129 p_atomic_dec(&csc->relocs_bo[i].bo->num_cs_references);
130 radeon_bo_reference(&csc->relocs_bo[i].bo, NULL);
133 p_atomic_dec(&csc->slab_buffers[i].bo->num_cs_references);
134 radeon_bo_reference(&csc->slab_buffers[i].bo, NULL);
198 int radeon_lookup_buffer(struct radeon_cs_context *csc, struct radeon_bo *bo)
200 unsigned hash = bo->hash & (ARRAY_SIZE(csc->reloc_indices_hashlist)-1);
205 if (bo->handle) {
214 if (i == -1 || (i < num_buffers && buffers[i].bo == bo))
217 /* Hash collision, look for the BO in the list of relocs linearly. *
348 struct radeon_bo *bo = (struct radeon_bo*)buf; local
602 struct radeon_bo *bo = cs->csc->slab_buffers[i].bo; local
723 struct radeon_bo *bo = (struct radeon_bo*)_buf; local
    [all...]
radeon_drm_cs.h 33 struct radeon_bo *bo; member in struct:radeon_bo_item
92 int radeon_lookup_buffer(struct radeon_cs_context *csc, struct radeon_bo *bo);
102 struct radeon_bo *bo)
104 int num_refs = bo->num_cs_references;
105 return num_refs == bo->rws->num_cs ||
106 (num_refs && radeon_lookup_buffer(cs->csc, bo) != -1);
111 struct radeon_bo *bo)
115 if (!bo->num_cs_references)
118 index = radeon_lookup_buffer(cs->csc, bo);
122 if (!bo->handle
    [all...]
  /external/mesa3d/src/mesa/drivers/dri/i915/
intel_buffer_objects.c 203 drm_intel_bo_references(intel->batch.bo, intel_obj->buffer);
207 /* Replace the current busy bo with fresh data. */
249 if (drm_intel_bo_references(intel->batch.bo, intel_obj->buffer)) {
271 * invalidated, then we can make a new little BO, let them write into that,
272 * and blit it into the real BO at unmap time.
322 if (drm_intel_bo_references(intel->batch.bo, intel_obj->buffer)) {
340 * BO, and we'll copy what they put in there out at unmap or
385 /* Ideally we'd use a BO to avoid taking up cache space for the temporary
506 if (!intel->upload.bo)
510 drm_intel_bo_subdata(intel->upload.bo,
687 drm_intel_bo *bo = intel_bufferobj_buffer(intel, intel_obj); local
    [all...]
  /external/mesa3d/src/mesa/drivers/dri/i965/
brw_draw.c 202 /* If indirect, emit a bunch of loads from the indirect BO. */
223 drm_intel_bo *bo = intel_bufferobj_buffer(brw, local
229 brw_load_register_mem(brw, GEN7_3DPRIM_VERTEX_COUNT, bo,
232 brw_load_register_mem(brw, GEN7_3DPRIM_INSTANCE_COUNT, bo,
236 brw_load_register_mem(brw, GEN7_3DPRIM_START_VERTEX, bo,
240 brw_load_register_mem(brw, GEN7_3DPRIM_BASE_VERTEX, bo,
243 brw_load_register_mem(brw, GEN7_3DPRIM_START_INSTANCE, bo,
247 brw_load_register_mem(brw, GEN7_3DPRIM_START_INSTANCE, bo,
294 drm_intel_bo_unreference(brw->vb.buffers[i].bo);
295 brw->vb.buffers[i].bo = NULL
    [all...]
gen7_sol_state.c 55 drm_intel_bo *bo; local
78 bo = intel_bufferobj_buffer(brw, bufferobj, start, end - start);
79 assert(end <= bo->size);
84 OUT_RELOC(bo, I915_GEM_DOMAIN_RENDER, I915_GEM_DOMAIN_RENDER, start);
85 OUT_RELOC(bo, I915_GEM_DOMAIN_RENDER, I915_GEM_DOMAIN_RENDER, end);
363 if (drm_intel_bo_references(brw->batch.bo, obj->prim_count_bo))
409 /* Gather up the results so far and release the BO. */

Completed in 353 milliseconds

1 2 3 45 6 7 8 910