1 #ifndef __NOUVEAU_BUFFER_H__ 2 #define __NOUVEAU_BUFFER_H__ 3 4 #include "util/u_range.h" 5 #include "util/u_transfer.h" 6 #include "util/list.h" 7 8 struct pipe_resource; 9 struct nouveau_context; 10 struct nouveau_bo; 11 12 /* DIRTY: buffer was (or will be after the next flush) written to by GPU and 13 * resource->data has not been updated to reflect modified VRAM contents 14 * 15 * USER_MEMORY: resource->data is a pointer to client memory and may change 16 * between GL calls 17 */ 18 #define NOUVEAU_BUFFER_STATUS_GPU_READING (1 << 0) 19 #define NOUVEAU_BUFFER_STATUS_GPU_WRITING (1 << 1) 20 #define NOUVEAU_BUFFER_STATUS_DIRTY (1 << 2) 21 #define NOUVEAU_BUFFER_STATUS_USER_MEMORY (1 << 7) 22 23 #define NOUVEAU_BUFFER_STATUS_REALLOC_MASK NOUVEAU_BUFFER_STATUS_USER_MEMORY 24 25 /* Resources, if mapped into the GPU's address space, are guaranteed to 26 * have constant virtual addresses (nv50+). 27 * 28 * The address of a resource will lie within the nouveau_bo referenced, 29 * and this bo should be added to the memory manager's validation list. 30 */ 31 struct nv04_resource { 32 struct pipe_resource base; 33 const struct u_resource_vtbl *vtbl; 34 35 uint64_t address; /* virtual address (nv50+) */ 36 37 uint8_t *data; /* resource's contents, if domain == 0, or cached */ 38 struct nouveau_bo *bo; 39 uint32_t offset; /* offset into the data/bo */ 40 41 uint8_t status; 42 uint8_t domain; 43 44 uint16_t cb_bindings[6]; /* per-shader per-slot bindings */ 45 46 struct nouveau_fence *fence; 47 struct nouveau_fence *fence_wr; 48 49 struct nouveau_mm_allocation *mm; 50 51 /* buffer range that has been initialized */ 52 struct util_range valid_buffer_range; 53 }; 54 55 void 56 nouveau_buffer_release_gpu_storage(struct nv04_resource *); 57 58 void 59 nouveau_copy_buffer(struct nouveau_context *, 60 struct nv04_resource *dst, unsigned dst_pos, 61 struct nv04_resource *src, unsigned src_pos, unsigned size); 62 63 bool 64 nouveau_buffer_migrate(struct nouveau_context *, 65 struct nv04_resource *, unsigned domain); 66 67 void * 68 nouveau_resource_map_offset(struct nouveau_context *, struct nv04_resource *, 69 uint32_t offset, uint32_t flags); 70 71 static inline void 72 nouveau_resource_unmap(struct nv04_resource *res) 73 { 74 /* no-op */ 75 } 76 77 static inline struct nv04_resource * 78 nv04_resource(struct pipe_resource *resource) 79 { 80 return (struct nv04_resource *)resource; 81 } 82 83 /* is resource mapped into the GPU's address space (i.e. VRAM or GART) ? */ 84 static inline bool 85 nouveau_resource_mapped_by_gpu(struct pipe_resource *resource) 86 { 87 return nv04_resource(resource)->domain != 0; 88 } 89 90 struct pipe_resource * 91 nouveau_buffer_create(struct pipe_screen *pscreen, 92 const struct pipe_resource *templ); 93 94 struct pipe_resource * 95 nouveau_user_buffer_create(struct pipe_screen *screen, void *ptr, 96 unsigned bytes, unsigned usage); 97 98 bool 99 nouveau_user_buffer_upload(struct nouveau_context *, struct nv04_resource *, 100 unsigned base, unsigned size); 101 102 void 103 nouveau_buffer_invalidate(struct pipe_context *pipe, 104 struct pipe_resource *resource); 105 106 /* Copy data to a scratch buffer and return address & bo the data resides in. 107 * Returns 0 on failure. 108 */ 109 uint64_t 110 nouveau_scratch_data(struct nouveau_context *, 111 const void *data, unsigned base, unsigned size, 112 struct nouveau_bo **); 113 114 #endif 115