1 /* 2 * Copyright 2016 Intel Corporation 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice (including the next 12 * paragraph) shall be included in all copies or substantial portions of the 13 * Software. 14 * 15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING 20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS 21 * IN THE SOFTWARE. 22 */ 23 24 #include "anv_private.h" 25 26 #include "genxml/gen_macros.h" 27 #include "genxml/genX_pack.h" 28 29 #include "common/gen_l3_config.h" 30 31 /** 32 * This file implements some lightweight memcpy/memset operations on the GPU 33 * using a vertex buffer and streamout. 34 */ 35 36 /** 37 * Returns the greatest common divisor of a and b that is a power of two. 38 */ 39 static inline uint64_t 40 gcd_pow2_u64(uint64_t a, uint64_t b) 41 { 42 assert(a > 0 || b > 0); 43 44 unsigned a_log2 = ffsll(a) - 1; 45 unsigned b_log2 = ffsll(b) - 1; 46 47 /* If either a or b is 0, then a_log2 or b_log2 will be UINT_MAX in which 48 * case, the MIN2() will take the other one. If both are 0 then we will 49 * hit the assert above. 50 */ 51 return 1 << MIN2(a_log2, b_log2); 52 } 53 54 void 55 genX(cmd_buffer_gpu_memcpy)(struct anv_cmd_buffer *cmd_buffer, 56 struct anv_bo *dst, uint32_t dst_offset, 57 struct anv_bo *src, uint32_t src_offset, 58 uint32_t size) 59 { 60 if (size == 0) 61 return; 62 63 assert(dst_offset + size <= dst->size); 64 assert(src_offset + size <= src->size); 65 66 /* The maximum copy block size is 4 32-bit components at a time. */ 67 unsigned bs = 16; 68 bs = gcd_pow2_u64(bs, src_offset); 69 bs = gcd_pow2_u64(bs, dst_offset); 70 bs = gcd_pow2_u64(bs, size); 71 72 enum isl_format format; 73 switch (bs) { 74 case 4: format = ISL_FORMAT_R32_UINT; break; 75 case 8: format = ISL_FORMAT_R32G32_UINT; break; 76 case 16: format = ISL_FORMAT_R32G32B32A32_UINT; break; 77 default: 78 unreachable("Invalid size"); 79 } 80 81 if (!cmd_buffer->state.current_l3_config) { 82 const struct gen_l3_config *cfg = 83 gen_get_default_l3_config(&cmd_buffer->device->info); 84 genX(cmd_buffer_config_l3)(cmd_buffer, cfg); 85 } 86 87 genX(cmd_buffer_apply_pipe_flushes)(cmd_buffer); 88 89 genX(flush_pipeline_select_3d)(cmd_buffer); 90 91 uint32_t *dw; 92 dw = anv_batch_emitn(&cmd_buffer->batch, 5, GENX(3DSTATE_VERTEX_BUFFERS)); 93 GENX(VERTEX_BUFFER_STATE_pack)(&cmd_buffer->batch, dw + 1, 94 &(struct GENX(VERTEX_BUFFER_STATE)) { 95 .VertexBufferIndex = 32, /* Reserved for this */ 96 .AddressModifyEnable = true, 97 .BufferStartingAddress = { src, src_offset }, 98 .BufferPitch = bs, 99 #if (GEN_GEN >= 8) 100 .MemoryObjectControlState = GENX(MOCS), 101 .BufferSize = size, 102 #else 103 .VertexBufferMemoryObjectControlState = GENX(MOCS), 104 .EndAddress = { src, src_offset + size - 1 }, 105 #endif 106 }); 107 108 dw = anv_batch_emitn(&cmd_buffer->batch, 3, GENX(3DSTATE_VERTEX_ELEMENTS)); 109 GENX(VERTEX_ELEMENT_STATE_pack)(&cmd_buffer->batch, dw + 1, 110 &(struct GENX(VERTEX_ELEMENT_STATE)) { 111 .VertexBufferIndex = 32, 112 .Valid = true, 113 .SourceElementFormat = format, 114 .SourceElementOffset = 0, 115 .Component0Control = (bs >= 4) ? VFCOMP_STORE_SRC : VFCOMP_STORE_0, 116 .Component1Control = (bs >= 8) ? VFCOMP_STORE_SRC : VFCOMP_STORE_0, 117 .Component2Control = (bs >= 12) ? VFCOMP_STORE_SRC : VFCOMP_STORE_0, 118 .Component3Control = (bs >= 16) ? VFCOMP_STORE_SRC : VFCOMP_STORE_0, 119 }); 120 121 #if GEN_GEN >= 8 122 anv_batch_emit(&cmd_buffer->batch, GENX(3DSTATE_VF_SGVS), sgvs); 123 #endif 124 125 /* Disable all shader stages */ 126 anv_batch_emit(&cmd_buffer->batch, GENX(3DSTATE_VS), vs); 127 anv_batch_emit(&cmd_buffer->batch, GENX(3DSTATE_HS), hs); 128 anv_batch_emit(&cmd_buffer->batch, GENX(3DSTATE_TE), te); 129 anv_batch_emit(&cmd_buffer->batch, GENX(3DSTATE_DS), DS); 130 anv_batch_emit(&cmd_buffer->batch, GENX(3DSTATE_GS), gs); 131 anv_batch_emit(&cmd_buffer->batch, GENX(3DSTATE_PS), gs); 132 133 anv_batch_emit(&cmd_buffer->batch, GENX(3DSTATE_SBE), sbe) { 134 sbe.VertexURBEntryReadOffset = 1; 135 sbe.NumberofSFOutputAttributes = 1; 136 sbe.VertexURBEntryReadLength = 1; 137 #if GEN_GEN >= 8 138 sbe.ForceVertexURBEntryReadLength = true; 139 sbe.ForceVertexURBEntryReadOffset = true; 140 #endif 141 142 #if GEN_GEN >= 9 143 for (unsigned i = 0; i < 32; i++) 144 sbe.AttributeActiveComponentFormat[i] = ACF_XYZW; 145 #endif 146 } 147 148 /* Emit URB setup. We tell it that the VS is active because we want it to 149 * allocate space for the VS. Even though one isn't run, we need VUEs to 150 * store the data that VF is going to pass to SOL. 151 */ 152 const unsigned entry_size[4] = { DIV_ROUND_UP(32, 64), 1, 1, 1 }; 153 154 genX(emit_urb_setup)(cmd_buffer->device, &cmd_buffer->batch, 155 cmd_buffer->state.current_l3_config, 156 VK_SHADER_STAGE_VERTEX_BIT, entry_size); 157 158 anv_batch_emit(&cmd_buffer->batch, GENX(3DSTATE_SO_BUFFER), sob) { 159 sob.SOBufferIndex = 0; 160 sob.SOBufferObjectControlState = GENX(MOCS); 161 sob.SurfaceBaseAddress = (struct anv_address) { dst, dst_offset }; 162 163 #if GEN_GEN >= 8 164 sob.SOBufferEnable = true; 165 sob.SurfaceSize = size - 1; 166 #else 167 sob.SurfacePitch = bs; 168 sob.SurfaceEndAddress = sob.SurfaceBaseAddress; 169 sob.SurfaceEndAddress.offset += size; 170 #endif 171 172 #if GEN_GEN >= 8 173 /* As SOL writes out data, it updates the SO_WRITE_OFFSET registers with 174 * the end position of the stream. We need to reset this value to 0 at 175 * the beginning of the run or else SOL will start at the offset from 176 * the previous draw. 177 */ 178 sob.StreamOffsetWriteEnable = true; 179 sob.StreamOffset = 0; 180 #endif 181 } 182 183 #if GEN_GEN <= 7 184 /* The hardware can do this for us on BDW+ (see above) */ 185 anv_batch_emit(&cmd_buffer->batch, GENX(MI_LOAD_REGISTER_IMM), load) { 186 load.RegisterOffset = GENX(SO_WRITE_OFFSET0_num); 187 load.DataDWord = 0; 188 } 189 #endif 190 191 dw = anv_batch_emitn(&cmd_buffer->batch, 5, GENX(3DSTATE_SO_DECL_LIST), 192 .StreamtoBufferSelects0 = (1 << 0), 193 .NumEntries0 = 1); 194 GENX(SO_DECL_ENTRY_pack)(&cmd_buffer->batch, dw + 3, 195 &(struct GENX(SO_DECL_ENTRY)) { 196 .Stream0Decl = { 197 .OutputBufferSlot = 0, 198 .RegisterIndex = 0, 199 .ComponentMask = (1 << (bs / 4)) - 1, 200 }, 201 }); 202 203 anv_batch_emit(&cmd_buffer->batch, GENX(3DSTATE_STREAMOUT), so) { 204 so.SOFunctionEnable = true; 205 so.RenderingDisable = true; 206 so.Stream0VertexReadOffset = 0; 207 so.Stream0VertexReadLength = DIV_ROUND_UP(32, 64); 208 #if GEN_GEN >= 8 209 so.Buffer0SurfacePitch = bs; 210 #else 211 so.SOBufferEnable0 = true; 212 #endif 213 } 214 215 #if GEN_GEN >= 8 216 anv_batch_emit(&cmd_buffer->batch, GENX(3DSTATE_VF_TOPOLOGY), topo) { 217 topo.PrimitiveTopologyType = _3DPRIM_POINTLIST; 218 } 219 #endif 220 221 anv_batch_emit(&cmd_buffer->batch, GENX(3DPRIMITIVE), prim) { 222 prim.VertexAccessType = SEQUENTIAL; 223 prim.PrimitiveTopologyType = _3DPRIM_POINTLIST; 224 prim.VertexCountPerInstance = size / bs; 225 prim.StartVertexLocation = 0; 226 prim.InstanceCount = 1; 227 prim.StartInstanceLocation = 0; 228 prim.BaseVertexLocation = 0; 229 } 230 231 cmd_buffer->state.dirty |= ANV_CMD_DIRTY_PIPELINE; 232 } 233