1 /* 2 * Copyright 2012 Advanced Micro Devices, Inc. 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * on the rights to use, copy, modify, merge, publish, distribute, sub 8 * license, and/or sell copies of the Software, and to permit persons to whom 9 * the Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice (including the next 12 * paragraph) shall be included in all copies or substantial portions of the 13 * Software. 14 * 15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 17 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL 18 * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM, 19 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR 20 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE 21 * USE OR OTHER DEALINGS IN THE SOFTWARE. 22 * 23 * Authors: 24 * Christian Knig <christian.koenig (at) amd.com> 25 */ 26 27 #include "util/u_memory.h" 28 #include "radeonsi_pipe.h" 29 #include "radeonsi_pm4.h" 30 #include "sid.h" 31 #include "r600_hw_context_priv.h" 32 33 #define NUMBER_OF_STATES (sizeof(union si_state) / sizeof(struct si_pm4_state *)) 34 35 void si_pm4_cmd_begin(struct si_pm4_state *state, unsigned opcode) 36 { 37 state->last_opcode = opcode; 38 state->last_pm4 = state->ndw++; 39 } 40 41 void si_pm4_cmd_add(struct si_pm4_state *state, uint32_t dw) 42 { 43 state->pm4[state->ndw++] = dw; 44 } 45 46 void si_pm4_cmd_end(struct si_pm4_state *state, bool predicate) 47 { 48 unsigned count; 49 count = state->ndw - state->last_pm4 - 2; 50 state->pm4[state->last_pm4] = PKT3(state->last_opcode, 51 count, predicate); 52 53 assert(state->ndw <= SI_PM4_MAX_DW); 54 } 55 56 void si_pm4_set_reg(struct si_pm4_state *state, unsigned reg, uint32_t val) 57 { 58 unsigned opcode; 59 60 if (reg >= SI_CONFIG_REG_OFFSET && reg <= SI_CONFIG_REG_END) { 61 opcode = PKT3_SET_CONFIG_REG; 62 reg -= SI_CONFIG_REG_OFFSET; 63 64 } else if (reg >= SI_SH_REG_OFFSET && reg <= SI_SH_REG_END) { 65 opcode = PKT3_SET_SH_REG; 66 reg -= SI_SH_REG_OFFSET; 67 68 } else if (reg >= SI_CONTEXT_REG_OFFSET && reg <= SI_CONTEXT_REG_END) { 69 opcode = PKT3_SET_CONTEXT_REG; 70 reg -= SI_CONTEXT_REG_OFFSET; 71 } else { 72 R600_ERR("Invalid register offset %08x!\n", reg); 73 return; 74 } 75 76 reg >>= 2; 77 78 if (opcode != state->last_opcode || reg != (state->last_reg + 1)) { 79 si_pm4_cmd_begin(state, opcode); 80 si_pm4_cmd_add(state, reg); 81 } 82 83 state->last_reg = reg; 84 si_pm4_cmd_add(state, val); 85 si_pm4_cmd_end(state, false); 86 } 87 88 void si_pm4_add_bo(struct si_pm4_state *state, 89 struct si_resource *bo, 90 enum radeon_bo_usage usage) 91 { 92 unsigned idx = state->nbo++; 93 assert(idx < SI_PM4_MAX_BO); 94 95 si_resource_reference(&state->bo[idx], bo); 96 state->bo_usage[idx] = usage; 97 } 98 99 void si_pm4_sh_data_begin(struct si_pm4_state *state) 100 { 101 si_pm4_cmd_begin(state, PKT3_NOP); 102 } 103 104 void si_pm4_sh_data_add(struct si_pm4_state *state, uint32_t dw) 105 { 106 si_pm4_cmd_add(state, dw); 107 } 108 109 void si_pm4_sh_data_end(struct si_pm4_state *state, unsigned reg) 110 { 111 unsigned offs = state->last_pm4 + 1; 112 113 /* Bail if no data was added */ 114 if (state->ndw == offs) { 115 state->ndw--; 116 return; 117 } 118 119 si_pm4_cmd_end(state, false); 120 121 si_pm4_cmd_begin(state, PKT3_SET_SH_REG_OFFSET); 122 si_pm4_cmd_add(state, (reg - SI_SH_REG_OFFSET) >> 2); 123 state->relocs[state->nrelocs++] = state->ndw; 124 si_pm4_cmd_add(state, offs << 2); 125 si_pm4_cmd_add(state, 0); 126 si_pm4_cmd_end(state, false); 127 } 128 129 void si_pm4_inval_shader_cache(struct si_pm4_state *state) 130 { 131 state->cp_coher_cntl |= S_0085F0_SH_ICACHE_ACTION_ENA(1); 132 state->cp_coher_cntl |= S_0085F0_SH_KCACHE_ACTION_ENA(1); 133 } 134 135 void si_pm4_inval_texture_cache(struct si_pm4_state *state) 136 { 137 state->cp_coher_cntl |= S_0085F0_TC_ACTION_ENA(1); 138 } 139 140 void si_pm4_inval_vertex_cache(struct si_pm4_state *state) 141 { 142 /* Some GPUs don't have the vertex cache and must use the texture cache instead. */ 143 state->cp_coher_cntl |= S_0085F0_TC_ACTION_ENA(1); 144 } 145 146 void si_pm4_inval_fb_cache(struct si_pm4_state *state, unsigned nr_cbufs) 147 { 148 state->cp_coher_cntl |= S_0085F0_CB_ACTION_ENA(1); 149 state->cp_coher_cntl |= ((1 << nr_cbufs) - 1) << S_0085F0_CB0_DEST_BASE_ENA_SHIFT; 150 } 151 152 void si_pm4_inval_zsbuf_cache(struct si_pm4_state *state) 153 { 154 state->cp_coher_cntl |= S_0085F0_DB_ACTION_ENA(1) | S_0085F0_DB_DEST_BASE_ENA(1); 155 } 156 157 void si_pm4_free_state(struct r600_context *rctx, 158 struct si_pm4_state *state, 159 unsigned idx) 160 { 161 if (state == NULL) 162 return; 163 164 if (idx != ~0 && rctx->emitted.array[idx] == state) { 165 rctx->emitted.array[idx] = NULL; 166 } 167 168 for (int i = 0; i < state->nbo; ++i) { 169 si_resource_reference(&state->bo[i], NULL); 170 } 171 FREE(state); 172 } 173 174 uint32_t si_pm4_sync_flags(struct r600_context *rctx) 175 { 176 uint32_t cp_coher_cntl = 0; 177 178 for (int i = 0; i < NUMBER_OF_STATES; ++i) { 179 struct si_pm4_state *state = rctx->queued.array[i]; 180 181 if (!state || rctx->emitted.array[i] == state) 182 continue; 183 184 cp_coher_cntl |= state->cp_coher_cntl; 185 } 186 return cp_coher_cntl; 187 } 188 189 unsigned si_pm4_dirty_dw(struct r600_context *rctx) 190 { 191 unsigned count = 0; 192 193 for (int i = 0; i < NUMBER_OF_STATES; ++i) { 194 struct si_pm4_state *state = rctx->queued.array[i]; 195 196 if (!state || rctx->emitted.array[i] == state) 197 continue; 198 199 count += state->ndw; 200 } 201 202 return count; 203 } 204 205 void si_pm4_emit(struct r600_context *rctx, struct si_pm4_state *state) 206 { 207 struct radeon_winsys_cs *cs = rctx->cs; 208 for (int i = 0; i < state->nbo; ++i) { 209 r600_context_bo_reloc(rctx, state->bo[i], 210 state->bo_usage[i]); 211 } 212 213 memcpy(&cs->buf[cs->cdw], state->pm4, state->ndw * 4); 214 215 for (int i = 0; i < state->nrelocs; ++i) { 216 cs->buf[cs->cdw + state->relocs[i]] += cs->cdw << 2; 217 } 218 219 cs->cdw += state->ndw; 220 } 221 222 void si_pm4_emit_dirty(struct r600_context *rctx) 223 { 224 for (int i = 0; i < NUMBER_OF_STATES; ++i) { 225 struct si_pm4_state *state = rctx->queued.array[i]; 226 227 if (!state || rctx->emitted.array[i] == state) 228 continue; 229 230 si_pm4_emit(rctx, state); 231 rctx->emitted.array[i] = state; 232 } 233 } 234 235 void si_pm4_reset_emitted(struct r600_context *rctx) 236 { 237 memset(&rctx->emitted, 0, sizeof(rctx->emitted)); 238 } 239