Home | History | Annotate | Download | only in r600
      1 /*
      2  * Copyright 2015 Advanced Micro Devices, Inc.
      3  *
      4  * Permission is hereby granted, free of charge, to any person obtaining a
      5  * copy of this software and associated documentation files (the "Software"),
      6  * to deal in the Software without restriction, including without limitation
      7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
      8  * and/or sell copies of the Software, and to permit persons to whom the
      9  * Software is furnished to do so, subject to the following conditions:
     10  *
     11  * The above copyright notice and this permission notice (including the next
     12  * paragraph) shall be included in all copies or substantial portions of the
     13  * Software.
     14  *
     15  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
     16  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
     17  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
     18  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
     19  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
     20  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
     21  * SOFTWARE.
     22  *
     23  * Authors: Marek Olk <maraeo (at) gmail.com>
     24  *
     25  */
     26 
     27 /* The GPU load is measured as follows.
     28  *
     29  * There is a thread which samples the GRBM_STATUS register at a certain
     30  * frequency and the "busy" or "idle" counter is incremented based on
     31  * whether the GUI_ACTIVE bit is set or not.
     32  *
     33  * Then, the user can sample the counters twice and calculate the average
     34  * GPU load between the two samples.
     35  */
     36 
     37 #include "r600_pipe_common.h"
     38 #include "r600_query.h"
     39 #include "util/os_time.h"
     40 
     41 /* For good accuracy at 1000 fps or lower. This will be inaccurate for higher
     42  * fps (there are too few samples per frame). */
     43 #define SAMPLES_PER_SEC 10000
     44 
     45 #define GRBM_STATUS		0x8010
     46 #define TA_BUSY(x)		(((x) >> 14) & 0x1)
     47 #define GDS_BUSY(x)		(((x) >> 15) & 0x1)
     48 #define VGT_BUSY(x)		(((x) >> 17) & 0x1)
     49 #define IA_BUSY(x)		(((x) >> 19) & 0x1)
     50 #define SX_BUSY(x)		(((x) >> 20) & 0x1)
     51 #define WD_BUSY(x)		(((x) >> 21) & 0x1)
     52 #define SPI_BUSY(x)		(((x) >> 22) & 0x1)
     53 #define BCI_BUSY(x)		(((x) >> 23) & 0x1)
     54 #define SC_BUSY(x)		(((x) >> 24) & 0x1)
     55 #define PA_BUSY(x)		(((x) >> 25) & 0x1)
     56 #define DB_BUSY(x)		(((x) >> 26) & 0x1)
     57 #define CP_BUSY(x)		(((x) >> 29) & 0x1)
     58 #define CB_BUSY(x)		(((x) >> 30) & 0x1)
     59 #define GUI_ACTIVE(x)		(((x) >> 31) & 0x1)
     60 
     61 #define SRBM_STATUS2		0x0e4c
     62 #define SDMA_BUSY(x)		(((x) >> 5) & 0x1)
     63 
     64 #define CP_STAT                 0x8680
     65 #define PFP_BUSY(x)		(((x) >> 15) & 0x1)
     66 #define MEQ_BUSY(x)		(((x) >> 16) & 0x1)
     67 #define ME_BUSY(x)		(((x) >> 17) & 0x1)
     68 #define SURFACE_SYNC_BUSY(x)	(((x) >> 21) & 0x1)
     69 #define DMA_BUSY(x)		(((x) >> 22) & 0x1)
     70 #define SCRATCH_RAM_BUSY(x)	(((x) >> 24) & 0x1)
     71 
     72 #define IDENTITY(x) x
     73 
     74 #define UPDATE_COUNTER(field, mask)					\
     75 	do {								\
     76 		if (mask(value))					\
     77 			p_atomic_inc(&counters->named.field.busy);	\
     78 		else							\
     79 			p_atomic_inc(&counters->named.field.idle);	\
     80 	} while (0)
     81 
     82 static void r600_update_mmio_counters(struct r600_common_screen *rscreen,
     83 				      union r600_mmio_counters *counters)
     84 {
     85 	uint32_t value = 0;
     86 	bool gui_busy, sdma_busy = false;
     87 
     88 	/* GRBM_STATUS */
     89 	rscreen->ws->read_registers(rscreen->ws, GRBM_STATUS, 1, &value);
     90 
     91 	UPDATE_COUNTER(ta, TA_BUSY);
     92 	UPDATE_COUNTER(gds, GDS_BUSY);
     93 	UPDATE_COUNTER(vgt, VGT_BUSY);
     94 	UPDATE_COUNTER(ia, IA_BUSY);
     95 	UPDATE_COUNTER(sx, SX_BUSY);
     96 	UPDATE_COUNTER(wd, WD_BUSY);
     97 	UPDATE_COUNTER(spi, SPI_BUSY);
     98 	UPDATE_COUNTER(bci, BCI_BUSY);
     99 	UPDATE_COUNTER(sc, SC_BUSY);
    100 	UPDATE_COUNTER(pa, PA_BUSY);
    101 	UPDATE_COUNTER(db, DB_BUSY);
    102 	UPDATE_COUNTER(cp, CP_BUSY);
    103 	UPDATE_COUNTER(cb, CB_BUSY);
    104 	UPDATE_COUNTER(gui, GUI_ACTIVE);
    105 	gui_busy = GUI_ACTIVE(value);
    106 
    107 	value = gui_busy || sdma_busy;
    108 	UPDATE_COUNTER(gpu, IDENTITY);
    109 }
    110 
    111 #undef UPDATE_COUNTER
    112 
    113 static int
    114 r600_gpu_load_thread(void *param)
    115 {
    116 	struct r600_common_screen *rscreen = (struct r600_common_screen*)param;
    117 	const int period_us = 1000000 / SAMPLES_PER_SEC;
    118 	int sleep_us = period_us;
    119 	int64_t cur_time, last_time = os_time_get();
    120 
    121 	while (!p_atomic_read(&rscreen->gpu_load_stop_thread)) {
    122 		if (sleep_us)
    123 			os_time_sleep(sleep_us);
    124 
    125 		/* Make sure we sleep the ideal amount of time to match
    126 		 * the expected frequency. */
    127 		cur_time = os_time_get();
    128 
    129 		if (os_time_timeout(last_time, last_time + period_us,
    130 				    cur_time))
    131 			sleep_us = MAX2(sleep_us - 1, 1);
    132 		else
    133 			sleep_us += 1;
    134 
    135 		/*printf("Hz: %.1f\n", 1000000.0 / (cur_time - last_time));*/
    136 		last_time = cur_time;
    137 
    138 		/* Update the counters. */
    139 		r600_update_mmio_counters(rscreen, &rscreen->mmio_counters);
    140 	}
    141 	p_atomic_dec(&rscreen->gpu_load_stop_thread);
    142 	return 0;
    143 }
    144 
    145 void r600_gpu_load_kill_thread(struct r600_common_screen *rscreen)
    146 {
    147 	if (!rscreen->gpu_load_thread)
    148 		return;
    149 
    150 	p_atomic_inc(&rscreen->gpu_load_stop_thread);
    151 	thrd_join(rscreen->gpu_load_thread, NULL);
    152 	rscreen->gpu_load_thread = 0;
    153 }
    154 
    155 static uint64_t r600_read_mmio_counter(struct r600_common_screen *rscreen,
    156 				       unsigned busy_index)
    157 {
    158 	/* Start the thread if needed. */
    159 	if (!rscreen->gpu_load_thread) {
    160 		mtx_lock(&rscreen->gpu_load_mutex);
    161 		/* Check again inside the mutex. */
    162 		if (!rscreen->gpu_load_thread)
    163 			rscreen->gpu_load_thread =
    164 				u_thread_create(r600_gpu_load_thread, rscreen);
    165 		mtx_unlock(&rscreen->gpu_load_mutex);
    166 	}
    167 
    168 	unsigned busy = p_atomic_read(&rscreen->mmio_counters.array[busy_index]);
    169 	unsigned idle = p_atomic_read(&rscreen->mmio_counters.array[busy_index + 1]);
    170 
    171 	return busy | ((uint64_t)idle << 32);
    172 }
    173 
    174 static unsigned r600_end_mmio_counter(struct r600_common_screen *rscreen,
    175 				      uint64_t begin, unsigned busy_index)
    176 {
    177 	uint64_t end = r600_read_mmio_counter(rscreen, busy_index);
    178 	unsigned busy = (end & 0xffffffff) - (begin & 0xffffffff);
    179 	unsigned idle = (end >> 32) - (begin >> 32);
    180 
    181 	/* Calculate the % of time the busy counter was being incremented.
    182 	 *
    183 	 * If no counters were incremented, return the current counter status.
    184 	 * It's for the case when the load is queried faster than
    185 	 * the counters are updated.
    186 	 */
    187 	if (idle || busy) {
    188 		return busy*100 / (busy + idle);
    189 	} else {
    190 		union r600_mmio_counters counters;
    191 
    192 		memset(&counters, 0, sizeof(counters));
    193 		r600_update_mmio_counters(rscreen, &counters);
    194 		return counters.array[busy_index] ? 100 : 0;
    195 	}
    196 }
    197 
    198 #define BUSY_INDEX(rscreen, field) (&rscreen->mmio_counters.named.field.busy - \
    199 				    rscreen->mmio_counters.array)
    200 
    201 static unsigned busy_index_from_type(struct r600_common_screen *rscreen,
    202 				     unsigned type)
    203 {
    204 	switch (type) {
    205 	case R600_QUERY_GPU_LOAD:
    206 		return BUSY_INDEX(rscreen, gpu);
    207 	case R600_QUERY_GPU_SHADERS_BUSY:
    208 		return BUSY_INDEX(rscreen, spi);
    209 	case R600_QUERY_GPU_TA_BUSY:
    210 		return BUSY_INDEX(rscreen, ta);
    211 	case R600_QUERY_GPU_GDS_BUSY:
    212 		return BUSY_INDEX(rscreen, gds);
    213 	case R600_QUERY_GPU_VGT_BUSY:
    214 		return BUSY_INDEX(rscreen, vgt);
    215 	case R600_QUERY_GPU_IA_BUSY:
    216 		return BUSY_INDEX(rscreen, ia);
    217 	case R600_QUERY_GPU_SX_BUSY:
    218 		return BUSY_INDEX(rscreen, sx);
    219 	case R600_QUERY_GPU_WD_BUSY:
    220 		return BUSY_INDEX(rscreen, wd);
    221 	case R600_QUERY_GPU_BCI_BUSY:
    222 		return BUSY_INDEX(rscreen, bci);
    223 	case R600_QUERY_GPU_SC_BUSY:
    224 		return BUSY_INDEX(rscreen, sc);
    225 	case R600_QUERY_GPU_PA_BUSY:
    226 		return BUSY_INDEX(rscreen, pa);
    227 	case R600_QUERY_GPU_DB_BUSY:
    228 		return BUSY_INDEX(rscreen, db);
    229 	case R600_QUERY_GPU_CP_BUSY:
    230 		return BUSY_INDEX(rscreen, cp);
    231 	case R600_QUERY_GPU_CB_BUSY:
    232 		return BUSY_INDEX(rscreen, cb);
    233 	case R600_QUERY_GPU_SDMA_BUSY:
    234 		return BUSY_INDEX(rscreen, sdma);
    235 	case R600_QUERY_GPU_PFP_BUSY:
    236 		return BUSY_INDEX(rscreen, pfp);
    237 	case R600_QUERY_GPU_MEQ_BUSY:
    238 		return BUSY_INDEX(rscreen, meq);
    239 	case R600_QUERY_GPU_ME_BUSY:
    240 		return BUSY_INDEX(rscreen, me);
    241 	case R600_QUERY_GPU_SURF_SYNC_BUSY:
    242 		return BUSY_INDEX(rscreen, surf_sync);
    243 	case R600_QUERY_GPU_CP_DMA_BUSY:
    244 		return BUSY_INDEX(rscreen, cp_dma);
    245 	case R600_QUERY_GPU_SCRATCH_RAM_BUSY:
    246 		return BUSY_INDEX(rscreen, scratch_ram);
    247 	default:
    248 		unreachable("invalid query type");
    249 	}
    250 }
    251 
    252 uint64_t r600_begin_counter(struct r600_common_screen *rscreen, unsigned type)
    253 {
    254 	unsigned busy_index = busy_index_from_type(rscreen, type);
    255 	return r600_read_mmio_counter(rscreen, busy_index);
    256 }
    257 
    258 unsigned r600_end_counter(struct r600_common_screen *rscreen, unsigned type,
    259 			  uint64_t begin)
    260 {
    261 	unsigned busy_index = busy_index_from_type(rscreen, type);
    262 	return r600_end_mmio_counter(rscreen, begin, busy_index);
    263 }
    264