Home | History | Annotate | Download | only in minigbm
      1 /*
      2  * Copyright 2014 The Chromium OS Authors. All rights reserved.
      3  * Use of this source code is governed by a BSD-style license that can be
      4  * found in the LICENSE file.
      5  */
      6 
      7 #ifdef DRV_TEGRA
      8 
      9 #include <assert.h>
     10 #include <stdio.h>
     11 #include <string.h>
     12 #include <sys/mman.h>
     13 #include <tegra_drm.h>
     14 #include <xf86drm.h>
     15 
     16 #include "drv_priv.h"
     17 #include "helpers.h"
     18 #include "util.h"
     19 
     20 /*
     21  * GOB (Group Of Bytes) is the basic unit of the blocklinear layout.
     22  * GOBs are arranged to blocks, where the height of the block (measured
     23  * in GOBs) is configurable.
     24  */
     25 #define NV_BLOCKLINEAR_GOB_HEIGHT 8
     26 #define NV_BLOCKLINEAR_GOB_WIDTH 64
     27 #define NV_DEFAULT_BLOCK_HEIGHT_LOG2 4
     28 #define NV_PREFERRED_PAGE_SIZE (128 * 1024)
     29 
     30 // clang-format off
     31 enum nv_mem_kind
     32 {
     33 	NV_MEM_KIND_PITCH = 0,
     34 	NV_MEM_KIND_C32_2CRA = 0xdb,
     35 	NV_MEM_KIND_GENERIC_16Bx2 = 0xfe,
     36 };
     37 
     38 enum tegra_map_type {
     39 	TEGRA_READ_TILED_BUFFER = 0,
     40 	TEGRA_WRITE_TILED_BUFFER = 1,
     41 };
     42 // clang-format on
     43 
     44 struct tegra_private_map_data {
     45 	void *tiled;
     46 	void *untiled;
     47 };
     48 
     49 static const uint32_t render_target_formats[] = { DRM_FORMAT_ARGB8888, DRM_FORMAT_XRGB8888 };
     50 
     51 static int compute_block_height_log2(int height)
     52 {
     53 	int block_height_log2 = NV_DEFAULT_BLOCK_HEIGHT_LOG2;
     54 
     55 	if (block_height_log2 > 0) {
     56 		/* Shrink, if a smaller block height could cover the whole
     57 		 * surface height. */
     58 		int proposed = NV_BLOCKLINEAR_GOB_HEIGHT << (block_height_log2 - 1);
     59 		while (proposed >= height) {
     60 			block_height_log2--;
     61 			if (block_height_log2 == 0)
     62 				break;
     63 			proposed /= 2;
     64 		}
     65 	}
     66 	return block_height_log2;
     67 }
     68 
     69 static void compute_layout_blocklinear(int width, int height, int format, enum nv_mem_kind *kind,
     70 				       uint32_t *block_height_log2, uint32_t *stride,
     71 				       uint32_t *size)
     72 {
     73 	int pitch = drv_stride_from_format(format, width, 0);
     74 
     75 	/* Align to blocklinear blocks. */
     76 	pitch = ALIGN(pitch, NV_BLOCKLINEAR_GOB_WIDTH);
     77 
     78 	/* Compute padded height. */
     79 	*block_height_log2 = compute_block_height_log2(height);
     80 	int block_height = 1 << *block_height_log2;
     81 	int padded_height = ALIGN(height, NV_BLOCKLINEAR_GOB_HEIGHT * block_height);
     82 
     83 	int bytes = pitch * padded_height;
     84 
     85 	/* Pad the allocation to the preferred page size.
     86 	 * This will reduce the required page table size (see discussion in NV
     87 	 * bug 1321091), and also acts as a WAR for NV bug 1325421.
     88 	 */
     89 	bytes = ALIGN(bytes, NV_PREFERRED_PAGE_SIZE);
     90 
     91 	*kind = NV_MEM_KIND_C32_2CRA;
     92 	*stride = pitch;
     93 	*size = bytes;
     94 }
     95 
     96 static void compute_layout_linear(int width, int height, int format, uint32_t *stride,
     97 				  uint32_t *size)
     98 {
     99 	*stride = ALIGN(drv_stride_from_format(format, width, 0), 64);
    100 	*size = *stride * height;
    101 }
    102 
    103 static void transfer_tile(struct bo *bo, uint8_t *tiled, uint8_t *untiled, enum tegra_map_type type,
    104 			  uint32_t bytes_per_pixel, uint32_t gob_top, uint32_t gob_left,
    105 			  uint32_t gob_size_pixels, uint8_t *tiled_last)
    106 {
    107 	uint8_t *tmp;
    108 	uint32_t x, y, k;
    109 	for (k = 0; k < gob_size_pixels; k++) {
    110 		/*
    111 		 * Given the kth pixel starting from the tile specified by
    112 		 * gob_top and gob_left, unswizzle to get the standard (x, y)
    113 		 * representation.
    114 		 */
    115 		x = gob_left + (((k >> 3) & 8) | ((k >> 1) & 4) | (k & 3));
    116 		y = gob_top + ((k >> 7 << 3) | ((k >> 3) & 6) | ((k >> 2) & 1));
    117 
    118 		if (tiled >= tiled_last)
    119 			return;
    120 
    121 		if (x >= bo->width || y >= bo->height) {
    122 			tiled += bytes_per_pixel;
    123 			continue;
    124 		}
    125 
    126 		tmp = untiled + y * bo->strides[0] + x * bytes_per_pixel;
    127 
    128 		if (type == TEGRA_READ_TILED_BUFFER)
    129 			memcpy(tmp, tiled, bytes_per_pixel);
    130 		else if (type == TEGRA_WRITE_TILED_BUFFER)
    131 			memcpy(tiled, tmp, bytes_per_pixel);
    132 
    133 		/* Move on to next pixel. */
    134 		tiled += bytes_per_pixel;
    135 	}
    136 }
    137 
    138 static void transfer_tiled_memory(struct bo *bo, uint8_t *tiled, uint8_t *untiled,
    139 				  enum tegra_map_type type)
    140 {
    141 	uint32_t gob_width, gob_height, gob_size_bytes, gob_size_pixels, gob_count_x, gob_count_y,
    142 	    gob_top, gob_left;
    143 	uint32_t i, j, offset;
    144 	uint8_t *tmp, *tiled_last;
    145 	uint32_t bytes_per_pixel = drv_stride_from_format(bo->format, 1, 0);
    146 
    147 	/*
    148 	 * The blocklinear format consists of 8*(2^n) x 64 byte sized tiles,
    149 	 * where 0 <= n <= 4.
    150 	 */
    151 	gob_width = DIV_ROUND_UP(NV_BLOCKLINEAR_GOB_WIDTH, bytes_per_pixel);
    152 	gob_height = NV_BLOCKLINEAR_GOB_HEIGHT * (1 << NV_DEFAULT_BLOCK_HEIGHT_LOG2);
    153 	/* Calculate the height from maximum possible gob height */
    154 	while (gob_height > NV_BLOCKLINEAR_GOB_HEIGHT && gob_height >= 2 * bo->height)
    155 		gob_height /= 2;
    156 
    157 	gob_size_bytes = gob_height * NV_BLOCKLINEAR_GOB_WIDTH;
    158 	gob_size_pixels = gob_height * gob_width;
    159 
    160 	gob_count_x = DIV_ROUND_UP(bo->strides[0], NV_BLOCKLINEAR_GOB_WIDTH);
    161 	gob_count_y = DIV_ROUND_UP(bo->height, gob_height);
    162 
    163 	tiled_last = tiled + bo->total_size;
    164 
    165 	offset = 0;
    166 	for (j = 0; j < gob_count_y; j++) {
    167 		gob_top = j * gob_height;
    168 		for (i = 0; i < gob_count_x; i++) {
    169 			tmp = tiled + offset;
    170 			gob_left = i * gob_width;
    171 
    172 			transfer_tile(bo, tmp, untiled, type, bytes_per_pixel, gob_top, gob_left,
    173 				      gob_size_pixels, tiled_last);
    174 
    175 			offset += gob_size_bytes;
    176 		}
    177 	}
    178 }
    179 
    180 static int tegra_init(struct driver *drv)
    181 {
    182 	struct format_metadata metadata;
    183 	uint64_t use_flags = BO_USE_RENDER_MASK;
    184 
    185 	metadata.tiling = NV_MEM_KIND_PITCH;
    186 	metadata.priority = 1;
    187 	metadata.modifier = DRM_FORMAT_MOD_LINEAR;
    188 
    189 	drv_add_combinations(drv, render_target_formats, ARRAY_SIZE(render_target_formats),
    190 			     &metadata, use_flags);
    191 
    192 	drv_modify_combination(drv, DRM_FORMAT_XRGB8888, &metadata, BO_USE_CURSOR | BO_USE_SCANOUT);
    193 	drv_modify_combination(drv, DRM_FORMAT_ARGB8888, &metadata, BO_USE_CURSOR | BO_USE_SCANOUT);
    194 
    195 	use_flags &= ~BO_USE_SW_WRITE_OFTEN;
    196 	use_flags &= ~BO_USE_SW_READ_OFTEN;
    197 	use_flags &= ~BO_USE_LINEAR;
    198 
    199 	metadata.tiling = NV_MEM_KIND_C32_2CRA;
    200 	metadata.priority = 2;
    201 
    202 	drv_add_combinations(drv, render_target_formats, ARRAY_SIZE(render_target_formats),
    203 			     &metadata, use_flags);
    204 
    205 	drv_modify_combination(drv, DRM_FORMAT_XRGB8888, &metadata, BO_USE_SCANOUT);
    206 	drv_modify_combination(drv, DRM_FORMAT_ARGB8888, &metadata, BO_USE_SCANOUT);
    207 	return 0;
    208 }
    209 
    210 static int tegra_bo_create(struct bo *bo, uint32_t width, uint32_t height, uint32_t format,
    211 			   uint64_t use_flags)
    212 {
    213 	uint32_t size, stride, block_height_log2 = 0;
    214 	enum nv_mem_kind kind = NV_MEM_KIND_PITCH;
    215 	struct drm_tegra_gem_create gem_create;
    216 	int ret;
    217 
    218 	if (use_flags &
    219 	    (BO_USE_CURSOR | BO_USE_LINEAR | BO_USE_SW_READ_OFTEN | BO_USE_SW_WRITE_OFTEN))
    220 		compute_layout_linear(width, height, format, &stride, &size);
    221 	else
    222 		compute_layout_blocklinear(width, height, format, &kind, &block_height_log2,
    223 					   &stride, &size);
    224 
    225 	memset(&gem_create, 0, sizeof(gem_create));
    226 	gem_create.size = size;
    227 	gem_create.flags = 0;
    228 
    229 	ret = drmIoctl(bo->drv->fd, DRM_IOCTL_TEGRA_GEM_CREATE, &gem_create);
    230 	if (ret) {
    231 		drv_log("DRM_IOCTL_TEGRA_GEM_CREATE failed (size=%zu)\n", size);
    232 		return ret;
    233 	}
    234 
    235 	bo->handles[0].u32 = gem_create.handle;
    236 	bo->offsets[0] = 0;
    237 	bo->total_size = bo->sizes[0] = size;
    238 	bo->strides[0] = stride;
    239 
    240 	if (kind != NV_MEM_KIND_PITCH) {
    241 		struct drm_tegra_gem_set_tiling gem_tile;
    242 
    243 		memset(&gem_tile, 0, sizeof(gem_tile));
    244 		gem_tile.handle = bo->handles[0].u32;
    245 		gem_tile.mode = DRM_TEGRA_GEM_TILING_MODE_BLOCK;
    246 		gem_tile.value = block_height_log2;
    247 
    248 		ret = drmCommandWriteRead(bo->drv->fd, DRM_TEGRA_GEM_SET_TILING, &gem_tile,
    249 					  sizeof(gem_tile));
    250 		if (ret < 0) {
    251 			drv_gem_bo_destroy(bo);
    252 			return ret;
    253 		}
    254 
    255 		/* Encode blocklinear parameters for EGLImage creation. */
    256 		bo->tiling = (kind & 0xff) | ((block_height_log2 & 0xf) << 8);
    257 		bo->format_modifiers[0] = fourcc_mod_code(NV, bo->tiling);
    258 	}
    259 
    260 	return 0;
    261 }
    262 
    263 static int tegra_bo_import(struct bo *bo, struct drv_import_fd_data *data)
    264 {
    265 	int ret;
    266 	struct drm_tegra_gem_get_tiling gem_get_tiling;
    267 
    268 	ret = drv_prime_bo_import(bo, data);
    269 	if (ret)
    270 		return ret;
    271 
    272 	/* TODO(gsingh): export modifiers and get rid of backdoor tiling. */
    273 	memset(&gem_get_tiling, 0, sizeof(gem_get_tiling));
    274 	gem_get_tiling.handle = bo->handles[0].u32;
    275 
    276 	ret = drmIoctl(bo->drv->fd, DRM_IOCTL_TEGRA_GEM_GET_TILING, &gem_get_tiling);
    277 	if (ret) {
    278 		drv_gem_bo_destroy(bo);
    279 		return ret;
    280 	}
    281 
    282 	/* NOTE(djmk): we only know about one tiled format, so if our drmIoctl call tells us we are
    283 	   tiled, assume it is this format (NV_MEM_KIND_C32_2CRA) otherwise linear (KIND_PITCH). */
    284 	if (gem_get_tiling.mode == DRM_TEGRA_GEM_TILING_MODE_PITCH) {
    285 		bo->tiling = NV_MEM_KIND_PITCH;
    286 	} else if (gem_get_tiling.mode == DRM_TEGRA_GEM_TILING_MODE_BLOCK) {
    287 		bo->tiling = NV_MEM_KIND_C32_2CRA;
    288 	} else {
    289 		drv_log("%s: unknown tile format %d\n", __func__, gem_get_tiling.mode);
    290 		drv_gem_bo_destroy(bo);
    291 		assert(0);
    292 	}
    293 
    294 	bo->format_modifiers[0] = fourcc_mod_code(NV, bo->tiling);
    295 	return 0;
    296 }
    297 
    298 static void *tegra_bo_map(struct bo *bo, struct vma *vma, size_t plane, uint32_t map_flags)
    299 {
    300 	int ret;
    301 	struct drm_tegra_gem_mmap gem_map;
    302 	struct tegra_private_map_data *priv;
    303 
    304 	memset(&gem_map, 0, sizeof(gem_map));
    305 	gem_map.handle = bo->handles[0].u32;
    306 
    307 	ret = drmCommandWriteRead(bo->drv->fd, DRM_TEGRA_GEM_MMAP, &gem_map, sizeof(gem_map));
    308 	if (ret < 0) {
    309 		drv_log("DRM_TEGRA_GEM_MMAP failed\n");
    310 		return MAP_FAILED;
    311 	}
    312 
    313 	void *addr = mmap(0, bo->total_size, drv_get_prot(map_flags), MAP_SHARED, bo->drv->fd,
    314 			  gem_map.offset);
    315 	vma->length = bo->total_size;
    316 	if ((bo->tiling & 0xFF) == NV_MEM_KIND_C32_2CRA && addr != MAP_FAILED) {
    317 		priv = calloc(1, sizeof(*priv));
    318 		priv->untiled = calloc(1, bo->total_size);
    319 		priv->tiled = addr;
    320 		vma->priv = priv;
    321 		transfer_tiled_memory(bo, priv->tiled, priv->untiled, TEGRA_READ_TILED_BUFFER);
    322 		addr = priv->untiled;
    323 	}
    324 
    325 	return addr;
    326 }
    327 
    328 static int tegra_bo_unmap(struct bo *bo, struct vma *vma)
    329 {
    330 	if (vma->priv) {
    331 		struct tegra_private_map_data *priv = vma->priv;
    332 		vma->addr = priv->tiled;
    333 		free(priv->untiled);
    334 		free(priv);
    335 		vma->priv = NULL;
    336 	}
    337 
    338 	return munmap(vma->addr, vma->length);
    339 }
    340 
    341 static int tegra_bo_flush(struct bo *bo, struct mapping *mapping)
    342 {
    343 	struct tegra_private_map_data *priv = mapping->vma->priv;
    344 
    345 	if (priv && (mapping->vma->map_flags & BO_MAP_WRITE))
    346 		transfer_tiled_memory(bo, priv->tiled, priv->untiled, TEGRA_WRITE_TILED_BUFFER);
    347 
    348 	return 0;
    349 }
    350 
    351 const struct backend backend_tegra = {
    352 	.name = "tegra",
    353 	.init = tegra_init,
    354 	.bo_create = tegra_bo_create,
    355 	.bo_destroy = drv_gem_bo_destroy,
    356 	.bo_import = tegra_bo_import,
    357 	.bo_map = tegra_bo_map,
    358 	.bo_unmap = tegra_bo_unmap,
    359 	.bo_flush = tegra_bo_flush,
    360 };
    361 
    362 #endif
    363