/* * Copyright © 2012 Intel Corporation * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. * * Authors: * Daniel Vetter * */ /** @file gem_set_tiling_vs_blt.c * * Testcase: Check for proper synchronization of tiling changes vs. tiled gpu * access * * The blitter on gen3 and earlier needs properly set up fences. Which also * means that for untiled blits we may not set up a fence before that blt has * finished. * * Current kernels have a bug there, but it's pretty hard to hit because you * need: * - a blt on an untiled object which is aligned correctly for tiling. * - a set_tiling to switch that object to tiling * - another blt without any intervening cpu access that uses this object. * * Testcase has been extended to also check tiled->untiled and tiled->tiled * transitions (i.e. changing stride). */ #include "igt.h" #include #include #include #include #include #include #include #include #include #include "drm.h" #include "intel_bufmgr.h" IGT_TEST_DESCRIPTION("Check for proper synchronization of tiling changes vs." " tiled gpu access."); static drm_intel_bufmgr *bufmgr; struct intel_batchbuffer *batch; uint32_t devid; #define TEST_SIZE (1024*1024) #define TEST_STRIDE (4*1024) #define TEST_HEIGHT(stride) (TEST_SIZE/(stride)) #define TEST_WIDTH(stride) ((stride)/4) uint32_t data[TEST_SIZE/4]; static void do_test(uint32_t tiling, unsigned stride, uint32_t tiling_after, unsigned stride_after) { drm_intel_bo *busy_bo, *test_bo, *target_bo; int i, ret; uint32_t *ptr; uint32_t test_bo_handle; uint32_t blt_stride, blt_bits; bool tiling_changed = false; igt_info("filling ring .. "); busy_bo = drm_intel_bo_alloc(bufmgr, "busy bo bo", 16*1024*1024, 4096); for (i = 0; i < 250; i++) { BLIT_COPY_BATCH_START(0); OUT_BATCH((3 << 24) | /* 32 bits */ (0xcc << 16) | /* copy ROP */ 2*1024*4); OUT_BATCH(0 << 16 | 1024); OUT_BATCH((2048) << 16 | (2048)); OUT_RELOC_FENCED(busy_bo, I915_GEM_DOMAIN_RENDER, I915_GEM_DOMAIN_RENDER, 0); OUT_BATCH(0 << 16 | 0); OUT_BATCH(2*1024*4); OUT_RELOC_FENCED(busy_bo, I915_GEM_DOMAIN_RENDER, 0, 0); ADVANCE_BATCH(); if (batch->gen >= 6) { BEGIN_BATCH(3, 0); OUT_BATCH(XY_SETUP_CLIP_BLT_CMD); OUT_BATCH(0); OUT_BATCH(0); ADVANCE_BATCH(); } } intel_batchbuffer_flush(batch); igt_info("playing tricks .. "); /* first allocate the target so it gets out of the way of playing funky * tricks */ target_bo = drm_intel_bo_alloc(bufmgr, "target bo", TEST_SIZE, 4096); /* allocate buffer with parameters _after_ transition we want to check * and touch it, so that it's properly aligned in the gtt. */ test_bo = drm_intel_bo_alloc(bufmgr, "tiled busy bo", TEST_SIZE, 4096); test_bo_handle = test_bo->handle; ret = drm_intel_bo_set_tiling(test_bo, &tiling_after, stride_after); igt_assert_eq(ret, 0); drm_intel_gem_bo_map_gtt(test_bo); ptr = test_bo->virtual; *ptr = 0; ptr = NULL; drm_intel_gem_bo_unmap_gtt(test_bo); drm_intel_bo_unreference(test_bo); test_bo = NULL; /* note we need a bo bigger than batches, otherwise the buffer reuse * trick will fail. */ test_bo = drm_intel_bo_alloc(bufmgr, "busy bo", TEST_SIZE, 4096); /* double check that the reuse trick worked */ igt_assert(test_bo_handle == test_bo->handle); test_bo_handle = test_bo->handle; /* ensure we have the right tiling before we start. */ ret = drm_intel_bo_set_tiling(test_bo, &tiling, stride); igt_assert_eq(ret, 0); if (tiling == I915_TILING_NONE) { drm_intel_bo_subdata(test_bo, 0, TEST_SIZE, data); } else { drm_intel_gem_bo_map_gtt(test_bo); ptr = test_bo->virtual; memcpy(ptr, data, TEST_SIZE); ptr = NULL; drm_intel_gem_bo_unmap_gtt(test_bo); } blt_stride = stride; blt_bits = 0; if (intel_gen(devid) >= 4 && tiling != I915_TILING_NONE) { blt_stride /= 4; blt_bits = XY_SRC_COPY_BLT_SRC_TILED; } BLIT_COPY_BATCH_START(blt_bits); OUT_BATCH((3 << 24) | /* 32 bits */ (0xcc << 16) | /* copy ROP */ stride); OUT_BATCH(0 << 16 | 0); OUT_BATCH((TEST_HEIGHT(stride)) << 16 | (TEST_WIDTH(stride))); OUT_RELOC_FENCED(target_bo, I915_GEM_DOMAIN_RENDER, I915_GEM_DOMAIN_RENDER, 0); OUT_BATCH(0 << 16 | 0); OUT_BATCH(blt_stride); OUT_RELOC_FENCED(test_bo, I915_GEM_DOMAIN_RENDER, 0, 0); ADVANCE_BATCH(); intel_batchbuffer_flush(batch); drm_intel_bo_unreference(test_bo); test_bo = drm_intel_bo_alloc_for_render(bufmgr, "tiled busy bo", TEST_SIZE, 4096); /* double check that the reuse trick worked */ igt_assert(test_bo_handle == test_bo->handle); ret = drm_intel_bo_set_tiling(test_bo, &tiling_after, stride_after); igt_assert_eq(ret, 0); /* Note: We don't care about gen4+ here because the blitter doesn't use * fences there. So not setting tiling flags on the tiled buffer is ok. */ BLIT_COPY_BATCH_START(0); OUT_BATCH((3 << 24) | /* 32 bits */ (0xcc << 16) | /* copy ROP */ stride_after); OUT_BATCH(0 << 16 | 0); OUT_BATCH((1) << 16 | (1)); OUT_RELOC_FENCED(test_bo, I915_GEM_DOMAIN_RENDER, I915_GEM_DOMAIN_RENDER, 0); OUT_BATCH(0 << 16 | 0); OUT_BATCH(stride_after); OUT_RELOC_FENCED(test_bo, I915_GEM_DOMAIN_RENDER, 0, 0); ADVANCE_BATCH(); intel_batchbuffer_flush(batch); /* Now try to trick the kernel the kernel into changing up the fencing * too early. */ igt_info("checking .. "); memset(data, 0, TEST_SIZE); drm_intel_bo_get_subdata(target_bo, 0, TEST_SIZE, data); for (i = 0; i < TEST_SIZE/4; i++) igt_assert(data[i] == i); /* check whether tiling on the test_bo actually changed. */ drm_intel_gem_bo_map_gtt(test_bo); ptr = test_bo->virtual; for (i = 0; i < TEST_SIZE/4; i++) if (ptr[i] != data[i]) tiling_changed = true; ptr = NULL; drm_intel_gem_bo_unmap_gtt(test_bo); igt_assert(tiling_changed); drm_intel_bo_unreference(test_bo); drm_intel_bo_unreference(target_bo); drm_intel_bo_unreference(busy_bo); igt_info("done\n"); } int fd; igt_main { int i; uint32_t tiling, tiling_after; igt_skip_on_simulation(); igt_fixture { for (i = 0; i < 1024*256; i++) data[i] = i; fd = drm_open_any(); bufmgr = drm_intel_bufmgr_gem_init(fd, 4096); drm_intel_bufmgr_gem_enable_reuse(bufmgr); devid = intel_get_drm_devid(fd); batch = intel_batchbuffer_alloc(bufmgr, devid); } igt_subtest("untiled-to-tiled") { tiling = I915_TILING_NONE; tiling_after = I915_TILING_X; do_test(tiling, TEST_STRIDE, tiling_after, TEST_STRIDE); igt_assert(tiling == I915_TILING_NONE); igt_assert(tiling_after == I915_TILING_X); } igt_subtest("tiled-to-untiled") { tiling = I915_TILING_X; tiling_after = I915_TILING_NONE; do_test(tiling, TEST_STRIDE, tiling_after, TEST_STRIDE); igt_assert(tiling == I915_TILING_X); igt_assert(tiling_after == I915_TILING_NONE); } igt_subtest("tiled-to-tiled") { tiling = I915_TILING_X; tiling_after = I915_TILING_X; do_test(tiling, TEST_STRIDE/2, tiling_after, TEST_STRIDE); igt_assert(tiling == I915_TILING_X); igt_assert(tiling_after == I915_TILING_X); } }