/* * Copyright © 2013 Intel Corporation * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. * */ #include #include #include #include #include #include "drmtest.h" #include "ioctl_wrappers.h" #include "intel_chipset.h" #ifndef I915_PARAM_CMD_PARSER_VERSION #define I915_PARAM_CMD_PARSER_VERSION 28 #endif static int __gem_execbuf(int fd, struct drm_i915_gem_execbuffer2 *execbuf) { if (drmIoctl(fd, DRM_IOCTL_I915_GEM_EXECBUFFER2, execbuf)) return -errno; return 0; } static void exec_batch_patched(int fd, uint32_t cmd_bo, uint32_t *cmds, int size, int patch_offset, uint64_t expected_value) { struct drm_i915_gem_execbuffer2 execbuf; struct drm_i915_gem_exec_object2 objs[2]; struct drm_i915_gem_relocation_entry reloc[1]; uint32_t target_bo = gem_create(fd, 4096); uint64_t actual_value = 0; gem_write(fd, cmd_bo, 0, cmds, size); reloc[0].offset = patch_offset; reloc[0].delta = 0; reloc[0].target_handle = target_bo; reloc[0].read_domains = I915_GEM_DOMAIN_RENDER; reloc[0].write_domain = I915_GEM_DOMAIN_RENDER; reloc[0].presumed_offset = 0; objs[0].handle = target_bo; objs[0].relocation_count = 0; objs[0].relocs_ptr = 0; objs[0].alignment = 0; objs[0].offset = 0; objs[0].flags = 0; objs[0].rsvd1 = 0; objs[0].rsvd2 = 0; objs[1].handle = cmd_bo; objs[1].relocation_count = 1; objs[1].relocs_ptr = (uintptr_t)reloc; objs[1].alignment = 0; objs[1].offset = 0; objs[1].flags = 0; objs[1].rsvd1 = 0; objs[1].rsvd2 = 0; execbuf.buffers_ptr = (uintptr_t)objs; execbuf.buffer_count = 2; execbuf.batch_start_offset = 0; execbuf.batch_len = size; execbuf.cliprects_ptr = 0; execbuf.num_cliprects = 0; execbuf.DR1 = 0; execbuf.DR4 = 0; execbuf.flags = I915_EXEC_RENDER; i915_execbuffer2_set_context_id(execbuf, 0); execbuf.rsvd2 = 0; gem_execbuf(fd, &execbuf); gem_sync(fd, cmd_bo); gem_read(fd,target_bo, 0, &actual_value, sizeof(actual_value)); igt_assert_eq(expected_value, actual_value); gem_close(fd, target_bo); } static void exec_batch(int fd, uint32_t cmd_bo, uint32_t *cmds, int size, int ring, int expected_ret) { struct drm_i915_gem_execbuffer2 execbuf; struct drm_i915_gem_exec_object2 objs[1]; gem_write(fd, cmd_bo, 0, cmds, size); objs[0].handle = cmd_bo; objs[0].relocation_count = 0; objs[0].relocs_ptr = 0; objs[0].alignment = 0; objs[0].offset = 0; objs[0].flags = 0; objs[0].rsvd1 = 0; objs[0].rsvd2 = 0; execbuf.buffers_ptr = (uintptr_t)objs; execbuf.buffer_count = 1; execbuf.batch_start_offset = 0; execbuf.batch_len = size; execbuf.cliprects_ptr = 0; execbuf.num_cliprects = 0; execbuf.DR1 = 0; execbuf.DR4 = 0; execbuf.flags = ring; i915_execbuffer2_set_context_id(execbuf, 0); execbuf.rsvd2 = 0; igt_assert_eq(__gem_execbuf(fd, &execbuf), expected_ret); gem_sync(fd, cmd_bo); } static void exec_split_batch(int fd, uint32_t *cmds, int size, int ring, int expected_ret) { struct drm_i915_gem_execbuffer2 execbuf; struct drm_i915_gem_exec_object2 objs[1]; uint32_t cmd_bo; uint32_t noop[1024] = { 0 }; const int alloc_size = 4096 * 2; const int actual_start_offset = 4096-sizeof(uint32_t); // Allocate and fill a 2-page batch with noops cmd_bo = gem_create(fd, alloc_size); gem_write(fd, cmd_bo, 0, noop, sizeof(noop)); gem_write(fd, cmd_bo, 4096, noop, sizeof(noop)); // Write the provided commands such that the first dword // of the command buffer is the last dword of the first // page (i.e. the command is split across the two pages). gem_write(fd, cmd_bo, actual_start_offset, cmds, size); objs[0].handle = cmd_bo; objs[0].relocation_count = 0; objs[0].relocs_ptr = 0; objs[0].alignment = 0; objs[0].offset = 0; objs[0].flags = 0; objs[0].rsvd1 = 0; objs[0].rsvd2 = 0; execbuf.buffers_ptr = (uintptr_t)objs; execbuf.buffer_count = 1; // NB: We want batch_start_offset and batch_len to point to the block // of the actual commands (i.e. at the last dword of the first page), // but have to adjust both the start offset and length to meet the // kernel driver's requirements on the alignment of those fields. execbuf.batch_start_offset = actual_start_offset & ~0x7; execbuf.batch_len = ALIGN(size + actual_start_offset - execbuf.batch_start_offset, 0x8); execbuf.cliprects_ptr = 0; execbuf.num_cliprects = 0; execbuf.DR1 = 0; execbuf.DR4 = 0; execbuf.flags = ring; i915_execbuffer2_set_context_id(execbuf, 0); execbuf.rsvd2 = 0; igt_assert_eq(__gem_execbuf(fd, &execbuf), expected_ret); gem_sync(fd, cmd_bo); gem_close(fd, cmd_bo); } static void exec_batch_chained(int fd, uint32_t cmd_bo, uint32_t *cmds, int size, int patch_offset, uint64_t expected_value) { struct drm_i915_gem_execbuffer2 execbuf; struct drm_i915_gem_exec_object2 objs[3]; struct drm_i915_gem_relocation_entry reloc; struct drm_i915_gem_relocation_entry first_level_reloc; uint32_t target_bo = gem_create(fd, 4096); uint32_t first_level_bo = gem_create(fd, 4096); uint64_t actual_value = 0; static uint32_t first_level_cmds[] = { MI_BATCH_BUFFER_START | MI_BATCH_NON_SECURE_I965, 0, MI_BATCH_BUFFER_END, 0, }; if (IS_HASWELL(intel_get_drm_devid(fd))) first_level_cmds[0] |= MI_BATCH_NON_SECURE_HSW; gem_write(fd, first_level_bo, 0, first_level_cmds, sizeof(first_level_cmds)); gem_write(fd, cmd_bo, 0, cmds, size); reloc.offset = patch_offset; reloc.delta = 0; reloc.target_handle = target_bo; reloc.read_domains = I915_GEM_DOMAIN_RENDER; reloc.write_domain = I915_GEM_DOMAIN_RENDER; reloc.presumed_offset = 0; first_level_reloc.offset = 4; first_level_reloc.delta = 0; first_level_reloc.target_handle = cmd_bo; first_level_reloc.read_domains = I915_GEM_DOMAIN_INSTRUCTION; first_level_reloc.write_domain = 0; first_level_reloc.presumed_offset = 0; objs[0].handle = target_bo; objs[0].relocation_count = 0; objs[0].relocs_ptr = 0; objs[0].alignment = 0; objs[0].offset = 0; objs[0].flags = 0; objs[0].rsvd1 = 0; objs[0].rsvd2 = 0; objs[1].handle = cmd_bo; objs[1].relocation_count = 1; objs[1].relocs_ptr = (uintptr_t)&reloc; objs[1].alignment = 0; objs[1].offset = 0; objs[1].flags = 0; objs[1].rsvd1 = 0; objs[1].rsvd2 = 0; objs[2].handle = first_level_bo; objs[2].relocation_count = 1; objs[2].relocs_ptr = (uintptr_t)&first_level_reloc; objs[2].alignment = 0; objs[2].offset = 0; objs[2].flags = 0; objs[2].rsvd1 = 0; objs[2].rsvd2 = 0; execbuf.buffers_ptr = (uintptr_t)objs; execbuf.buffer_count = 3; execbuf.batch_start_offset = 0; execbuf.batch_len = sizeof(first_level_cmds); execbuf.cliprects_ptr = 0; execbuf.num_cliprects = 0; execbuf.DR1 = 0; execbuf.DR4 = 0; execbuf.flags = I915_EXEC_RENDER; i915_execbuffer2_set_context_id(execbuf, 0); execbuf.rsvd2 = 0; gem_execbuf(fd, &execbuf); gem_sync(fd, cmd_bo); gem_read(fd,target_bo, 0, &actual_value, sizeof(actual_value)); igt_assert_eq(expected_value, actual_value); gem_close(fd, first_level_bo); gem_close(fd, target_bo); } uint32_t handle; int fd; #define MI_ARB_ON_OFF (0x8 << 23) #define MI_DISPLAY_FLIP ((0x14 << 23) | 1) #define GFX_OP_PIPE_CONTROL ((0x3<<29)|(0x3<<27)|(0x2<<24)|2) #define PIPE_CONTROL_QW_WRITE (1<<14) #define PIPE_CONTROL_LRI_POST_OP (1<<23) #define OACONTROL 0x2360 igt_main { igt_fixture { int parser_version = 0; drm_i915_getparam_t gp; int rc; fd = drm_open_any(); gp.param = I915_PARAM_CMD_PARSER_VERSION; gp.value = &parser_version; rc = drmIoctl(fd, DRM_IOCTL_I915_GETPARAM, &gp); igt_require(!rc && parser_version > 0); igt_require(gem_uses_aliasing_ppgtt(fd)); handle = gem_create(fd, 4096); /* ATM cmd parser only exists on gen7. */ igt_require(intel_gen(intel_get_drm_devid(fd)) == 7); } igt_subtest("basic-allowed") { uint32_t pc[] = { GFX_OP_PIPE_CONTROL, PIPE_CONTROL_QW_WRITE, 0, // To be patched 0x12000000, 0, MI_BATCH_BUFFER_END, }; exec_batch_patched(fd, handle, pc, sizeof(pc), 8, // patch offset, 0x12000000); } igt_subtest("basic-rejected") { uint32_t arb_on_off[] = { MI_ARB_ON_OFF, MI_BATCH_BUFFER_END, }; uint32_t display_flip[] = { MI_DISPLAY_FLIP, 0, 0, 0, MI_BATCH_BUFFER_END, 0 }; exec_batch(fd, handle, arb_on_off, sizeof(arb_on_off), I915_EXEC_RENDER, -EINVAL); exec_batch(fd, handle, arb_on_off, sizeof(arb_on_off), I915_EXEC_BSD, -EINVAL); if (gem_has_vebox(fd)) { exec_batch(fd, handle, arb_on_off, sizeof(arb_on_off), I915_EXEC_VEBOX, -EINVAL); } exec_batch(fd, handle, display_flip, sizeof(display_flip), I915_EXEC_BLT, -EINVAL); } igt_subtest("registers") { uint32_t lri_bad[] = { MI_LOAD_REGISTER_IMM, 0, // disallowed register address 0x12000000, MI_BATCH_BUFFER_END, }; uint32_t lri_ok[] = { MI_LOAD_REGISTER_IMM, 0x5280, // allowed register address (SO_WRITE_OFFSET[0]) 0x1, MI_BATCH_BUFFER_END, }; exec_batch(fd, handle, lri_bad, sizeof(lri_bad), I915_EXEC_RENDER, -EINVAL); exec_batch(fd, handle, lri_ok, sizeof(lri_ok), I915_EXEC_RENDER, 0); } igt_subtest("bitmasks") { uint32_t pc[] = { GFX_OP_PIPE_CONTROL, (PIPE_CONTROL_QW_WRITE | PIPE_CONTROL_LRI_POST_OP), 0, // To be patched 0x12000000, 0, MI_BATCH_BUFFER_END, }; exec_batch(fd, handle, pc, sizeof(pc), I915_EXEC_RENDER, -EINVAL); } igt_subtest("batch-without-end") { uint32_t noop[1024] = { 0 }; exec_batch(fd, handle, noop, sizeof(noop), I915_EXEC_RENDER, -EINVAL); } igt_subtest("cmd-crossing-page") { uint32_t lri_ok[] = { MI_LOAD_REGISTER_IMM, 0x5280, // allowed register address (SO_WRITE_OFFSET[0]) 0x1, MI_BATCH_BUFFER_END, }; exec_split_batch(fd, lri_ok, sizeof(lri_ok), I915_EXEC_RENDER, 0); } igt_subtest("oacontrol-tracking") { uint32_t lri_ok[] = { MI_LOAD_REGISTER_IMM, OACONTROL, 0x31337000, MI_LOAD_REGISTER_IMM, OACONTROL, 0x0, MI_BATCH_BUFFER_END, 0 }; uint32_t lri_bad[] = { MI_LOAD_REGISTER_IMM, OACONTROL, 0x31337000, MI_BATCH_BUFFER_END, }; uint32_t lri_extra_bad[] = { MI_LOAD_REGISTER_IMM, OACONTROL, 0x31337000, MI_LOAD_REGISTER_IMM, OACONTROL, 0x0, MI_LOAD_REGISTER_IMM, OACONTROL, 0x31337000, MI_BATCH_BUFFER_END, }; exec_batch(fd, handle, lri_ok, sizeof(lri_ok), I915_EXEC_RENDER, 0); exec_batch(fd, handle, lri_bad, sizeof(lri_bad), I915_EXEC_RENDER, -EINVAL); exec_batch(fd, handle, lri_extra_bad, sizeof(lri_extra_bad), I915_EXEC_RENDER, -EINVAL); } igt_subtest("chained-batch") { uint32_t pc[] = { GFX_OP_PIPE_CONTROL, PIPE_CONTROL_QW_WRITE, 0, // To be patched 0x12000000, 0, MI_BATCH_BUFFER_END, }; exec_batch_chained(fd, handle, pc, sizeof(pc), 8, // patch offset, 0x12000000); } igt_fixture { gem_close(fd, handle); close(fd); } }