summaryrefslogtreecommitdiff
path: root/tests/i915/gem_exec_reloc.c
diff options
context:
space:
mode:
Diffstat (limited to 'tests/i915/gem_exec_reloc.c')
-rw-r--r--tests/i915/gem_exec_reloc.c761
1 files changed, 761 insertions, 0 deletions
diff --git a/tests/i915/gem_exec_reloc.c b/tests/i915/gem_exec_reloc.c
new file mode 100644
index 00000000..837f60a6
--- /dev/null
+++ b/tests/i915/gem_exec_reloc.c
@@ -0,0 +1,761 @@
+/*
+ * Copyright © 2016 Intel Corporation
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * the rights to use, copy, modify, merge, publish, distribute, sublicense,
+ * and/or sell copies of the Software, and to permit persons to whom the
+ * Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice (including the next
+ * paragraph) shall be included in all copies or substantial portions of the
+ * Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
+ * IN THE SOFTWARE.
+ */
+
+#include "igt.h"
+#include "igt_dummyload.h"
+
+IGT_TEST_DESCRIPTION("Basic sanity check of execbuf-ioctl relocations.");
+
+#define LOCAL_I915_EXEC_BSD_SHIFT (13)
+#define LOCAL_I915_EXEC_BSD_MASK (3 << LOCAL_I915_EXEC_BSD_SHIFT)
+
+#define LOCAL_I915_EXEC_NO_RELOC (1<<11)
+#define LOCAL_I915_EXEC_HANDLE_LUT (1<<12)
+
+#define ENGINE_MASK (I915_EXEC_RING_MASK | LOCAL_I915_EXEC_BSD_MASK)
+
+static uint32_t find_last_set(uint64_t x)
+{
+ uint32_t i = 0;
+ while (x) {
+ x >>= 1;
+ i++;
+ }
+ return i;
+}
+
+static void write_dword(int fd,
+ uint32_t target_handle,
+ uint64_t target_offset,
+ uint32_t value)
+{
+ int gen = intel_gen(intel_get_drm_devid(fd));
+ struct drm_i915_gem_execbuffer2 execbuf;
+ struct drm_i915_gem_exec_object2 obj[2];
+ struct drm_i915_gem_relocation_entry reloc;
+ uint32_t buf[16];
+ int i;
+
+ memset(obj, 0, sizeof(obj));
+ obj[0].handle = target_handle;
+ obj[1].handle = gem_create(fd, 4096);
+
+ i = 0;
+ buf[i++] = MI_STORE_DWORD_IMM | (gen < 6 ? 1<<22 : 0);
+ if (gen >= 8) {
+ buf[i++] = target_offset;
+ buf[i++] = target_offset >> 32;
+ } else if (gen >= 4) {
+ buf[i++] = 0;
+ buf[i++] = target_offset;
+ } else {
+ buf[i-1]--;
+ buf[i++] = target_offset;
+ }
+ buf[i++] = value;
+ buf[i++] = MI_BATCH_BUFFER_END;
+ gem_write(fd, obj[1].handle, 0, buf, sizeof(buf));
+
+ memset(&reloc, 0, sizeof(reloc));
+ if (gen >= 8 || gen < 4)
+ reloc.offset = sizeof(uint32_t);
+ else
+ reloc.offset = 2*sizeof(uint32_t);
+ reloc.target_handle = target_handle;
+ reloc.delta = target_offset;
+ reloc.read_domains = I915_GEM_DOMAIN_INSTRUCTION;
+ reloc.write_domain = I915_GEM_DOMAIN_INSTRUCTION;
+
+ obj[1].relocation_count = 1;
+ obj[1].relocs_ptr = to_user_pointer(&reloc);
+
+ memset(&execbuf, 0, sizeof(execbuf));
+ execbuf.buffers_ptr = to_user_pointer(obj);
+ execbuf.buffer_count = 2;
+ execbuf.flags = I915_EXEC_SECURE;
+ gem_execbuf(fd, &execbuf);
+ gem_close(fd, obj[1].handle);
+}
+
+enum mode { MEM, CPU, WC, GTT };
+#define RO 0x100
+static void from_mmap(int fd, uint64_t size, enum mode mode)
+{
+ uint32_t bbe = MI_BATCH_BUFFER_END;
+ struct drm_i915_gem_execbuffer2 execbuf;
+ struct drm_i915_gem_exec_object2 obj;
+ struct drm_i915_gem_relocation_entry *relocs;
+ uint32_t reloc_handle;
+ uint64_t value;
+ uint64_t max, i;
+ int retry = 2;
+
+ /* Worst case is that the kernel has to copy the entire incoming
+ * reloc[], so double the memory requirements.
+ */
+ intel_require_memory(2, size, CHECK_RAM);
+
+ memset(&obj, 0, sizeof(obj));
+ obj.handle = gem_create(fd, 4096);
+ gem_write(fd, obj.handle, 0, &bbe, sizeof(bbe));
+
+ max = size / sizeof(*relocs);
+ switch (mode & ~RO) {
+ case MEM:
+ relocs = mmap(0, size,
+ PROT_WRITE, MAP_PRIVATE | MAP_ANON,
+ -1, 0);
+ igt_assert(relocs != (void *)-1);
+ break;
+ case GTT:
+ reloc_handle = gem_create(fd, size);
+ relocs = gem_mmap__gtt(fd, reloc_handle, size, PROT_WRITE);
+ gem_set_domain(fd, reloc_handle,
+ I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT);
+ gem_close(fd, reloc_handle);
+ break;
+ case CPU:
+ reloc_handle = gem_create(fd, size);
+ relocs = gem_mmap__cpu(fd, reloc_handle, 0, size, PROT_WRITE);
+ gem_set_domain(fd, reloc_handle,
+ I915_GEM_DOMAIN_CPU, I915_GEM_DOMAIN_CPU);
+ gem_close(fd, reloc_handle);
+ break;
+ case WC:
+ reloc_handle = gem_create(fd, size);
+ relocs = gem_mmap__wc(fd, reloc_handle, 0, size, PROT_WRITE);
+ gem_set_domain(fd, reloc_handle,
+ I915_GEM_DOMAIN_WC, I915_GEM_DOMAIN_WC);
+ gem_close(fd, reloc_handle);
+ break;
+ }
+
+ for (i = 0; i < max; i++) {
+ relocs[i].target_handle = obj.handle;
+ relocs[i].presumed_offset = ~0ull;
+ relocs[i].offset = 1024;
+ relocs[i].delta = i;
+ relocs[i].read_domains = I915_GEM_DOMAIN_INSTRUCTION;
+ relocs[i].write_domain = 0;
+ }
+ obj.relocation_count = max;
+ obj.relocs_ptr = to_user_pointer(relocs);
+
+ if (mode & RO)
+ mprotect(relocs, size, PROT_READ);
+
+ memset(&execbuf, 0, sizeof(execbuf));
+ execbuf.buffers_ptr = to_user_pointer(&obj);
+ execbuf.buffer_count = 1;
+ while (relocs[0].presumed_offset == ~0ull && retry--)
+ gem_execbuf(fd, &execbuf);
+ gem_read(fd, obj.handle, 1024, &value, sizeof(value));
+ gem_close(fd, obj.handle);
+
+ igt_assert_eq_u64(value, obj.offset + max - 1);
+ if (relocs[0].presumed_offset != ~0ull) {
+ for (i = 0; i < max; i++)
+ igt_assert_eq_u64(relocs[i].presumed_offset,
+ obj.offset);
+ }
+ munmap(relocs, size);
+}
+
+static void from_gpu(int fd)
+{
+ uint32_t bbe = MI_BATCH_BUFFER_END;
+ struct drm_i915_gem_execbuffer2 execbuf;
+ struct drm_i915_gem_exec_object2 obj;
+ struct drm_i915_gem_relocation_entry *relocs;
+ uint32_t reloc_handle;
+ uint64_t value;
+
+ igt_require(gem_can_store_dword(fd, 0));
+
+ memset(&obj, 0, sizeof(obj));
+ obj.handle = gem_create(fd, 4096);
+ gem_write(fd, obj.handle, 0, &bbe, sizeof(bbe));
+
+ reloc_handle = gem_create(fd, 4096);
+ write_dword(fd,
+ reloc_handle,
+ offsetof(struct drm_i915_gem_relocation_entry,
+ target_handle),
+ obj.handle);
+ write_dword(fd,
+ reloc_handle,
+ offsetof(struct drm_i915_gem_relocation_entry,
+ offset),
+ 1024);
+ write_dword(fd,
+ reloc_handle,
+ offsetof(struct drm_i915_gem_relocation_entry,
+ read_domains),
+ I915_GEM_DOMAIN_INSTRUCTION);
+
+ relocs = gem_mmap__cpu(fd, reloc_handle, 0, 4096, PROT_READ);
+ gem_set_domain(fd, reloc_handle,
+ I915_GEM_DOMAIN_CPU, I915_GEM_DOMAIN_CPU);
+ gem_close(fd, reloc_handle);
+
+ obj.relocation_count = 1;
+ obj.relocs_ptr = to_user_pointer(relocs);
+
+ memset(&execbuf, 0, sizeof(execbuf));
+ execbuf.buffers_ptr = to_user_pointer(&obj);
+ execbuf.buffer_count = 1;
+ gem_execbuf(fd, &execbuf);
+ gem_read(fd, obj.handle, 1024, &value, sizeof(value));
+ gem_close(fd, obj.handle);
+
+ igt_assert_eq_u64(value, obj.offset);
+ igt_assert_eq_u64(relocs->presumed_offset, obj.offset);
+ munmap(relocs, 4096);
+}
+
+static void check_bo(int fd, uint32_t handle)
+{
+ uint32_t *map;
+ int i;
+
+ igt_debug("Verifying result\n");
+ map = gem_mmap__cpu(fd, handle, 0, 4096, PROT_READ);
+ gem_set_domain(fd, handle, I915_GEM_DOMAIN_CPU, 0);
+ for (i = 0; i < 1024; i++)
+ igt_assert_eq(map[i], i);
+ munmap(map, 4096);
+}
+
+static void active(int fd, unsigned engine)
+{
+ const int gen = intel_gen(intel_get_drm_devid(fd));
+ struct drm_i915_gem_relocation_entry reloc;
+ struct drm_i915_gem_exec_object2 obj[2];
+ struct drm_i915_gem_execbuffer2 execbuf;
+ unsigned engines[16];
+ unsigned nengine;
+ int pass;
+
+ nengine = 0;
+ if (engine == ALL_ENGINES) {
+ for_each_physical_engine(fd, engine) {
+ if (gem_can_store_dword(fd, engine))
+ engines[nengine++] = engine;
+ }
+ } else {
+ igt_require(gem_has_ring(fd, engine));
+ igt_require(gem_can_store_dword(fd, engine));
+ engines[nengine++] = engine;
+ }
+ igt_require(nengine);
+
+ memset(obj, 0, sizeof(obj));
+ obj[0].handle = gem_create(fd, 4096);
+ obj[1].handle = gem_create(fd, 64*1024);
+ obj[1].relocs_ptr = to_user_pointer(&reloc);
+ obj[1].relocation_count = 1;
+
+ memset(&reloc, 0, sizeof(reloc));
+ reloc.offset = sizeof(uint32_t);
+ reloc.target_handle = obj[0].handle;
+ if (gen < 8 && gen >= 4)
+ reloc.offset += sizeof(uint32_t);
+ reloc.read_domains = I915_GEM_DOMAIN_INSTRUCTION;
+ reloc.write_domain = I915_GEM_DOMAIN_INSTRUCTION;
+
+ memset(&execbuf, 0, sizeof(execbuf));
+ execbuf.buffers_ptr = to_user_pointer(obj);
+ execbuf.buffer_count = 2;
+ if (gen < 6)
+ execbuf.flags |= I915_EXEC_SECURE;
+
+ for (pass = 0; pass < 1024; pass++) {
+ uint32_t batch[16];
+ int i = 0;
+ batch[i] = MI_STORE_DWORD_IMM | (gen < 6 ? 1 << 22 : 0);
+ if (gen >= 8) {
+ batch[++i] = 0;
+ batch[++i] = 0;
+ } else if (gen >= 4) {
+ batch[++i] = 0;
+ batch[++i] = 0;
+ } else {
+ batch[i]--;
+ batch[++i] = 0;
+ }
+ batch[++i] = pass;
+ batch[++i] = MI_BATCH_BUFFER_END;
+ gem_write(fd, obj[1].handle, pass*sizeof(batch),
+ batch, sizeof(batch));
+ }
+
+ for (pass = 0; pass < 1024; pass++) {
+ reloc.delta = 4*pass;
+ reloc.presumed_offset = -1;
+ execbuf.flags &= ~ENGINE_MASK;
+ execbuf.flags |= engines[rand() % nengine];
+ gem_execbuf(fd, &execbuf);
+ execbuf.batch_start_offset += 64;
+ reloc.offset += 64;
+ }
+ gem_close(fd, obj[1].handle);
+
+ check_bo(fd, obj[0].handle);
+ gem_close(fd, obj[0].handle);
+}
+
+static bool has_64b_reloc(int fd)
+{
+ return intel_gen(intel_get_drm_devid(fd)) >= 8;
+}
+
+#define NORELOC 1
+#define ACTIVE 2
+#define HANG 4
+static void basic_reloc(int fd, unsigned before, unsigned after, unsigned flags)
+{
+#define OBJSZ 8192
+ struct drm_i915_gem_relocation_entry reloc;
+ struct drm_i915_gem_exec_object2 obj;
+ struct drm_i915_gem_execbuffer2 execbuf;
+ uint64_t address_mask = has_64b_reloc(fd) ? ~(uint64_t)0 : ~(uint32_t)0;
+ const uint32_t bbe = MI_BATCH_BUFFER_END;
+ unsigned int reloc_offset;
+
+ memset(&obj, 0, sizeof(obj));
+ obj.handle = gem_create(fd, OBJSZ);
+ obj.relocs_ptr = to_user_pointer(&reloc);
+ obj.relocation_count = 1;
+ gem_write(fd, obj.handle, 0, &bbe, sizeof(bbe));
+
+ memset(&execbuf, 0, sizeof(execbuf));
+ execbuf.buffers_ptr = to_user_pointer(&obj);
+ execbuf.buffer_count = 1;
+ if (flags & NORELOC)
+ execbuf.flags |= LOCAL_I915_EXEC_NO_RELOC;
+
+ for (reloc_offset = 4096 - 8; reloc_offset <= 4096 + 8; reloc_offset += 4) {
+ igt_spin_t *spin = NULL;
+ uint32_t trash = 0;
+ uint64_t offset;
+
+ obj.offset = -1;
+
+ memset(&reloc, 0, sizeof(reloc));
+ reloc.offset = reloc_offset;
+ reloc.target_handle = obj.handle;
+ reloc.read_domains = I915_GEM_DOMAIN_INSTRUCTION;
+ reloc.presumed_offset = -1;
+
+ if (before) {
+ char *wc;
+
+ if (before == I915_GEM_DOMAIN_CPU)
+ wc = gem_mmap__cpu(fd, obj.handle, 0, OBJSZ, PROT_WRITE);
+ else if (before == I915_GEM_DOMAIN_GTT)
+ wc = gem_mmap__gtt(fd, obj.handle, OBJSZ, PROT_WRITE);
+ else if (before == I915_GEM_DOMAIN_WC)
+ wc = gem_mmap__wc(fd, obj.handle, 0, OBJSZ, PROT_WRITE);
+ else
+ igt_assert(0);
+ gem_set_domain(fd, obj.handle, before, before);
+ offset = -1;
+ memcpy(wc + reloc_offset, &offset, sizeof(offset));
+ munmap(wc, OBJSZ);
+ } else {
+ offset = -1;
+ gem_write(fd, obj.handle, reloc_offset, &offset, sizeof(offset));
+ }
+
+ if (flags & ACTIVE) {
+ spin = igt_spin_batch_new(fd,
+ .engine = I915_EXEC_DEFAULT,
+ .dependency = obj.handle);
+ if (!(flags & HANG))
+ igt_spin_batch_set_timeout(spin, NSEC_PER_SEC/100);
+ igt_assert(gem_bo_busy(fd, obj.handle));
+ }
+
+ gem_execbuf(fd, &execbuf);
+
+ if (after) {
+ char *wc;
+
+ if (after == I915_GEM_DOMAIN_CPU)
+ wc = gem_mmap__cpu(fd, obj.handle, 0, OBJSZ, PROT_READ);
+ else if (after == I915_GEM_DOMAIN_GTT)
+ wc = gem_mmap__gtt(fd, obj.handle, OBJSZ, PROT_READ);
+ else if (after == I915_GEM_DOMAIN_WC)
+ wc = gem_mmap__wc(fd, obj.handle, 0, OBJSZ, PROT_READ);
+ else
+ igt_assert(0);
+ gem_set_domain(fd, obj.handle, after, 0);
+ offset = ~reloc.presumed_offset & address_mask;
+ memcpy(&offset, wc + reloc_offset, has_64b_reloc(fd) ? 8 : 4);
+ munmap(wc, OBJSZ);
+ } else {
+ offset = ~reloc.presumed_offset & address_mask;
+ gem_read(fd, obj.handle, reloc_offset, &offset, has_64b_reloc(fd) ? 8 : 4);
+ }
+
+ if (reloc.presumed_offset == -1)
+ igt_warn("reloc.presumed_offset == -1\n");
+ else
+ igt_assert_eq_u64(reloc.presumed_offset, offset);
+ igt_assert_eq_u64(obj.offset, offset);
+
+ igt_spin_batch_free(fd, spin);
+
+ /* Simulate relocation */
+ if (flags & NORELOC) {
+ obj.offset += OBJSZ;
+ reloc.presumed_offset += OBJSZ;
+ } else {
+ trash = obj.handle;
+ obj.handle = gem_create(fd, OBJSZ);
+ gem_write(fd, obj.handle, 0, &bbe, sizeof(bbe));
+ reloc.target_handle = obj.handle;
+ }
+
+ if (before) {
+ char *wc;
+
+ if (before == I915_GEM_DOMAIN_CPU)
+ wc = gem_mmap__cpu(fd, obj.handle, 0, OBJSZ, PROT_WRITE);
+ else if (before == I915_GEM_DOMAIN_GTT)
+ wc = gem_mmap__gtt(fd, obj.handle, OBJSZ, PROT_WRITE);
+ else if (before == I915_GEM_DOMAIN_WC)
+ wc = gem_mmap__wc(fd, obj.handle, 0, OBJSZ, PROT_WRITE);
+ else
+ igt_assert(0);
+ gem_set_domain(fd, obj.handle, before, before);
+ memcpy(wc + reloc_offset, &reloc.presumed_offset, sizeof(reloc.presumed_offset));
+ munmap(wc, OBJSZ);
+ } else {
+ gem_write(fd, obj.handle, reloc_offset, &reloc.presumed_offset, sizeof(reloc.presumed_offset));
+ }
+
+ if (flags & ACTIVE) {
+ spin = igt_spin_batch_new(fd,
+ .engine = I915_EXEC_DEFAULT,
+ .dependency = obj.handle);
+ if (!(flags & HANG))
+ igt_spin_batch_set_timeout(spin, NSEC_PER_SEC/100);
+ igt_assert(gem_bo_busy(fd, obj.handle));
+ }
+
+ gem_execbuf(fd, &execbuf);
+
+ if (after) {
+ char *wc;
+
+ if (after == I915_GEM_DOMAIN_CPU)
+ wc = gem_mmap__cpu(fd, obj.handle, 0, OBJSZ, PROT_READ);
+ else if (after == I915_GEM_DOMAIN_GTT)
+ wc = gem_mmap__gtt(fd, obj.handle, OBJSZ, PROT_READ);
+ else if (after == I915_GEM_DOMAIN_WC)
+ wc = gem_mmap__wc(fd, obj.handle, 0, OBJSZ, PROT_READ);
+ else
+ igt_assert(0);
+ gem_set_domain(fd, obj.handle, after, 0);
+ offset = ~reloc.presumed_offset & address_mask;
+ memcpy(&offset, wc + reloc_offset, has_64b_reloc(fd) ? 8 : 4);
+ munmap(wc, OBJSZ);
+ } else {
+ offset = ~reloc.presumed_offset & address_mask;
+ gem_read(fd, obj.handle, reloc_offset, &offset, has_64b_reloc(fd) ? 8 : 4);
+ }
+
+ if (reloc.presumed_offset == -1)
+ igt_warn("reloc.presumed_offset == -1\n");
+ else
+ igt_assert_eq_u64(reloc.presumed_offset, offset);
+ igt_assert_eq_u64(obj.offset, offset);
+
+ igt_spin_batch_free(fd, spin);
+ if (trash)
+ gem_close(fd, trash);
+ }
+
+ gem_close(fd, obj.handle);
+}
+
+static inline uint64_t sign_extend(uint64_t x, int index)
+{
+ int shift = 63 - index;
+ return (int64_t)(x << shift) >> shift;
+}
+
+static uint64_t gen8_canonical_address(uint64_t address)
+{
+ return sign_extend(address, 47);
+}
+
+static void basic_range(int fd, unsigned flags)
+{
+ struct drm_i915_gem_relocation_entry reloc[128];
+ struct drm_i915_gem_exec_object2 obj[128];
+ struct drm_i915_gem_execbuffer2 execbuf;
+ uint64_t address_mask = has_64b_reloc(fd) ? ~(uint64_t)0 : ~(uint32_t)0;
+ uint64_t gtt_size = gem_aperture_size(fd);
+ const uint32_t bbe = MI_BATCH_BUFFER_END;
+ igt_spin_t *spin = NULL;
+ int count, n;
+
+ igt_require(gem_has_softpin(fd));
+
+ for (count = 12; gtt_size >> (count + 1); count++)
+ ;
+
+ count -= 12;
+
+ memset(obj, 0, sizeof(obj));
+ memset(reloc, 0, sizeof(reloc));
+ memset(&execbuf, 0, sizeof(execbuf));
+
+ n = 0;
+ for (int i = 0; i <= count; i++) {
+ obj[n].handle = gem_create(fd, 4096);
+ obj[n].offset = (1ull << (i + 12)) - 4096;
+ obj[n].offset = gen8_canonical_address(obj[n].offset);
+ obj[n].flags = EXEC_OBJECT_PINNED | EXEC_OBJECT_SUPPORTS_48B_ADDRESS;
+ gem_write(fd, obj[n].handle, 0, &bbe, sizeof(bbe));
+ execbuf.buffers_ptr = to_user_pointer(&obj[n]);
+ execbuf.buffer_count = 1;
+ if (__gem_execbuf(fd, &execbuf))
+ continue;
+
+ igt_debug("obj[%d] handle=%d, address=%llx\n",
+ n, obj[n].handle, (long long)obj[n].offset);
+
+ reloc[n].offset = 8 * (n + 1);
+ reloc[n].target_handle = obj[n].handle;
+ reloc[n].read_domains = I915_GEM_DOMAIN_INSTRUCTION;
+ reloc[n].presumed_offset = -1;
+ n++;
+ }
+ for (int i = 1; i < count; i++) {
+ obj[n].handle = gem_create(fd, 4096);
+ obj[n].offset = 1ull << (i + 12);
+ obj[n].offset = gen8_canonical_address(obj[n].offset);
+ obj[n].flags = EXEC_OBJECT_PINNED | EXEC_OBJECT_SUPPORTS_48B_ADDRESS;
+ gem_write(fd, obj[n].handle, 0, &bbe, sizeof(bbe));
+ execbuf.buffers_ptr = to_user_pointer(&obj[n]);
+ execbuf.buffer_count = 1;
+ if (__gem_execbuf(fd, &execbuf))
+ continue;
+
+ igt_debug("obj[%d] handle=%d, address=%llx\n",
+ n, obj[n].handle, (long long)obj[n].offset);
+
+ reloc[n].offset = 8 * (n + 1);
+ reloc[n].target_handle = obj[n].handle;
+ reloc[n].read_domains = I915_GEM_DOMAIN_INSTRUCTION;
+ reloc[n].presumed_offset = -1;
+ n++;
+ }
+ igt_require(n);
+
+ obj[n].handle = gem_create(fd, 4096);
+ obj[n].relocs_ptr = to_user_pointer(reloc);
+ obj[n].relocation_count = n;
+ gem_write(fd, obj[n].handle, 0, &bbe, sizeof(bbe));
+
+ execbuf.buffers_ptr = to_user_pointer(obj);
+ execbuf.buffer_count = n + 1;
+
+ if (flags & ACTIVE) {
+ spin = igt_spin_batch_new(fd, .dependency = obj[n].handle);
+ if (!(flags & HANG))
+ igt_spin_batch_set_timeout(spin, NSEC_PER_SEC/100);
+ igt_assert(gem_bo_busy(fd, obj[n].handle));
+ }
+
+ gem_execbuf(fd, &execbuf);
+ igt_spin_batch_free(fd, spin);
+
+ for (int i = 0; i < n; i++) {
+ uint64_t offset;
+
+ offset = ~reloc[i].presumed_offset & address_mask;
+ gem_read(fd, obj[n].handle, reloc[i].offset,
+ &offset, has_64b_reloc(fd) ? 8 : 4);
+
+ igt_debug("obj[%d] handle=%d, offset=%llx, found=%llx, presumed=%llx\n",
+ i, obj[i].handle,
+ (long long)obj[i].offset,
+ (long long)offset,
+ (long long)reloc[i].presumed_offset);
+
+ igt_assert_eq_u64(obj[i].offset, offset);
+ if (reloc[i].presumed_offset == -1)
+ igt_warn("reloc.presumed_offset == -1\n");
+ else
+ igt_assert_eq_u64(reloc[i].presumed_offset, offset);
+ }
+
+ for (int i = 0; i <= n; i++)
+ gem_close(fd, obj[i].handle);
+}
+
+static void basic_softpin(int fd)
+{
+ struct drm_i915_gem_exec_object2 obj[2];
+ struct drm_i915_gem_execbuffer2 execbuf;
+ uint64_t offset;
+ uint32_t bbe = MI_BATCH_BUFFER_END;
+
+ igt_require(gem_has_softpin(fd));
+
+ memset(obj, 0, sizeof(obj));
+ obj[1].handle = gem_create(fd, 4096);
+ gem_write(fd, obj[1].handle, 0, &bbe, sizeof(bbe));
+
+ memset(&execbuf, 0, sizeof(execbuf));
+ execbuf.buffers_ptr = to_user_pointer(&obj[1]);
+ execbuf.buffer_count = 1;
+ gem_execbuf(fd, &execbuf);
+
+ offset = obj[1].offset;
+
+ obj[0].handle = gem_create(fd, 4096);
+ obj[0].offset = obj[1].offset;
+ obj[0].flags = EXEC_OBJECT_PINNED;
+
+ execbuf.buffers_ptr = to_user_pointer(&obj[0]);
+ execbuf.buffer_count = 2;
+
+ gem_execbuf(fd, &execbuf);
+ igt_assert_eq_u64(obj[0].offset, offset);
+
+ gem_close(fd, obj[0].handle);
+ gem_close(fd, obj[1].handle);
+}
+
+igt_main
+{
+ const struct mode {
+ const char *name;
+ unsigned before, after;
+ } modes[] = {
+ { "cpu", I915_GEM_DOMAIN_CPU, I915_GEM_DOMAIN_CPU },
+ { "gtt", I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT },
+ { "wc", I915_GEM_DOMAIN_WC, I915_GEM_DOMAIN_WC },
+ { "cpu-gtt", I915_GEM_DOMAIN_CPU, I915_GEM_DOMAIN_GTT },
+ { "gtt-cpu", I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_CPU },
+ { "cpu-wc", I915_GEM_DOMAIN_CPU, I915_GEM_DOMAIN_WC },
+ { "wc-cpu", I915_GEM_DOMAIN_WC, I915_GEM_DOMAIN_CPU },
+ { "gtt-wc", I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_WC },
+ { "wc-gtt", I915_GEM_DOMAIN_WC, I915_GEM_DOMAIN_GTT },
+ { "cpu-read", I915_GEM_DOMAIN_CPU, 0 },
+ { "gtt-read", I915_GEM_DOMAIN_GTT, 0 },
+ { "wc-read", I915_GEM_DOMAIN_WC, 0 },
+ { "write-cpu", 0, I915_GEM_DOMAIN_CPU },
+ { "write-gtt", 0, I915_GEM_DOMAIN_GTT },
+ { "write-wc", 0, I915_GEM_DOMAIN_WC },
+ { "write-read", 0, 0 },
+ { },
+ }, *m;
+ const struct flags {
+ const char *name;
+ unsigned flags;
+ bool basic;
+ } flags[] = {
+ { "", 0 , true},
+ { "-noreloc", NORELOC, true },
+ { "-active", ACTIVE, true },
+ { "-hang", ACTIVE | HANG },
+ { },
+ }, *f;
+ uint64_t size;
+ int fd = -1;
+
+ igt_fixture {
+ fd = drm_open_driver_master(DRIVER_INTEL);
+ igt_require_gem(fd);
+ }
+
+ for (f = flags; f->name; f++) {
+ igt_hang_t hang;
+
+ igt_subtest_group {
+ igt_fixture {
+ if (f->flags & HANG)
+ hang = igt_allow_hang(fd, 0, 0);
+ }
+
+ for (m = modes; m->name; m++) {
+ igt_subtest_f("%s%s%s",
+ f->basic ? "basic-" : "",
+ m->name,
+ f->name) {
+ if ((m->before | m->after) & I915_GEM_DOMAIN_WC)
+ igt_require(gem_mmap__has_wc(fd));
+ basic_reloc(fd, m->before, m->after, f->flags);
+ }
+ }
+
+ if (!(f->flags & NORELOC)) {
+ igt_subtest_f("%srange%s",
+ f->basic ? "basic-" : "", f->name)
+ basic_range(fd, f->flags);
+ }
+
+ igt_fixture {
+ if (f->flags & HANG)
+ igt_disallow_hang(fd, hang);
+ }
+ }
+ }
+
+ igt_subtest("basic-softpin")
+ basic_softpin(fd);
+
+ for (size = 4096; size <= 4ull*1024*1024*1024; size <<= 1) {
+ igt_subtest_f("mmap-%u", find_last_set(size) - 1)
+ from_mmap(fd, size, MEM);
+ igt_subtest_f("readonly-%u", find_last_set(size) - 1)
+ from_mmap(fd, size, MEM | RO);
+ igt_subtest_f("cpu-%u", find_last_set(size) - 1)
+ from_mmap(fd, size, CPU);
+ igt_subtest_f("wc-%u", find_last_set(size) - 1) {
+ igt_require(gem_mmap__has_wc(fd));
+ from_mmap(fd, size, WC);
+ }
+ igt_subtest_f("gtt-%u", find_last_set(size) - 1)
+ from_mmap(fd, size, GTT);
+ }
+
+ igt_subtest("gpu")
+ from_gpu(fd);
+
+ igt_subtest("active")
+ active(fd, ALL_ENGINES);
+ for (const struct intel_execution_engine *e = intel_execution_engines;
+ e->name; e++) {
+ igt_subtest_f("active-%s", e->name)
+ active(fd, e->exec_id | e->flags);
+ }
+ igt_fixture
+ close(fd);
+}