/* * Copyright © 2016 Intel Corporation * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. */ #include #include #include "i915/gem.h" #include "i915/gem_create.h" #include "igt.h" #include "igt_device.h" #include "igt_rand.h" #include "igt_sysfs.h" IGT_TEST_DESCRIPTION("Check that we capture the user specified objects on a hang"); static void check_error_state(int dir, struct drm_i915_gem_exec_object2 *obj) { char *error, *str; bool found = false; error = igt_sysfs_get(dir, "error"); igt_sysfs_set(dir, "error", "Begone!"); igt_assert(error); igt_debug("%s\n", error); /* render ring --- user = 0x00000000 ffffd000 */ for (str = error; (str = strstr(str, "--- user = ")); str++) { uint64_t addr; uint32_t hi, lo; igt_assert(sscanf(str, "--- user = 0x%x %x", &hi, &lo) == 2); addr = hi; addr <<= 32; addr |= lo; igt_assert_eq_u64(addr, obj->offset); found = true; } free(error); igt_assert(found); } static void __capture1(int fd, int dir, uint64_t ahnd, const intel_ctx_t *ctx, unsigned ring, uint32_t target, uint64_t target_size) { const unsigned int gen = intel_gen(intel_get_drm_devid(fd)); struct drm_i915_gem_exec_object2 obj[4]; #define SCRATCH 0 #define CAPTURE 1 #define NOCAPTURE 2 #define BATCH 3 struct drm_i915_gem_relocation_entry reloc[2]; struct drm_i915_gem_execbuffer2 execbuf; uint32_t *batch, *seqno; int i; memset(obj, 0, sizeof(obj)); obj[SCRATCH].handle = gem_create(fd, 4096); obj[SCRATCH].flags = EXEC_OBJECT_WRITE; obj[CAPTURE].handle = target; obj[CAPTURE].flags = EXEC_OBJECT_CAPTURE; obj[NOCAPTURE].handle = gem_create(fd, 4096); obj[BATCH].handle = gem_create(fd, 4096); obj[BATCH].relocs_ptr = (uintptr_t)reloc; obj[BATCH].relocation_count = !ahnd ? ARRAY_SIZE(reloc) : 0; for (i = 0; i < ARRAY_SIZE(obj); i++) { obj[i].offset = get_offset(ahnd, obj[i].handle, i == CAPTURE ? target_size : 4096, 0); obj[i].flags |= ahnd ? EXEC_OBJECT_PINNED : 0; } memset(reloc, 0, sizeof(reloc)); reloc[0].target_handle = obj[BATCH].handle; /* recurse */ reloc[0].presumed_offset = obj[BATCH].offset; reloc[0].offset = 5*sizeof(uint32_t); reloc[0].delta = 0; reloc[0].read_domains = I915_GEM_DOMAIN_COMMAND; reloc[0].write_domain = 0; reloc[1].target_handle = obj[SCRATCH].handle; /* breadcrumb */ reloc[1].presumed_offset = obj[SCRATCH].offset; reloc[1].offset = sizeof(uint32_t); reloc[1].delta = 0; reloc[1].read_domains = I915_GEM_DOMAIN_RENDER; reloc[1].write_domain = I915_GEM_DOMAIN_RENDER; seqno = gem_mmap__device_coherent(fd, obj[SCRATCH].handle, 0, 4096, PROT_READ); gem_set_domain(fd, obj[SCRATCH].handle, I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT); batch = gem_mmap__cpu(fd, obj[BATCH].handle, 0, 4096, PROT_WRITE); gem_set_domain(fd, obj[BATCH].handle, I915_GEM_DOMAIN_CPU, I915_GEM_DOMAIN_CPU); i = 0; batch[i] = MI_STORE_DWORD_IMM | (gen < 6 ? 1 << 22 : 0); if (gen >= 8) { batch[++i] = obj[SCRATCH].offset; batch[++i] = obj[SCRATCH].offset >> 32; } else if (gen >= 4) { batch[++i] = 0; batch[++i] = 0; reloc[1].offset += sizeof(uint32_t); } else { batch[i]--; batch[++i] = 0; } batch[++i] = 0xc0ffee; if (gen < 4) batch[++i] = MI_NOOP; batch[++i] = MI_BATCH_BUFFER_START; /* not crashed? try again! */ if (gen >= 8) { batch[i] |= 1 << 8 | 1; batch[++i] = obj[BATCH].offset; batch[++i] = obj[BATCH].offset >> 32; } else if (gen >= 6) { batch[i] |= 1 << 8; batch[++i] = 0; } else { batch[i] |= 2 << 6; batch[++i] = 0; if (gen < 4) { batch[i] |= 1; reloc[0].delta = 1; } } munmap(batch, 4096); memset(&execbuf, 0, sizeof(execbuf)); execbuf.buffers_ptr = (uintptr_t)obj; execbuf.buffer_count = ARRAY_SIZE(obj); execbuf.flags = ring; if (gen > 3 && gen < 6) execbuf.flags |= I915_EXEC_SECURE; execbuf.rsvd1 = ctx->id; igt_assert(!READ_ONCE(*seqno)); gem_execbuf(fd, &execbuf); /* Wait for the request to start */ while (READ_ONCE(*seqno) != 0xc0ffee) igt_assert(gem_bo_busy(fd, obj[SCRATCH].handle)); munmap(seqno, 4096); /* Check that only the buffer we marked is reported in the error */ igt_force_gpu_reset(fd); check_error_state(dir, &obj[CAPTURE]); gem_sync(fd, obj[BATCH].handle); for (i = 0; i < ARRAY_SIZE(obj); i++) put_offset(ahnd, obj[i].handle); gem_close(fd, obj[BATCH].handle); gem_close(fd, obj[NOCAPTURE].handle); gem_close(fd, obj[SCRATCH].handle); } static void capture(int fd, int dir, const intel_ctx_t *ctx, unsigned ring) { uint32_t handle; uint64_t ahnd; handle = gem_create(fd, 4096); ahnd = get_reloc_ahnd(fd, ctx->id); __capture1(fd, dir, ahnd, ctx, ring, handle, 4096); gem_close(fd, handle); put_ahnd(ahnd); } static int cmp(const void *A, const void *B) { const uint64_t *a = A, *b = B; if (*a < *b) return -1; if (*a > *b) return 1; return 0; } static struct offset { uint64_t addr; unsigned long idx; } *__captureN(int fd, int dir, uint64_t ahnd, unsigned ring, unsigned int size, int count, unsigned int flags) #define INCREMENTAL 0x1 #define ASYNC 0x2 { const unsigned int gen = intel_gen(intel_get_drm_devid(fd)); struct drm_i915_gem_exec_object2 *obj; struct drm_i915_gem_relocation_entry reloc[2]; struct drm_i915_gem_execbuffer2 execbuf; uint32_t *batch, *seqno; struct offset *offsets; int i; offsets = calloc(count, sizeof(*offsets)); igt_assert(offsets); obj = calloc(count + 2, sizeof(*obj)); igt_assert(obj); obj[0].handle = gem_create(fd, 4096); obj[0].offset = get_offset(ahnd, obj[0].handle, 4096, 0); obj[0].flags = EXEC_OBJECT_WRITE | ahnd ? EXEC_OBJECT_PINNED : 0; for (i = 0; i < count; i++) { obj[i + 1].handle = gem_create(fd, size); obj[i + 1].offset = get_offset(ahnd, obj[i + 1].handle, size, 0); obj[i + 1].flags = EXEC_OBJECT_CAPTURE | EXEC_OBJECT_SUPPORTS_48B_ADDRESS; if (ahnd) obj[i + 1].flags |= EXEC_OBJECT_PINNED; if (flags & INCREMENTAL) { uint32_t *ptr; ptr = gem_mmap__cpu(fd, obj[i + 1].handle, 0, size, PROT_WRITE); for (unsigned int n = 0; n < size / sizeof(*ptr); n++) ptr[n] = i * size + n; munmap(ptr, size); } } obj[count + 1].handle = gem_create(fd, 4096); obj[count + 1].relocs_ptr = (uintptr_t)reloc; obj[count + 1].relocation_count = !ahnd ? ARRAY_SIZE(reloc) : 0; obj[count + 1].offset = get_offset(ahnd, obj[count + 1].handle, 4096, 0); obj[count + 1].flags = ahnd ? EXEC_OBJECT_PINNED : 0; memset(reloc, 0, sizeof(reloc)); reloc[0].target_handle = obj[count + 1].handle; /* recurse */ reloc[0].presumed_offset = obj[count + 1].offset; reloc[0].offset = 5*sizeof(uint32_t); reloc[0].delta = 0; reloc[0].read_domains = I915_GEM_DOMAIN_COMMAND; reloc[0].write_domain = 0; reloc[1].target_handle = obj[0].handle; /* breadcrumb */ reloc[1].presumed_offset = obj[0].offset; reloc[1].offset = sizeof(uint32_t); reloc[1].delta = 0; reloc[1].read_domains = I915_GEM_DOMAIN_RENDER; reloc[1].write_domain = I915_GEM_DOMAIN_RENDER; if (!ahnd) { obj[count + 1].relocs_ptr = (uintptr_t)reloc; obj[count + 1].relocation_count = ARRAY_SIZE(reloc); } seqno = gem_mmap__device_coherent(fd, obj[0].handle, 0, 4096, PROT_READ); gem_set_domain(fd, obj[0].handle, I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT); batch = gem_mmap__cpu(fd, obj[count + 1].handle, 0, 4096, PROT_WRITE); gem_set_domain(fd, obj[count + 1].handle, I915_GEM_DOMAIN_CPU, I915_GEM_DOMAIN_CPU); i = 0; batch[i] = MI_STORE_DWORD_IMM | (gen < 6 ? 1 << 22 : 0); if (gen >= 8) { batch[++i] = obj[0].offset; batch[++i] = obj[0].offset >> 32; } else if (gen >= 4) { batch[++i] = 0; batch[++i] = 0; reloc[1].offset += sizeof(uint32_t); } else { batch[i]--; batch[++i] = 0; } batch[++i] = 0xc0ffee; if (gen < 4) batch[++i] = MI_NOOP; batch[++i] = MI_BATCH_BUFFER_START; /* not crashed? try again! */ if (gen >= 8) { batch[i] |= 1 << 8 | 1; batch[++i] = obj[count + 1].offset; batch[++i] = obj[count + 1].offset >> 32; } else if (gen >= 6) { batch[i] |= 1 << 8; batch[++i] = 0; } else { batch[i] |= 2 << 6; batch[++i] = 0; if (gen < 4) { batch[i] |= 1; reloc[0].delta = 1; } } munmap(batch, 4096); memset(&execbuf, 0, sizeof(execbuf)); execbuf.buffers_ptr = (uintptr_t)obj; execbuf.buffer_count = count + 2; execbuf.flags = ring; if (gen > 3 && gen < 6) execbuf.flags |= I915_EXEC_SECURE; igt_assert(!READ_ONCE(*seqno)); gem_execbuf(fd, &execbuf); /* Wait for the request to start */ while (READ_ONCE(*seqno) != 0xc0ffee) igt_assert(gem_bo_busy(fd, obj[0].handle)); munmap(seqno, 4096); if (!(flags & ASYNC)) { igt_force_gpu_reset(fd); gem_sync(fd, obj[count + 1].handle); } gem_close(fd, obj[count + 1].handle); put_offset(ahnd, obj[count + 1].handle); for (i = 0; i < count; i++) { offsets[i].addr = obj[i + 1].offset; offsets[i].idx = i; gem_close(fd, obj[i + 1].handle); put_offset(ahnd, obj[i + 1].handle); } gem_close(fd, obj[0].handle); put_offset(ahnd, obj[0].handle); qsort(offsets, count, sizeof(*offsets), cmp); igt_assert(offsets[0].addr <= offsets[count-1].addr); return offsets; } static unsigned long zlib_inflate(uint32_t **ptr, unsigned long len) { struct z_stream_s zstream; void *out; memset(&zstream, 0, sizeof(zstream)); zstream.next_in = (unsigned char *)*ptr; zstream.avail_in = 4*len; if (inflateInit(&zstream) != Z_OK) return 0; out = malloc(128*4096); /* approximate obj size */ zstream.next_out = out; zstream.avail_out = 128*4096; do { switch (inflate(&zstream, Z_SYNC_FLUSH)) { case Z_STREAM_END: goto end; case Z_OK: break; default: inflateEnd(&zstream); return 0; } if (zstream.avail_out) break; out = realloc(out, 2*zstream.total_out); if (out == NULL) { inflateEnd(&zstream); return 0; } zstream.next_out = (unsigned char *)out + zstream.total_out; zstream.avail_out = zstream.total_out; } while (1); end: inflateEnd(&zstream); free(*ptr); *ptr = out; return zstream.total_out / 4; } static unsigned long ascii85_decode(char *in, uint32_t **out, bool inflate, char **end) { unsigned long len = 0, size = 1024; *out = realloc(*out, sizeof(uint32_t)*size); if (*out == NULL) return 0; while (*in >= '!' && *in <= 'z') { uint32_t v = 0; if (len == size) { size *= 2; *out = realloc(*out, sizeof(uint32_t)*size); if (*out == NULL) return 0; } if (*in == 'z') { in++; } else { v += in[0] - 33; v *= 85; v += in[1] - 33; v *= 85; v += in[2] - 33; v *= 85; v += in[3] - 33; v *= 85; v += in[4] - 33; in += 5; } (*out)[len++] = v; } *end = in; if (!inflate) return len; return zlib_inflate(out, len); } static void many(int fd, int dir, uint64_t size, unsigned int flags) { uint64_t ram, gtt, ahnd; unsigned long count, blobs; struct offset *offsets; char *error, *str; gtt = gem_aperture_size(fd) / size; ram = (intel_get_avail_ram_mb() << 20) / size; igt_debug("Available objects in GTT:%"PRIu64", RAM:%"PRIu64"\n", gtt, ram); count = min(gtt, ram) / 4; igt_require(count > 1); intel_require_memory(count, size, CHECK_RAM); ahnd = get_reloc_ahnd(fd, 0); offsets = __captureN(fd, dir, ahnd, 0, size, count, flags); error = igt_sysfs_get(dir, "error"); igt_sysfs_set(dir, "error", "Begone!"); igt_assert(error); blobs = 0; /* render ring --- user = 0x00000000 ffffd000 */ str = strstr(error, "--- user = "); while (str) { uint32_t *data = NULL; unsigned long i, sz; uint64_t addr; if (strncmp(str, "--- user = 0x", 13)) break; str += 13; addr = strtoul(str, &str, 16); addr <<= 32; addr |= strtoul(str + 1, &str, 16); igt_assert(*str++ == '\n'); if (!(*str == ':' || *str == '~')) continue; igt_debug("blob:%.64s\n", str); sz = ascii85_decode(str + 1, &data, *str == ':', &str); igt_assert_eq(4 * sz, size); igt_assert(*str++ == '\n'); str = strchr(str, '-'); if (flags & INCREMENTAL) { unsigned long start = 0; unsigned long end = count; uint32_t expect; while (end > start) { i = (end - start) / 2 + start; if (offsets[i].addr < addr) start = i + 1; else if (offsets[i].addr > addr) end = i; else break; } igt_assert(offsets[i].addr == addr); igt_debug("offset:%"PRIx64", index:%ld\n", addr, offsets[i].idx); expect = offsets[i].idx * size; for (i = 0; i < sz; i++) igt_assert_eq(data[i], expect++); } else { for (i = 0; i < sz; i++) igt_assert_eq(data[i], 0); } blobs++; free(data); } igt_info("Captured %lu %"PRId64"-blobs out of a total of %lu\n", blobs, size >> 12, count); igt_assert(count); free(error); free(offsets); put_ahnd(ahnd); } static void prioinv(int fd, int dir, const intel_ctx_t *ctx, unsigned ring, const char *name) { const uint32_t bbe = MI_BATCH_BUFFER_END; struct drm_i915_gem_exec_object2 obj = { .handle = gem_create(fd, 4096), }; struct drm_i915_gem_execbuffer2 execbuf = { .buffers_ptr = to_user_pointer(&obj), .buffer_count = 1, .flags = ring, .rsvd1 = ctx->id, }; int64_t timeout = NSEC_PER_SEC; /* 1s, feeling generous, blame debug */ uint64_t ram, gtt, ahnd, size = 4 << 20; unsigned long count; int link[2], dummy; ahnd = get_reloc_ahnd(fd, ctx->id); obj.offset = get_offset(ahnd, obj.handle, 4096, 0); igt_require(gem_scheduler_enabled(fd)); igt_require(igt_params_set(fd, "reset", "%u", -1)); /* engine resets! */ igt_require(gem_gpu_reset_type(fd) > 1); /* Needs to be fast enough for the hangcheck to return within 1s */ igt_require(gem_engine_property_printf(fd, name, "preempt_timeout_ms", "%d", 0) > 0); gem_engine_property_printf(fd, name, "preempt_timeout_ms", "%d", 500); gtt = gem_aperture_size(fd) / size; ram = (intel_get_avail_ram_mb() << 20) / size; igt_debug("Available objects in GTT:%"PRIu64", RAM:%"PRIu64"\n", gtt, ram); count = min(gtt, ram) / 4; count = min(count, 256); /* Keep the duration within reason */ igt_require(count > 1); intel_require_memory(count, size, CHECK_RAM); gem_write(fd, obj.handle, 0, &bbe, sizeof(bbe)); gem_execbuf(fd, &execbuf); gem_sync(fd, obj.handle); igt_assert(pipe(link) == 0); igt_fork(child, 1) { fd = gem_reopen_driver(fd); igt_debug("Submitting large capture [%ld x %dMiB objects]\n", count, (int)(size >> 20)); intel_allocator_init(); /* Reopen the allocator in the new process. */ ahnd = get_reloc_ahnd(fd, 0); free(__captureN(fd, dir, ahnd, ring, size, count, ASYNC)); put_ahnd(ahnd); write(link[1], &fd, sizeof(fd)); /* wake the parent up */ igt_force_gpu_reset(fd); write(link[1], &fd, sizeof(fd)); /* wake the parent up */ } read(link[0], &dummy, sizeof(dummy)); igt_require_f(poll(&(struct pollfd){link[0], POLLIN}, 1, 500) == 0, "Capture completed too quickly! Will not block\n"); igt_debug("Submitting nop\n"); gem_execbuf(fd, &execbuf); igt_assert_eq(gem_wait(fd, obj.handle, &timeout), 0); gem_close(fd, obj.handle); igt_assert_f(poll(&(struct pollfd){link[0], POLLIN}, 1, 0) == 0, "Capture completed before nop!\n"); igt_debug("Waiting for capture/reset to complete\n"); igt_waitchildren(); close(link[0]); close(link[1]); gem_quiescent_gpu(fd); put_offset(ahnd, obj.handle); put_ahnd(ahnd); } static void userptr(int fd, int dir) { const intel_ctx_t *ctx = intel_ctx_0(fd); uint32_t handle; uint64_t ahnd; void *ptr; igt_assert(posix_memalign(&ptr, 4096, 4096) == 0); igt_require(__gem_userptr(fd, ptr, 4096, 0, 0, &handle) == 0); ahnd = get_reloc_ahnd(fd, ctx->id); __capture1(fd, dir, ahnd, intel_ctx_0(fd), 0, handle, 4096); gem_close(fd, handle); put_ahnd(ahnd); free(ptr); } static bool has_capture(int fd) { drm_i915_getparam_t gp; int async = -1; gp.param = I915_PARAM_HAS_EXEC_CAPTURE; gp.value = &async; drmIoctl(fd, DRM_IOCTL_I915_GETPARAM, &gp); return async > 0; } static size_t safer_strlen(const char *s) { return s ? strlen(s) : 0; } #define test_each_engine(T, i915, ctx, e) \ igt_subtest_with_dynamic(T) for_each_ctx_engine(i915, ctx, e) \ for_each_if(gem_class_can_store_dword(i915, (e)->class)) \ igt_dynamic_f("%s", (e)->name) igt_main { const struct intel_execution_engine2 *e; const intel_ctx_t *ctx; igt_hang_t hang; int fd = -1; int dir = -1; igt_fixture { int gen; fd = drm_open_driver(DRIVER_INTEL); gen = intel_gen(intel_get_drm_devid(fd)); if (gen > 3 && gen < 6) /* ctg and ilk need secure batches */ igt_device_set_master(fd); igt_require_gem(fd); gem_require_mmap_wc(fd); igt_require(has_capture(fd)); ctx = intel_ctx_create_all_physical(fd); igt_allow_hang(fd, ctx->id, HANG_ALLOW_CAPTURE); dir = igt_sysfs_open(fd); igt_require(igt_sysfs_set(dir, "error", "Begone!")); igt_require(safer_strlen(igt_sysfs_get(dir, "error")) > 0); } test_each_engine("capture", fd, ctx, e) capture(fd, dir, ctx, e->flags); igt_subtest_f("many-4K-zero") { igt_require(gem_can_store_dword(fd, 0)); many(fd, dir, 1<<12, 0); } igt_subtest_f("many-4K-incremental") { igt_require(gem_can_store_dword(fd, 0)); many(fd, dir, 1<<12, INCREMENTAL); } igt_subtest_f("many-2M-zero") { igt_require(gem_can_store_dword(fd, 0)); many(fd, dir, 2<<20, 0); } igt_subtest_f("many-2M-incremental") { igt_require(gem_can_store_dword(fd, 0)); many(fd, dir, 2<<20, INCREMENTAL); } igt_subtest_f("many-256M-incremental") { igt_require(gem_can_store_dword(fd, 0)); many(fd, dir, 256<<20, INCREMENTAL); } /* And check we can read from different types of objects */ igt_subtest_f("userptr") { igt_require(gem_can_store_dword(fd, 0)); userptr(fd, dir); } test_each_engine("pi", fd, ctx, e) prioinv(fd, dir, ctx, e->flags, e->name); igt_fixture { close(dir); igt_disallow_hang(fd, hang); intel_ctx_destroy(fd, ctx); close(fd); } }