summaryrefslogtreecommitdiff
path: root/tests/i915/gem_streaming_writes.c
diff options
context:
space:
mode:
authorArkadiusz Hiler <arkadiusz.hiler@intel.com>2018-10-18 14:06:42 +0300
committerArkadiusz Hiler <arkadiusz.hiler@intel.com>2018-10-23 10:55:51 +0300
commit741bf7064c467df725c14cc0b3b8b50436f9ee09 (patch)
tree0ad6fb217dca79a8f1175fb289979b574222fefa /tests/i915/gem_streaming_writes.c
parent78619fde4008424c472906041edb1d204e014f7c (diff)
tests: Introduce i915 directory
We can already move all the tests with distinct prefixes: gem_, gen3_ and i915_. pm_ and drv_ tests will follow in batches, so we can do the adjustments in the reporting/filtering layer of the CI system. v2: Fix test-list.txt generation with meson v3: Fix docs build (Petri) Cc: Chris Wilson <chris@chris-wilson.co.uk> Cc: Petri Latvala <petri.latvala@intel.com> Cc: Martin Peres <martin.peres@linux.intel.com> Signed-off-by: Arkadiusz Hiler <arkadiusz.hiler@intel.com> Reviewed-by: Petri Latvala <petri.latvala@intel.com> Tested-by: Petri Latvala <petri.latvala@intel.com> Acked-by: Daniel Vetter <daniel.vetter@ffwll.ch> Acked-by: Chris Wilson <chris@chris-wilson.co.uk>
Diffstat (limited to 'tests/i915/gem_streaming_writes.c')
-rw-r--r--tests/i915/gem_streaming_writes.c405
1 files changed, 405 insertions, 0 deletions
diff --git a/tests/i915/gem_streaming_writes.c b/tests/i915/gem_streaming_writes.c
new file mode 100644
index 00000000..e83d69de
--- /dev/null
+++ b/tests/i915/gem_streaming_writes.c
@@ -0,0 +1,405 @@
+/*
+ * Copyright © 2015 Intel Corporation
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * the rights to use, copy, modify, merge, publish, distribute, sublicense,
+ * and/or sell copies of the Software, and to permit persons to whom the
+ * Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice (including the next
+ * paragraph) shall be included in all copies or substantial portions of the
+ * Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
+ * IN THE SOFTWARE.
+ *
+ * Authors:
+ * Chris Wilson <chris@chris-wilson.co.uk>
+ *
+ */
+
+#include "igt.h"
+#include <unistd.h>
+#include <stdlib.h>
+#include <stdio.h>
+#include <string.h>
+#include <fcntl.h>
+#include <inttypes.h>
+#include <pthread.h>
+#include <errno.h>
+#include <sys/stat.h>
+#include <sys/ioctl.h>
+#include "drm.h"
+
+#define OBJECT_SIZE 1024*1024
+#define CHUNK_SIZE 32
+
+#define COPY_BLT_CMD (2<<29|0x53<<22|0x6)
+#define BLT_WRITE_ALPHA (1<<21)
+#define BLT_WRITE_RGB (1<<20)
+#define BLT_WRITE_ARGB (BLT_WRITE_ALPHA | BLT_WRITE_RGB)
+
+#define LOCAL_I915_EXEC_HANDLE_LUT (1<<12)
+
+IGT_TEST_DESCRIPTION("Test of streaming writes into active GPU sources");
+
+#define SRC 0
+#define DST 1
+#define BATCH 2
+
+#define src exec[SRC].handle
+#define src_offset exec[SRC].offset
+#define dst exec[DST].handle
+#define dst_offset exec[DST].offset
+
+static void test_streaming(int fd, int mode, int sync)
+{
+ const int has_64bit_reloc = intel_gen(intel_get_drm_devid(fd)) >= 8;
+ struct drm_i915_gem_execbuffer2 execbuf;
+ struct drm_i915_gem_exec_object2 exec[3];
+ struct drm_i915_gem_relocation_entry reloc[128];
+ uint32_t tmp[] = { MI_BATCH_BUFFER_END };
+ uint64_t __src_offset, __dst_offset;
+ uint32_t *s, *d;
+ uint32_t offset;
+ struct {
+ uint32_t handle;
+ uint64_t offset;
+ } *batch;
+ int i, n;
+
+ memset(exec, 0, sizeof(exec));
+ exec[SRC].handle = gem_create(fd, OBJECT_SIZE);
+ exec[DST].handle = gem_create(fd, OBJECT_SIZE);
+
+ switch (mode) {
+ case 0: /* cpu/snoop */
+ gem_set_caching(fd, src, I915_CACHING_CACHED);
+ s = gem_mmap__cpu(fd, src, 0, OBJECT_SIZE,
+ PROT_READ | PROT_WRITE);
+ break;
+ case 1: /* gtt */
+ s = gem_mmap__gtt(fd, src, OBJECT_SIZE,
+ PROT_READ | PROT_WRITE);
+ break;
+ case 2: /* wc */
+ s = gem_mmap__wc(fd, src, 0, OBJECT_SIZE,
+ PROT_READ | PROT_WRITE);
+ break;
+ }
+ *s = 0; /* fault the object into the mappable range first (for GTT) */
+
+ d = gem_mmap__cpu(fd, dst, 0, OBJECT_SIZE, PROT_READ);
+
+ gem_write(fd, dst, 0, tmp, sizeof(tmp));
+ memset(&execbuf, 0, sizeof(execbuf));
+ execbuf.buffers_ptr = to_user_pointer(exec);
+ execbuf.buffer_count = 2;
+ execbuf.flags = LOCAL_I915_EXEC_HANDLE_LUT;
+ if (__gem_execbuf(fd, &execbuf)) {
+ execbuf.flags = 0;
+ igt_require(__gem_execbuf(fd, &execbuf) == 0);
+ }
+ /* We assume that the active objects are fixed to avoid relocations */
+ __src_offset = src_offset;
+ __dst_offset = dst_offset;
+
+ memset(reloc, 0, sizeof(reloc));
+ for (i = 0; i < 64; i++) {
+ reloc[2*i+0].offset = 64*i + 4 * sizeof(uint32_t);
+ reloc[2*i+0].delta = 0;
+ reloc[2*i+0].target_handle = execbuf.flags & LOCAL_I915_EXEC_HANDLE_LUT ? DST : dst;
+ reloc[2*i+0].presumed_offset = dst_offset;
+ reloc[2*i+0].read_domains = I915_GEM_DOMAIN_RENDER;
+ reloc[2*i+0].write_domain = I915_GEM_DOMAIN_RENDER;
+
+ reloc[2*i+1].offset = 64*i + 7 * sizeof(uint32_t);
+ if (has_64bit_reloc)
+ reloc[2*i+1].offset += sizeof(uint32_t);
+ reloc[2*i+1].delta = 0;
+ reloc[2*i+1].target_handle = execbuf.flags & LOCAL_I915_EXEC_HANDLE_LUT ? SRC : src;
+ reloc[2*i+1].presumed_offset = src_offset;
+ reloc[2*i+1].read_domains = I915_GEM_DOMAIN_RENDER;
+ reloc[2*i+1].write_domain = 0;
+ }
+ gem_execbuf(fd, &execbuf);
+ igt_assert_eq_u64(__src_offset, src_offset);
+ igt_assert_eq_u64(__dst_offset, dst_offset);
+
+ exec[DST].flags = EXEC_OBJECT_WRITE;
+ exec[BATCH].relocation_count = 2;
+ execbuf.buffer_count = 3;
+ execbuf.flags |= I915_EXEC_NO_RELOC;
+ if (gem_has_blt(fd))
+ execbuf.flags |= I915_EXEC_BLT;
+
+ batch = malloc(sizeof(*batch) * (OBJECT_SIZE / CHUNK_SIZE / 64));
+ for (i = n = 0; i < OBJECT_SIZE / CHUNK_SIZE / 64; i++) {
+ uint32_t *base;
+
+ batch[i].handle = gem_create(fd, 4096);
+ batch[i].offset = 0;
+
+ base = gem_mmap__cpu(fd, batch[i].handle, 0, 4096, PROT_WRITE);
+ gem_set_domain(fd, batch[i].handle,
+ I915_GEM_DOMAIN_CPU, I915_GEM_DOMAIN_CPU);
+
+ for (int j = 0; j < 64; j++) {
+ unsigned x = (n * CHUNK_SIZE) % 4096 >> 2;
+ unsigned y = (n * CHUNK_SIZE) / 4096;
+ uint32_t *b = base + 16 * j;
+ int k = 0;
+
+ b[k] = COPY_BLT_CMD | BLT_WRITE_ARGB;
+ if (has_64bit_reloc)
+ b[k] += 2;
+ k++;
+ b[k++] = 0xcc << 16 | 1 << 25 | 1 << 24 | 4096;
+ b[k++] = (y << 16) | x;
+ b[k++] = ((y+1) << 16) | (x + (CHUNK_SIZE >> 2));
+ b[k++] = dst_offset;
+ if (has_64bit_reloc)
+ b[k++] = dst_offset >> 32;
+ b[k++] = (y << 16) | x;
+ b[k++] = 4096;
+ b[k++] = src_offset;
+ if (has_64bit_reloc)
+ b[k++] = src_offset >> 32;
+ b[k++] = MI_BATCH_BUFFER_END;
+
+ n++;
+ }
+
+ munmap(base, 4096);
+ }
+
+ for (int pass = 0; pass < 256; pass++) {
+ int domain = mode ? I915_GEM_DOMAIN_GTT : I915_GEM_DOMAIN_CPU;
+ gem_set_domain(fd, src, domain, domain);
+
+ if (pass == 0) {
+ for (i = 0; i < OBJECT_SIZE/4; i++)
+ s[i] = i;
+ }
+
+ /* Now copy from the src to the dst in 32byte chunks */
+ for (offset = 0; offset < OBJECT_SIZE; offset += CHUNK_SIZE) {
+ int b;
+
+ if (pass) {
+ if (sync)
+ gem_set_domain(fd, src, domain, domain);
+ for (i = 0; i < CHUNK_SIZE/4; i++)
+ s[offset/4 + i] = (OBJECT_SIZE*pass + offset)/4 + i;
+ }
+
+ igt_assert(exec[DST].flags & EXEC_OBJECT_WRITE);
+
+ b = offset / CHUNK_SIZE / 64;
+ n = offset / CHUNK_SIZE % 64;
+ exec[BATCH].relocs_ptr = to_user_pointer((reloc + 2*n));
+ exec[BATCH].handle = batch[b].handle;
+ exec[BATCH].offset = batch[b].offset;
+ execbuf.batch_start_offset = 64*n;
+
+ gem_execbuf(fd, &execbuf);
+ igt_assert_eq_u64(__src_offset, src_offset);
+ igt_assert_eq_u64(__dst_offset, dst_offset);
+
+ batch[b].offset = exec[BATCH].offset;
+ }
+
+ gem_set_domain(fd, dst, I915_GEM_DOMAIN_CPU, 0);
+ for (offset = 0; offset < OBJECT_SIZE/4; offset++)
+ igt_assert_eq(pass*OBJECT_SIZE/4 + offset, d[offset]);
+ }
+
+ for (i = 0; i < OBJECT_SIZE / CHUNK_SIZE / 64; i++)
+ gem_close(fd, batch[i].handle);
+ free(batch);
+
+ munmap(s, OBJECT_SIZE);
+ gem_close(fd, src);
+ munmap(d, OBJECT_SIZE);
+ gem_close(fd, dst);
+}
+
+static void test_batch(int fd, int mode, int reverse)
+{
+ const int has_64bit_reloc = intel_gen(intel_get_drm_devid(fd)) >= 8;
+ struct drm_i915_gem_execbuffer2 execbuf;
+ struct drm_i915_gem_exec_object2 exec[3];
+ struct drm_i915_gem_relocation_entry reloc[2];
+ uint32_t tmp[] = { MI_BATCH_BUFFER_END };
+ uint64_t __src_offset, __dst_offset;
+ bool need_64b_start_offset = true;
+ uint64_t batch_size;
+ uint32_t *s, *d;
+ uint32_t *base;
+ uint32_t offset;
+
+ memset(exec, 0, sizeof(exec));
+ exec[DST].handle = gem_create(fd, OBJECT_SIZE);
+ exec[SRC].handle = gem_create(fd, OBJECT_SIZE);
+
+ s = gem_mmap__wc(fd, src, 0, OBJECT_SIZE, PROT_READ | PROT_WRITE);
+
+ d = gem_mmap__cpu(fd, dst, 0, OBJECT_SIZE, PROT_READ);
+
+ memset(reloc, 0, sizeof(reloc));
+ reloc[0].offset = 4 * sizeof(uint32_t);
+ reloc[0].delta = 0;
+ reloc[0].target_handle = execbuf.flags & LOCAL_I915_EXEC_HANDLE_LUT ? DST : dst;
+ reloc[0].presumed_offset = dst_offset;
+ reloc[0].read_domains = I915_GEM_DOMAIN_RENDER;
+ reloc[0].write_domain = I915_GEM_DOMAIN_RENDER;
+
+ reloc[1].offset = 7 * sizeof(uint32_t);
+ if (has_64bit_reloc)
+ reloc[1].offset += sizeof(uint32_t);
+ reloc[1].delta = 0;
+ reloc[1].target_handle = execbuf.flags & LOCAL_I915_EXEC_HANDLE_LUT ? SRC : src;
+ reloc[1].presumed_offset = src_offset;
+ reloc[1].read_domains = I915_GEM_DOMAIN_RENDER;
+ reloc[1].write_domain = 0;
+
+ batch_size = ALIGN(OBJECT_SIZE / CHUNK_SIZE * 128, 4096);
+ exec[BATCH].relocs_ptr = to_user_pointer(reloc);
+ exec[BATCH].relocation_count = 2;
+ exec[BATCH].handle = gem_create(fd, batch_size);
+
+ switch (mode) {
+ case 0: /* cpu/snoop */
+ igt_require(gem_has_llc(fd));
+ base = gem_mmap__cpu(fd, exec[BATCH].handle, 0, batch_size,
+ PROT_READ | PROT_WRITE);
+ break;
+ case 1: /* gtt */
+ base = gem_mmap__gtt(fd, exec[BATCH].handle, batch_size,
+ PROT_READ | PROT_WRITE);
+ break;
+ case 2: /* wc */
+ base = gem_mmap__wc(fd, exec[BATCH].handle, 0, batch_size,
+ PROT_READ | PROT_WRITE);
+ break;
+ }
+ *base = 0; /* fault the object into the mappable range first */
+
+ gem_write(fd, exec[BATCH].handle, 0, tmp, sizeof(tmp));
+ memset(&execbuf, 0, sizeof(execbuf));
+ execbuf.buffers_ptr = to_user_pointer(exec);
+ execbuf.buffer_count = 3;
+ execbuf.flags = LOCAL_I915_EXEC_HANDLE_LUT;
+ if (gem_has_blt(fd))
+ execbuf.flags |= I915_EXEC_BLT;
+ if (__gem_execbuf(fd, &execbuf)) {
+ execbuf.flags &= ~LOCAL_I915_EXEC_HANDLE_LUT;
+ gem_execbuf(fd, &execbuf);
+ }
+ execbuf.flags |= I915_EXEC_NO_RELOC;
+ exec[DST].flags = EXEC_OBJECT_WRITE;
+ /* We assume that the active objects are fixed to avoid relocations */
+ exec[BATCH].relocation_count = 0;
+ __src_offset = src_offset;
+ __dst_offset = dst_offset;
+
+ offset = mode ? I915_GEM_DOMAIN_GTT : I915_GEM_DOMAIN_CPU;
+ gem_set_domain(fd, exec[BATCH].handle, offset, offset);
+ for (int pass = 0; pass < 256; pass++) {
+ gem_set_domain(fd, src, I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT);
+ for (offset = 0; offset < OBJECT_SIZE/4; offset++)
+ s[offset] = OBJECT_SIZE*pass/4 + offset;
+
+ /* Now copy from the src to the dst in 32byte chunks */
+ for (offset = 0; offset < OBJECT_SIZE / CHUNK_SIZE; offset++) {
+ unsigned x = (offset * CHUNK_SIZE) % 4096 >> 2;
+ unsigned y = (offset * CHUNK_SIZE) / 4096;
+ int k;
+
+ execbuf.batch_start_offset = 128 * offset;
+ if (!need_64b_start_offset)
+ execbuf.batch_start_offset += 8 * (pass & 7);
+ igt_assert(execbuf.batch_start_offset <= batch_size - 64);
+ if (reverse)
+ execbuf.batch_start_offset = batch_size - execbuf.batch_start_offset - 64;
+ igt_assert(execbuf.batch_start_offset <= batch_size - 64);
+ k = execbuf.batch_start_offset / 4;
+
+ base[k] = COPY_BLT_CMD | BLT_WRITE_ARGB;
+ if (has_64bit_reloc)
+ base[k] += 2;
+ k++;
+ base[k++] = 0xcc << 16 | 1 << 25 | 1 << 24 | 4096;
+ base[k++] = (y << 16) | x;
+ base[k++] = ((y+1) << 16) | (x + (CHUNK_SIZE >> 2));
+ base[k++] = dst_offset;
+ if (has_64bit_reloc)
+ base[k++] = dst_offset >> 32;
+ base[k++] = (y << 16) | x;
+ base[k++] = 4096;
+ base[k++] = src_offset;
+ if (has_64bit_reloc)
+ base[k++] = src_offset >> 32;
+ base[k++] = MI_BATCH_BUFFER_END;
+
+ igt_assert(exec[DST].flags & EXEC_OBJECT_WRITE);
+ gem_execbuf(fd, &execbuf);
+ igt_assert_eq_u64(__src_offset, src_offset);
+ igt_assert_eq_u64(__dst_offset, dst_offset);
+ }
+
+ gem_set_domain(fd, dst, I915_GEM_DOMAIN_CPU, 0);
+ for (offset = 0; offset < OBJECT_SIZE/4; offset++)
+ igt_assert_eq(pass*OBJECT_SIZE/4 + offset, d[offset]);
+ }
+
+ munmap(base, OBJECT_SIZE / CHUNK_SIZE * 128);
+ gem_close(fd, exec[BATCH].handle);
+
+ munmap(s, OBJECT_SIZE);
+ gem_close(fd, src);
+ munmap(d, OBJECT_SIZE);
+ gem_close(fd, dst);
+}
+
+igt_main
+{
+ int fd, sync;
+
+ igt_fixture {
+ fd = drm_open_driver(DRIVER_INTEL);
+ igt_require_gem(fd);
+ }
+
+ for (sync = 2; sync--; ) {
+ igt_subtest_f("cpu%s", sync ? "-sync":"")
+ test_streaming(fd, 0, sync);
+ igt_subtest_f("gtt%s", sync ? "-sync":"")
+ test_streaming(fd, 1, sync);
+ igt_subtest_f("wc%s", sync ? "-sync":"")
+ test_streaming(fd, 2, sync);
+ }
+
+ igt_subtest("batch-cpu")
+ test_batch(fd, 0, 0);
+ igt_subtest("batch-gtt")
+ test_batch(fd, 1, 0);
+ igt_subtest("batch-wc")
+ test_batch(fd, 2, 0);
+ igt_subtest("batch-reverse-cpu")
+ test_batch(fd, 0, 1);
+ igt_subtest("batch-reverse-gtt")
+ test_batch(fd, 1, 1);
+ igt_subtest("batch-reverse-wc")
+ test_batch(fd, 2, 1);
+
+ igt_fixture
+ close(fd);
+}