diff options
author | Arkadiusz Hiler <arkadiusz.hiler@intel.com> | 2018-10-18 14:06:42 +0300 |
---|---|---|
committer | Arkadiusz Hiler <arkadiusz.hiler@intel.com> | 2018-10-23 10:55:51 +0300 |
commit | 741bf7064c467df725c14cc0b3b8b50436f9ee09 (patch) | |
tree | 0ad6fb217dca79a8f1175fb289979b574222fefa /tests/i915/gem_streaming_writes.c | |
parent | 78619fde4008424c472906041edb1d204e014f7c (diff) |
tests: Introduce i915 directory
We can already move all the tests with distinct prefixes: gem_, gen3_
and i915_.
pm_ and drv_ tests will follow in batches, so we can do the
adjustments in the reporting/filtering layer of the CI system.
v2: Fix test-list.txt generation with meson
v3: Fix docs build (Petri)
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Petri Latvala <petri.latvala@intel.com>
Cc: Martin Peres <martin.peres@linux.intel.com>
Signed-off-by: Arkadiusz Hiler <arkadiusz.hiler@intel.com>
Reviewed-by: Petri Latvala <petri.latvala@intel.com>
Tested-by: Petri Latvala <petri.latvala@intel.com>
Acked-by: Daniel Vetter <daniel.vetter@ffwll.ch>
Acked-by: Chris Wilson <chris@chris-wilson.co.uk>
Diffstat (limited to 'tests/i915/gem_streaming_writes.c')
-rw-r--r-- | tests/i915/gem_streaming_writes.c | 405 |
1 files changed, 405 insertions, 0 deletions
diff --git a/tests/i915/gem_streaming_writes.c b/tests/i915/gem_streaming_writes.c new file mode 100644 index 00000000..e83d69de --- /dev/null +++ b/tests/i915/gem_streaming_writes.c @@ -0,0 +1,405 @@ +/* + * Copyright © 2015 Intel Corporation + * + * Permission is hereby granted, free of charge, to any person obtaining a + * copy of this software and associated documentation files (the "Software"), + * to deal in the Software without restriction, including without limitation + * the rights to use, copy, modify, merge, publish, distribute, sublicense, + * and/or sell copies of the Software, and to permit persons to whom the + * Software is furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice (including the next + * paragraph) shall be included in all copies or substantial portions of the + * Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS + * IN THE SOFTWARE. + * + * Authors: + * Chris Wilson <chris@chris-wilson.co.uk> + * + */ + +#include "igt.h" +#include <unistd.h> +#include <stdlib.h> +#include <stdio.h> +#include <string.h> +#include <fcntl.h> +#include <inttypes.h> +#include <pthread.h> +#include <errno.h> +#include <sys/stat.h> +#include <sys/ioctl.h> +#include "drm.h" + +#define OBJECT_SIZE 1024*1024 +#define CHUNK_SIZE 32 + +#define COPY_BLT_CMD (2<<29|0x53<<22|0x6) +#define BLT_WRITE_ALPHA (1<<21) +#define BLT_WRITE_RGB (1<<20) +#define BLT_WRITE_ARGB (BLT_WRITE_ALPHA | BLT_WRITE_RGB) + +#define LOCAL_I915_EXEC_HANDLE_LUT (1<<12) + +IGT_TEST_DESCRIPTION("Test of streaming writes into active GPU sources"); + +#define SRC 0 +#define DST 1 +#define BATCH 2 + +#define src exec[SRC].handle +#define src_offset exec[SRC].offset +#define dst exec[DST].handle +#define dst_offset exec[DST].offset + +static void test_streaming(int fd, int mode, int sync) +{ + const int has_64bit_reloc = intel_gen(intel_get_drm_devid(fd)) >= 8; + struct drm_i915_gem_execbuffer2 execbuf; + struct drm_i915_gem_exec_object2 exec[3]; + struct drm_i915_gem_relocation_entry reloc[128]; + uint32_t tmp[] = { MI_BATCH_BUFFER_END }; + uint64_t __src_offset, __dst_offset; + uint32_t *s, *d; + uint32_t offset; + struct { + uint32_t handle; + uint64_t offset; + } *batch; + int i, n; + + memset(exec, 0, sizeof(exec)); + exec[SRC].handle = gem_create(fd, OBJECT_SIZE); + exec[DST].handle = gem_create(fd, OBJECT_SIZE); + + switch (mode) { + case 0: /* cpu/snoop */ + gem_set_caching(fd, src, I915_CACHING_CACHED); + s = gem_mmap__cpu(fd, src, 0, OBJECT_SIZE, + PROT_READ | PROT_WRITE); + break; + case 1: /* gtt */ + s = gem_mmap__gtt(fd, src, OBJECT_SIZE, + PROT_READ | PROT_WRITE); + break; + case 2: /* wc */ + s = gem_mmap__wc(fd, src, 0, OBJECT_SIZE, + PROT_READ | PROT_WRITE); + break; + } + *s = 0; /* fault the object into the mappable range first (for GTT) */ + + d = gem_mmap__cpu(fd, dst, 0, OBJECT_SIZE, PROT_READ); + + gem_write(fd, dst, 0, tmp, sizeof(tmp)); + memset(&execbuf, 0, sizeof(execbuf)); + execbuf.buffers_ptr = to_user_pointer(exec); + execbuf.buffer_count = 2; + execbuf.flags = LOCAL_I915_EXEC_HANDLE_LUT; + if (__gem_execbuf(fd, &execbuf)) { + execbuf.flags = 0; + igt_require(__gem_execbuf(fd, &execbuf) == 0); + } + /* We assume that the active objects are fixed to avoid relocations */ + __src_offset = src_offset; + __dst_offset = dst_offset; + + memset(reloc, 0, sizeof(reloc)); + for (i = 0; i < 64; i++) { + reloc[2*i+0].offset = 64*i + 4 * sizeof(uint32_t); + reloc[2*i+0].delta = 0; + reloc[2*i+0].target_handle = execbuf.flags & LOCAL_I915_EXEC_HANDLE_LUT ? DST : dst; + reloc[2*i+0].presumed_offset = dst_offset; + reloc[2*i+0].read_domains = I915_GEM_DOMAIN_RENDER; + reloc[2*i+0].write_domain = I915_GEM_DOMAIN_RENDER; + + reloc[2*i+1].offset = 64*i + 7 * sizeof(uint32_t); + if (has_64bit_reloc) + reloc[2*i+1].offset += sizeof(uint32_t); + reloc[2*i+1].delta = 0; + reloc[2*i+1].target_handle = execbuf.flags & LOCAL_I915_EXEC_HANDLE_LUT ? SRC : src; + reloc[2*i+1].presumed_offset = src_offset; + reloc[2*i+1].read_domains = I915_GEM_DOMAIN_RENDER; + reloc[2*i+1].write_domain = 0; + } + gem_execbuf(fd, &execbuf); + igt_assert_eq_u64(__src_offset, src_offset); + igt_assert_eq_u64(__dst_offset, dst_offset); + + exec[DST].flags = EXEC_OBJECT_WRITE; + exec[BATCH].relocation_count = 2; + execbuf.buffer_count = 3; + execbuf.flags |= I915_EXEC_NO_RELOC; + if (gem_has_blt(fd)) + execbuf.flags |= I915_EXEC_BLT; + + batch = malloc(sizeof(*batch) * (OBJECT_SIZE / CHUNK_SIZE / 64)); + for (i = n = 0; i < OBJECT_SIZE / CHUNK_SIZE / 64; i++) { + uint32_t *base; + + batch[i].handle = gem_create(fd, 4096); + batch[i].offset = 0; + + base = gem_mmap__cpu(fd, batch[i].handle, 0, 4096, PROT_WRITE); + gem_set_domain(fd, batch[i].handle, + I915_GEM_DOMAIN_CPU, I915_GEM_DOMAIN_CPU); + + for (int j = 0; j < 64; j++) { + unsigned x = (n * CHUNK_SIZE) % 4096 >> 2; + unsigned y = (n * CHUNK_SIZE) / 4096; + uint32_t *b = base + 16 * j; + int k = 0; + + b[k] = COPY_BLT_CMD | BLT_WRITE_ARGB; + if (has_64bit_reloc) + b[k] += 2; + k++; + b[k++] = 0xcc << 16 | 1 << 25 | 1 << 24 | 4096; + b[k++] = (y << 16) | x; + b[k++] = ((y+1) << 16) | (x + (CHUNK_SIZE >> 2)); + b[k++] = dst_offset; + if (has_64bit_reloc) + b[k++] = dst_offset >> 32; + b[k++] = (y << 16) | x; + b[k++] = 4096; + b[k++] = src_offset; + if (has_64bit_reloc) + b[k++] = src_offset >> 32; + b[k++] = MI_BATCH_BUFFER_END; + + n++; + } + + munmap(base, 4096); + } + + for (int pass = 0; pass < 256; pass++) { + int domain = mode ? I915_GEM_DOMAIN_GTT : I915_GEM_DOMAIN_CPU; + gem_set_domain(fd, src, domain, domain); + + if (pass == 0) { + for (i = 0; i < OBJECT_SIZE/4; i++) + s[i] = i; + } + + /* Now copy from the src to the dst in 32byte chunks */ + for (offset = 0; offset < OBJECT_SIZE; offset += CHUNK_SIZE) { + int b; + + if (pass) { + if (sync) + gem_set_domain(fd, src, domain, domain); + for (i = 0; i < CHUNK_SIZE/4; i++) + s[offset/4 + i] = (OBJECT_SIZE*pass + offset)/4 + i; + } + + igt_assert(exec[DST].flags & EXEC_OBJECT_WRITE); + + b = offset / CHUNK_SIZE / 64; + n = offset / CHUNK_SIZE % 64; + exec[BATCH].relocs_ptr = to_user_pointer((reloc + 2*n)); + exec[BATCH].handle = batch[b].handle; + exec[BATCH].offset = batch[b].offset; + execbuf.batch_start_offset = 64*n; + + gem_execbuf(fd, &execbuf); + igt_assert_eq_u64(__src_offset, src_offset); + igt_assert_eq_u64(__dst_offset, dst_offset); + + batch[b].offset = exec[BATCH].offset; + } + + gem_set_domain(fd, dst, I915_GEM_DOMAIN_CPU, 0); + for (offset = 0; offset < OBJECT_SIZE/4; offset++) + igt_assert_eq(pass*OBJECT_SIZE/4 + offset, d[offset]); + } + + for (i = 0; i < OBJECT_SIZE / CHUNK_SIZE / 64; i++) + gem_close(fd, batch[i].handle); + free(batch); + + munmap(s, OBJECT_SIZE); + gem_close(fd, src); + munmap(d, OBJECT_SIZE); + gem_close(fd, dst); +} + +static void test_batch(int fd, int mode, int reverse) +{ + const int has_64bit_reloc = intel_gen(intel_get_drm_devid(fd)) >= 8; + struct drm_i915_gem_execbuffer2 execbuf; + struct drm_i915_gem_exec_object2 exec[3]; + struct drm_i915_gem_relocation_entry reloc[2]; + uint32_t tmp[] = { MI_BATCH_BUFFER_END }; + uint64_t __src_offset, __dst_offset; + bool need_64b_start_offset = true; + uint64_t batch_size; + uint32_t *s, *d; + uint32_t *base; + uint32_t offset; + + memset(exec, 0, sizeof(exec)); + exec[DST].handle = gem_create(fd, OBJECT_SIZE); + exec[SRC].handle = gem_create(fd, OBJECT_SIZE); + + s = gem_mmap__wc(fd, src, 0, OBJECT_SIZE, PROT_READ | PROT_WRITE); + + d = gem_mmap__cpu(fd, dst, 0, OBJECT_SIZE, PROT_READ); + + memset(reloc, 0, sizeof(reloc)); + reloc[0].offset = 4 * sizeof(uint32_t); + reloc[0].delta = 0; + reloc[0].target_handle = execbuf.flags & LOCAL_I915_EXEC_HANDLE_LUT ? DST : dst; + reloc[0].presumed_offset = dst_offset; + reloc[0].read_domains = I915_GEM_DOMAIN_RENDER; + reloc[0].write_domain = I915_GEM_DOMAIN_RENDER; + + reloc[1].offset = 7 * sizeof(uint32_t); + if (has_64bit_reloc) + reloc[1].offset += sizeof(uint32_t); + reloc[1].delta = 0; + reloc[1].target_handle = execbuf.flags & LOCAL_I915_EXEC_HANDLE_LUT ? SRC : src; + reloc[1].presumed_offset = src_offset; + reloc[1].read_domains = I915_GEM_DOMAIN_RENDER; + reloc[1].write_domain = 0; + + batch_size = ALIGN(OBJECT_SIZE / CHUNK_SIZE * 128, 4096); + exec[BATCH].relocs_ptr = to_user_pointer(reloc); + exec[BATCH].relocation_count = 2; + exec[BATCH].handle = gem_create(fd, batch_size); + + switch (mode) { + case 0: /* cpu/snoop */ + igt_require(gem_has_llc(fd)); + base = gem_mmap__cpu(fd, exec[BATCH].handle, 0, batch_size, + PROT_READ | PROT_WRITE); + break; + case 1: /* gtt */ + base = gem_mmap__gtt(fd, exec[BATCH].handle, batch_size, + PROT_READ | PROT_WRITE); + break; + case 2: /* wc */ + base = gem_mmap__wc(fd, exec[BATCH].handle, 0, batch_size, + PROT_READ | PROT_WRITE); + break; + } + *base = 0; /* fault the object into the mappable range first */ + + gem_write(fd, exec[BATCH].handle, 0, tmp, sizeof(tmp)); + memset(&execbuf, 0, sizeof(execbuf)); + execbuf.buffers_ptr = to_user_pointer(exec); + execbuf.buffer_count = 3; + execbuf.flags = LOCAL_I915_EXEC_HANDLE_LUT; + if (gem_has_blt(fd)) + execbuf.flags |= I915_EXEC_BLT; + if (__gem_execbuf(fd, &execbuf)) { + execbuf.flags &= ~LOCAL_I915_EXEC_HANDLE_LUT; + gem_execbuf(fd, &execbuf); + } + execbuf.flags |= I915_EXEC_NO_RELOC; + exec[DST].flags = EXEC_OBJECT_WRITE; + /* We assume that the active objects are fixed to avoid relocations */ + exec[BATCH].relocation_count = 0; + __src_offset = src_offset; + __dst_offset = dst_offset; + + offset = mode ? I915_GEM_DOMAIN_GTT : I915_GEM_DOMAIN_CPU; + gem_set_domain(fd, exec[BATCH].handle, offset, offset); + for (int pass = 0; pass < 256; pass++) { + gem_set_domain(fd, src, I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT); + for (offset = 0; offset < OBJECT_SIZE/4; offset++) + s[offset] = OBJECT_SIZE*pass/4 + offset; + + /* Now copy from the src to the dst in 32byte chunks */ + for (offset = 0; offset < OBJECT_SIZE / CHUNK_SIZE; offset++) { + unsigned x = (offset * CHUNK_SIZE) % 4096 >> 2; + unsigned y = (offset * CHUNK_SIZE) / 4096; + int k; + + execbuf.batch_start_offset = 128 * offset; + if (!need_64b_start_offset) + execbuf.batch_start_offset += 8 * (pass & 7); + igt_assert(execbuf.batch_start_offset <= batch_size - 64); + if (reverse) + execbuf.batch_start_offset = batch_size - execbuf.batch_start_offset - 64; + igt_assert(execbuf.batch_start_offset <= batch_size - 64); + k = execbuf.batch_start_offset / 4; + + base[k] = COPY_BLT_CMD | BLT_WRITE_ARGB; + if (has_64bit_reloc) + base[k] += 2; + k++; + base[k++] = 0xcc << 16 | 1 << 25 | 1 << 24 | 4096; + base[k++] = (y << 16) | x; + base[k++] = ((y+1) << 16) | (x + (CHUNK_SIZE >> 2)); + base[k++] = dst_offset; + if (has_64bit_reloc) + base[k++] = dst_offset >> 32; + base[k++] = (y << 16) | x; + base[k++] = 4096; + base[k++] = src_offset; + if (has_64bit_reloc) + base[k++] = src_offset >> 32; + base[k++] = MI_BATCH_BUFFER_END; + + igt_assert(exec[DST].flags & EXEC_OBJECT_WRITE); + gem_execbuf(fd, &execbuf); + igt_assert_eq_u64(__src_offset, src_offset); + igt_assert_eq_u64(__dst_offset, dst_offset); + } + + gem_set_domain(fd, dst, I915_GEM_DOMAIN_CPU, 0); + for (offset = 0; offset < OBJECT_SIZE/4; offset++) + igt_assert_eq(pass*OBJECT_SIZE/4 + offset, d[offset]); + } + + munmap(base, OBJECT_SIZE / CHUNK_SIZE * 128); + gem_close(fd, exec[BATCH].handle); + + munmap(s, OBJECT_SIZE); + gem_close(fd, src); + munmap(d, OBJECT_SIZE); + gem_close(fd, dst); +} + +igt_main +{ + int fd, sync; + + igt_fixture { + fd = drm_open_driver(DRIVER_INTEL); + igt_require_gem(fd); + } + + for (sync = 2; sync--; ) { + igt_subtest_f("cpu%s", sync ? "-sync":"") + test_streaming(fd, 0, sync); + igt_subtest_f("gtt%s", sync ? "-sync":"") + test_streaming(fd, 1, sync); + igt_subtest_f("wc%s", sync ? "-sync":"") + test_streaming(fd, 2, sync); + } + + igt_subtest("batch-cpu") + test_batch(fd, 0, 0); + igt_subtest("batch-gtt") + test_batch(fd, 1, 0); + igt_subtest("batch-wc") + test_batch(fd, 2, 0); + igt_subtest("batch-reverse-cpu") + test_batch(fd, 0, 1); + igt_subtest("batch-reverse-gtt") + test_batch(fd, 1, 1); + igt_subtest("batch-reverse-wc") + test_batch(fd, 2, 1); + + igt_fixture + close(fd); +} |