summaryrefslogtreecommitdiff
path: root/tests/i915/gem_ring_sync_copy.c
diff options
context:
space:
mode:
authorArkadiusz Hiler <arkadiusz.hiler@intel.com>2018-10-18 14:06:42 +0300
committerArkadiusz Hiler <arkadiusz.hiler@intel.com>2018-10-23 10:55:51 +0300
commit741bf7064c467df725c14cc0b3b8b50436f9ee09 (patch)
tree0ad6fb217dca79a8f1175fb289979b574222fefa /tests/i915/gem_ring_sync_copy.c
parent78619fde4008424c472906041edb1d204e014f7c (diff)
tests: Introduce i915 directory
We can already move all the tests with distinct prefixes: gem_, gen3_ and i915_. pm_ and drv_ tests will follow in batches, so we can do the adjustments in the reporting/filtering layer of the CI system. v2: Fix test-list.txt generation with meson v3: Fix docs build (Petri) Cc: Chris Wilson <chris@chris-wilson.co.uk> Cc: Petri Latvala <petri.latvala@intel.com> Cc: Martin Peres <martin.peres@linux.intel.com> Signed-off-by: Arkadiusz Hiler <arkadiusz.hiler@intel.com> Reviewed-by: Petri Latvala <petri.latvala@intel.com> Tested-by: Petri Latvala <petri.latvala@intel.com> Acked-by: Daniel Vetter <daniel.vetter@ffwll.ch> Acked-by: Chris Wilson <chris@chris-wilson.co.uk>
Diffstat (limited to 'tests/i915/gem_ring_sync_copy.c')
-rw-r--r--tests/i915/gem_ring_sync_copy.c374
1 files changed, 374 insertions, 0 deletions
diff --git a/tests/i915/gem_ring_sync_copy.c b/tests/i915/gem_ring_sync_copy.c
new file mode 100644
index 00000000..8d372355
--- /dev/null
+++ b/tests/i915/gem_ring_sync_copy.c
@@ -0,0 +1,374 @@
+/*
+ * Copyright © 2013 Intel Corporation
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * the rights to use, copy, modify, merge, publish, distribute, sublicense,
+ * and/or sell copies of the Software, and to permit persons to whom the
+ * Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice (including the next
+ * paragraph) shall be included in all copies or substantial portions of the
+ * Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
+ * IN THE SOFTWARE.
+ *
+ * Authors:
+ * Damien Lespiau <damien.lespiau@intel.com>
+ */
+
+/*
+ * The goal of this test is to ensure that we respect inter ring dependencies
+ *
+ * For each pair of rings R1, R2 where we have copy support (i.e. blt,
+ * rendercpy and mediafill) do:
+ * - Throw a busy load onto R1. gem_concurrent_blt just uses lots of buffers
+ * for this effect.
+ * - Fill three buffers A, B, C with unique data.
+ * - Copy A to B on ring R1
+ *
+ * Then come the three different variants.
+ * - Copy B to C on ring R2, check that C now contains what A originally
+ * contained. This is the write->read hazard. gem_concurrent_blt calls this
+ * early read.
+ * - Copy C to A on ring R2, check that B now contains what A originally
+ * contained. This is the read->write hazard, gem_concurrent_blt calls it
+ * overwrite_source.
+ * - Copy C to B on ring R2 and check that B contains what C originally
+ * contained. This is the write/write hazard. gem_concurrent_blt doesn't
+ * have that since for the cpu case it's too boring.
+ *
+ */
+
+#include "igt.h"
+#include <stdlib.h>
+#include <stdbool.h>
+
+
+IGT_TEST_DESCRIPTION("Ensure inter-ring dependencies are respected.");
+
+#define WIDTH 512
+#define HEIGHT 512
+#define NUM_BUSY_BUFFERS 32
+
+typedef struct {
+ int drm_fd;
+ uint32_t devid;
+ drm_intel_bufmgr *bufmgr;
+ struct intel_batchbuffer *batch;
+
+ /* number of buffers to keep the ring busy for a while */
+ unsigned int n_buffers_load;
+
+ uint32_t linear[WIDTH * HEIGHT];
+
+ struct {
+ igt_render_copyfunc_t copy;
+ struct igt_buf *srcs;
+ struct igt_buf *dsts;
+ } render;
+
+ struct {
+ drm_intel_bo **srcs;
+ drm_intel_bo **dsts;
+ } blitter;
+
+} data_t;
+
+enum ring {
+ RENDER,
+ BLITTER,
+};
+
+enum test {
+ TEST_WRITE_READ,
+ TEST_READ_WRITE,
+ TEST_WRITE_WRITE,
+};
+
+static const char *ring_name(enum ring ring)
+{
+ const char *names[] = {
+ "render",
+ "blitter",
+ };
+
+ return names[ring];
+}
+
+static drm_intel_bo *bo_create(data_t *data, int width, int height, int val)
+{
+ drm_intel_bo *bo;
+ int i;
+
+ bo = drm_intel_bo_alloc(data->bufmgr, "", 4 * width * height, 4096);
+ igt_assert(bo);
+
+ for (i = 0; i < width * height; i++)
+ data->linear[i] = val;
+ gem_write(data->drm_fd, bo->handle, 0, data->linear,
+ sizeof(data->linear));
+
+ return bo;
+}
+
+static void bo_check(data_t *data, drm_intel_bo *bo, uint32_t val)
+{
+ int i;
+
+ gem_read(data->drm_fd, bo->handle, 0,
+ data->linear, sizeof(data->linear));
+ for (i = 0; i < WIDTH * HEIGHT; i++)
+ igt_assert_eq_u32(data->linear[i], val);
+}
+
+static void scratch_buf_init_from_bo(struct igt_buf *buf, drm_intel_bo *bo)
+{
+ memset(buf, 0, sizeof(*buf));
+
+ buf->bo = bo;
+ buf->stride = 4 * WIDTH;
+ buf->tiling = I915_TILING_NONE;
+ buf->size = 4 * WIDTH * HEIGHT;
+}
+
+static void scratch_buf_init(data_t *data, struct igt_buf *buf,
+ int width, int height, uint32_t color)
+{
+ drm_intel_bo *bo;
+
+ bo = bo_create(data, width, height, color);
+ scratch_buf_init_from_bo(buf, bo);
+}
+
+/*
+ * Provide a few ring specific vfuncs for run_test().
+ *
+ * busy() Queue a n_buffers_load workloads onto the ring to keep it busy
+ * busy_fini() Clean up after busy
+ * copy() Copy one BO to another
+ */
+
+/*
+ * Render ring
+ */
+
+static void render_busy(data_t *data)
+{
+ size_t array_size;
+ int i;
+
+ /* allocate 32 buffer objects and re-use them as needed */
+ array_size = NUM_BUSY_BUFFERS * sizeof(struct igt_buf);
+
+ data->render.srcs = malloc(array_size);
+ data->render.dsts = malloc(array_size);
+
+ for (i = 0; i < NUM_BUSY_BUFFERS; i++) {
+ scratch_buf_init(data, &data->render.srcs[i], WIDTH, HEIGHT,
+ 0xdeadbeef);
+ scratch_buf_init(data, &data->render.dsts[i], WIDTH, HEIGHT,
+ 0xdeadbeef);
+ }
+
+ for (i = 0; i < data->n_buffers_load; i++) {
+ data->render.copy(data->batch,
+ NULL, /* context */
+ &data->render.srcs[i % NUM_BUSY_BUFFERS],
+ 0, 0, /* src_x, src_y */
+ WIDTH, HEIGHT,
+ &data->render.dsts[i % NUM_BUSY_BUFFERS],
+ 0, 0 /* dst_x, dst_y */);
+ }
+}
+
+static void render_busy_fini(data_t *data)
+{
+ int i;
+
+ for (i = 0; i < NUM_BUSY_BUFFERS; i++) {
+ drm_intel_bo_unreference(data->render.srcs[i].bo);
+ drm_intel_bo_unreference(data->render.dsts[i].bo);
+ }
+
+ free(data->render.srcs);
+ free(data->render.dsts);
+ data->render.srcs = NULL;
+ data->render.dsts = NULL;
+}
+
+static void render_copy(data_t *data, drm_intel_bo *src, drm_intel_bo *dst)
+{
+ struct igt_buf src_buf, dst_buf;
+
+ scratch_buf_init_from_bo(&src_buf, src);
+ scratch_buf_init_from_bo(&dst_buf, dst);
+
+ data->render.copy(data->batch,
+ NULL, /* context */
+ &src_buf,
+ 0, 0, /* src_x, src_y */
+ WIDTH, HEIGHT,
+ &dst_buf,
+ 0, 0 /* dst_x, dst_y */);
+}
+
+/*
+ * Blitter ring
+ */
+
+static void blitter_busy(data_t *data)
+{
+ size_t array_size;
+ int i;
+
+ /* allocate 32 buffer objects and re-use them as needed */
+ array_size = NUM_BUSY_BUFFERS * sizeof(drm_intel_bo *);
+
+ data->blitter.srcs = malloc(array_size);
+ data->blitter.dsts = malloc(array_size);
+
+ for (i = 0; i < NUM_BUSY_BUFFERS; i++) {
+ data->blitter.srcs[i] = bo_create(data,
+ WIDTH, HEIGHT,
+ 0xdeadbeef);
+ data->blitter.dsts[i] = bo_create(data,
+ WIDTH, HEIGHT,
+ 0xdeadbeef);
+ }
+
+ for (i = 0; i < data->n_buffers_load; i++) {
+ intel_copy_bo(data->batch,
+ data->blitter.srcs[i % NUM_BUSY_BUFFERS],
+ data->blitter.dsts[i % NUM_BUSY_BUFFERS],
+ WIDTH*HEIGHT*4);
+ }
+}
+
+static void blitter_busy_fini(data_t *data)
+{
+ int i;
+
+ for (i = 0; i < NUM_BUSY_BUFFERS; i++) {
+ drm_intel_bo_unreference(data->blitter.srcs[i]);
+ drm_intel_bo_unreference(data->blitter.dsts[i]);
+ }
+
+ free(data->blitter.srcs);
+ free(data->blitter.dsts);
+ data->blitter.srcs = NULL;
+ data->blitter.dsts = NULL;
+}
+
+static void blitter_copy(data_t *data, drm_intel_bo *src, drm_intel_bo *dst)
+{
+ intel_copy_bo(data->batch, dst, src, WIDTH*HEIGHT*4);
+}
+
+struct ring_ops {
+ void (*busy)(data_t *data);
+ void (*busy_fini)(data_t *data);
+ void (*copy)(data_t *data, drm_intel_bo *src, drm_intel_bo *dst);
+} ops [] = {
+ {
+ .busy = render_busy,
+ .busy_fini = render_busy_fini,
+ .copy = render_copy,
+ },
+ {
+ .busy = blitter_busy,
+ .busy_fini = blitter_busy_fini,
+ .copy = blitter_copy,
+ },
+};
+
+static void run_test(data_t *data, enum ring r1, enum ring r2, enum test test)
+{
+ struct ring_ops *r1_ops = &ops[r1];
+ struct ring_ops *r2_ops = &ops[r2];
+ drm_intel_bo *a, *b, *c;
+
+ a = bo_create(data, WIDTH, HEIGHT, 0xa);
+ b = bo_create(data, WIDTH, HEIGHT, 0xb);
+ c = bo_create(data, WIDTH, HEIGHT, 0xc);
+
+ r1_ops->busy(data);
+ r1_ops->copy(data, a, b);
+
+ switch (test) {
+ case TEST_WRITE_READ:
+ r2_ops->copy(data, b, c);
+ bo_check(data, c, 0xa);
+ break;
+ case TEST_READ_WRITE:
+ r2_ops->copy(data, c, a);
+ bo_check(data, b, 0xa);
+ break;
+ case TEST_WRITE_WRITE:
+ r2_ops->copy(data, c, b);
+ bo_check(data, b, 0xc);
+ break;
+ default:
+ igt_fail(IGT_EXIT_FAILURE);
+ }
+
+ r1_ops->busy_fini(data);
+}
+
+igt_main
+{
+ data_t data = {0, };
+ int i;
+ struct combination {
+ int r1, r2;
+ } ring_combinations [] = {
+ { RENDER, BLITTER },
+ { BLITTER, RENDER },
+ };
+
+ igt_fixture {
+ data.drm_fd = drm_open_driver_render(DRIVER_INTEL);
+ data.devid = intel_get_drm_devid(data.drm_fd);
+
+ data.n_buffers_load = 1000;
+
+ data.bufmgr = drm_intel_bufmgr_gem_init(data.drm_fd, 4096);
+ igt_assert(data.bufmgr);
+ drm_intel_bufmgr_gem_enable_reuse(data.bufmgr);
+
+ data.render.copy = igt_get_render_copyfunc(data.devid);
+ igt_require_f(data.render.copy,
+ "no render-copy function\n");
+
+ data.batch = intel_batchbuffer_alloc(data.bufmgr, data.devid);
+ igt_assert(data.batch);
+ }
+
+ for (i = 0; i < ARRAY_SIZE(ring_combinations); i++) {
+ struct combination *c = &ring_combinations[i];
+
+ igt_subtest_f("sync-%s-%s-write-read",
+ ring_name(c->r1), ring_name(c->r2))
+ run_test(&data, c->r1, c->r2, TEST_WRITE_READ);
+
+ igt_subtest_f("sync-%s-%s-read-write",
+ ring_name(c->r1), ring_name(c->r2))
+ run_test(&data, c->r1, c->r2, TEST_READ_WRITE);
+ igt_subtest_f("sync-%s-%s-write-write",
+ ring_name(c->r1), ring_name(c->r2))
+ run_test(&data, c->r1, c->r2, TEST_WRITE_WRITE);
+ }
+
+ igt_fixture {
+ intel_batchbuffer_free(data.batch);
+ drm_intel_bufmgr_destroy(data.bufmgr);
+ close(data.drm_fd);
+ }
+}