summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorZbigniew Kempczyński <zbigniew.kempczynski@intel.com>2020-01-17 15:14:49 +0100
committerChris Wilson <chris@chris-wilson.co.uk>2020-01-17 21:12:42 +0000
commit621c5d83db5dc573dedcb9f5cc9aa88cb1553a61 (patch)
tree2e87d617fffa247b42e8244cef6d7065299e5e8f
parent1b2816124ec3dbd53b81725d39292f45d41d895b (diff)
lib/intel_bufops: Introduce buffer operations
Different GENs supports different tile surfaces. Older GENs have HW fences to allow X / Y surface tiling / detiling. Newer GENs have to tile / detile such surface in software. To make test developer life easier this code adds buffer operations (short buf_ops) to use appropriate functions allowing copying linear buffer to BO and from BO to linear buffer regardless GPU generation and tiling/swizzling within BO. For GENs having fences support preference is to use them if they are available (X / Y tiling is probed on buf_ops initialization). Note: on Gen2 code supports only HW tiling at the moment. Signed-off-by: Zbigniew Kempczyński <zbigniew.kempczynski@intel.com> Cc: Chris Wilson <chris@chris-wilson.co.uk> Cc: Katarzyna Dec <katarzyna.dec@intel.com> Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
-rw-r--r--docs/reference/igt-gpu-tools/igt-gpu-tools-docs.xml1
-rw-r--r--lib/Makefile.sources2
-rw-r--r--lib/intel_bufops.c1157
-rw-r--r--lib/intel_bufops.h86
-rw-r--r--lib/meson.build1
5 files changed, 1247 insertions, 0 deletions
diff --git a/docs/reference/igt-gpu-tools/igt-gpu-tools-docs.xml b/docs/reference/igt-gpu-tools/igt-gpu-tools-docs.xml
index 454d64e5..aa9fef20 100644
--- a/docs/reference/igt-gpu-tools/igt-gpu-tools-docs.xml
+++ b/docs/reference/igt-gpu-tools/igt-gpu-tools-docs.xml
@@ -42,6 +42,7 @@
<xi:include href="xml/igt_vgem.xml"/>
<xi:include href="xml/igt_x86.xml"/>
<xi:include href="xml/intel_batchbuffer.xml"/>
+ <xi:include href="xml/intel_bufops.xml"/>
<xi:include href="xml/intel_chipset.xml"/>
<xi:include href="xml/intel_io.xml"/>
<xi:include href="xml/ioctl_wrappers.xml"/>
diff --git a/lib/Makefile.sources b/lib/Makefile.sources
index 5dd3962e..da34eae7 100644
--- a/lib/Makefile.sources
+++ b/lib/Makefile.sources
@@ -67,6 +67,8 @@ lib_source_list = \
intel_aub.h \
intel_batchbuffer.c \
intel_batchbuffer.h \
+ intel_bufops.c \
+ intel_bufops.h \
intel_chipset.c \
intel_chipset.h \
intel_device_info.c \
diff --git a/lib/intel_bufops.c b/lib/intel_bufops.c
new file mode 100644
index 00000000..804b2a0a
--- /dev/null
+++ b/lib/intel_bufops.c
@@ -0,0 +1,1157 @@
+/*
+ * Copyright © 2020 Intel Corporation
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * the rights to use, copy, modify, merge, publish, distribute, sublicense,
+ * and/or sell copies of the Software, and to permit persons to whom the
+ * Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice (including the next
+ * paragraph) shall be included in all copies or substantial portions of the
+ * Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
+ * IN THE SOFTWARE.
+ *
+ */
+
+#include <sys/ioctl.h>
+#include "igt.h"
+#include "igt_x86.h"
+#include "intel_bufops.h"
+
+/**
+ * SECTION:intel_bufops
+ * @short_description: Buffer operation on tiled surfaces
+ * @title: Buffer operations
+ * @include: igt.h
+ *
+ * # Buffer operations
+ *
+ * Intel GPU devices supports different set of tiled surfaces.
+ * Checking each time what tile formats are supports is cumbersome and
+ * error prone.
+ *
+ * Buffer operation (buf_ops) provide a wrapper to conditional code which
+ * can be used without worrying of implementation details giving:
+ * - copy linear to tiled buffer
+ * - copy tiled buffer to linear
+ *
+ * Following code order should be used (linear is plain memory with some
+ * image data):
+ *
+ * |[<!-- language="c" -->
+ * struct buf_ops *bops;
+ * struct intel_buf ibuf;
+ * ...
+ * bops = buf_ops_create(fd);
+ * intel_buf_init(bops, &ibuf, 512, 512, 32, I915_TILING_X, false);
+ * ...
+ * linear_to_intel_buf(bops, &ibuf, linear);
+ * ...
+ * intel_buf_to_linear(bops, &ibuf, linear);
+ * ...
+ * intel_buf_close(bops, &ibuf);
+ * ...
+ * buf_ops_destroy(bops);
+ * ]|
+ *
+ * Calling buf_ops_create(fd) probes hardware capabilities (supported fences,
+ * swizzling) and returns opaque pointer to buf_ops. From now on
+ * intel_buf_to_linear() and linear_to_intel_buf() will choose appropriate
+ * function to do the job.
+ *
+ * Note: bufops doesn't support SW tiling code yet.
+ */
+
+//#define BUFOPS_DEBUGGING
+#ifdef BUFOPS_DEBUGGING
+#define DEBUG(...) printf(__VA_ARGS__)
+#define DEBUGFN() DEBUG("\t -> %s()\n", __FUNCTION__)
+#else
+#define DEBUG(...)
+#define DEBUGFN()
+#endif
+
+#define TILE_DEF(x) (1 << (x))
+#define TILE_NONE TILE_DEF(I915_TILING_NONE)
+#define TILE_X TILE_DEF(I915_TILING_X)
+#define TILE_Y TILE_DEF(I915_TILING_Y)
+#define TILE_Yf TILE_DEF(I915_TILING_Yf)
+#define TILE_Ys TILE_DEF(I915_TILING_Ys)
+
+#define CCS_OFFSET(buf) (buf->aux.offset)
+#define CCS_SIZE(gen, buf) \
+ (intel_buf_aux_width(gen, buf) * intel_buf_aux_height(gen, buf))
+
+typedef void (*bo_copy)(struct buf_ops *, struct intel_buf *, uint32_t *);
+
+struct buf_ops {
+ int fd;
+ int gen_start;
+ int gen_end;
+ int intel_gen;
+ uint32_t supported_tiles;
+ uint32_t supported_hw_tiles;
+ uint32_t swizzle_x;
+ uint32_t swizzle_y;
+ bo_copy linear_to;
+ bo_copy linear_to_x;
+ bo_copy linear_to_y;
+ bo_copy linear_to_yf;
+ bo_copy linear_to_ys;
+ bo_copy to_linear;
+ bo_copy x_to_linear;
+ bo_copy y_to_linear;
+ bo_copy yf_to_linear;
+ bo_copy ys_to_linear;
+};
+
+static const char *tiling_str(uint32_t tiling)
+{
+ switch (tiling) {
+ case I915_TILING_NONE: return "NONE";
+ case I915_TILING_X: return "X";
+ case I915_TILING_Y: return "Y";
+ case I915_TILING_Yf: return "Yf";
+ case I915_TILING_Ys: return "Ys";
+ default: return "UNKNOWN";
+ }
+}
+
+static const char *bool_str(bool v)
+{
+ return v ? "yes" : "no";
+}
+
+static inline bool is_hw_tiling_supported(struct buf_ops *bops, uint32_t tiling)
+{
+ return bops->supported_hw_tiles & TILE_DEF(tiling);
+}
+
+static inline bool is_tiling_supported(struct buf_ops *bops, uint32_t tiling)
+{
+ return bops->supported_tiles & TILE_DEF(tiling);
+}
+
+static int __gem_get_tiling(int fd, struct drm_i915_gem_get_tiling *arg)
+{
+ int err;
+
+ err = 0;
+ if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_GET_TILING, arg)) {
+ err = -errno;
+ igt_assume(err);
+ }
+ errno = 0;
+
+ return err;
+}
+
+static bool __get_tiling(int fd, uint32_t handle, uint32_t *tiling,
+ uint32_t *swizzle)
+{
+ struct drm_i915_gem_get_tiling get_tiling = { .handle = handle };
+
+ if (__gem_get_tiling(fd, &get_tiling) != 0)
+ return false;
+
+ *tiling = get_tiling.tiling_mode;
+ *swizzle = get_tiling.swizzle_mode;
+ igt_debug("buf tiling: %s, swizzle: %x, phys_swizzle: %x\n",
+ tiling_str(get_tiling.tiling_mode),
+ get_tiling.swizzle_mode,
+ get_tiling.phys_swizzle_mode);
+
+ return get_tiling.phys_swizzle_mode == get_tiling.swizzle_mode;
+}
+
+static int __set_tiling(int fd, uint32_t handle, uint32_t tiling,
+ uint32_t stride)
+{
+ do {
+ struct drm_i915_gem_set_tiling st;
+ int err;
+
+ st.handle = handle;
+ st.tiling_mode = tiling;
+ st.stride = tiling ? stride : 0;
+
+ err = 0;
+ if (ioctl(fd, DRM_IOCTL_I915_GEM_SET_TILING, &st))
+ err = -errno;
+ errno = 0;
+ if (err != -EINTR)
+ return err;
+ } while (1);
+}
+
+static void set_hw_tiled(struct buf_ops *bops, struct intel_buf *buf)
+{
+ if (buf->tiling != I915_TILING_X && buf->tiling != I915_TILING_Y)
+ return;
+
+ if (!buf_ops_has_hw_fence(bops, buf->tiling))
+ return;
+
+ igt_assert_eq(__set_tiling(bops->fd,
+ buf->handle, buf->tiling, buf->stride),
+ 0);
+}
+
+static unsigned long swizzle_bit(unsigned int bit, unsigned long offset)
+{
+ return (offset & (1ul << bit)) >> (bit - 6);
+}
+
+static unsigned long swizzle_addr(void *ptr, uint32_t swizzle)
+{
+ unsigned long addr = to_user_pointer(ptr);
+
+ switch (swizzle) {
+ case I915_BIT_6_SWIZZLE_NONE:
+ return addr;
+ case I915_BIT_6_SWIZZLE_9:
+ return addr ^ swizzle_bit(9, addr);
+ case I915_BIT_6_SWIZZLE_9_10:
+ return addr ^ swizzle_bit(9, addr) ^ swizzle_bit(10, addr);
+ case I915_BIT_6_SWIZZLE_9_11:
+ return addr ^ swizzle_bit(9, addr) ^ swizzle_bit(11, addr);
+ case I915_BIT_6_SWIZZLE_9_10_11:
+ return (addr ^
+ swizzle_bit(9, addr) ^
+ swizzle_bit(10, addr) ^
+ swizzle_bit(11, addr));
+
+ case I915_BIT_6_SWIZZLE_UNKNOWN:
+ case I915_BIT_6_SWIZZLE_9_17:
+ case I915_BIT_6_SWIZZLE_9_10_17:
+ default:
+ igt_skip("physical swizzling mode impossible to handle in userspace\n");
+ return addr;
+ }
+}
+
+static void *x_ptr(void *ptr,
+ unsigned int x, unsigned int y,
+ unsigned int stride, unsigned int cpp)
+{
+ const int tile_width = 512;
+ const int tile_height = 8;
+ const int tile_size = tile_width * tile_height;
+ int offset_x, offset_y, pos;
+ int tile_x, tile_y;
+
+ x *= cpp;
+ tile_x = x / tile_width;
+ tile_y = y / tile_height;
+ offset_x = tile_x * tile_size;
+ offset_y = tile_y * stride * tile_height;
+
+ pos = (offset_y + (y % tile_height * tile_width) +
+ offset_x + (x % tile_width));
+
+ return ptr + pos;
+}
+
+static void *y_ptr(void *ptr,
+ unsigned int x, unsigned int y,
+ unsigned int stride, unsigned int cpp)
+{
+ const int tile_width = 128;
+ const int tile_height = 32;
+ const int owords = 16;
+ const int tile_size = tile_width * tile_height;
+ int offset_x, offset_y, pos;
+ int shift_x, shift_y;
+ int tile_x, tile_y;
+
+ x *= cpp;
+ tile_x = x / tile_width;
+ tile_y = y / tile_height;
+ offset_x = tile_x * tile_size;
+ offset_y = tile_y * stride * tile_height;
+ shift_x = x % owords + (x % tile_width) / owords * tile_width * cpp;
+ shift_y = y % tile_height * owords;
+
+ pos = offset_y + offset_x + shift_x + shift_y;
+
+ return ptr + pos;
+}
+
+static void *yf_ptr(void *ptr,
+ unsigned int x, unsigned int y,
+ unsigned int stride, unsigned int cpp)
+{
+ const int tile_size = 4 * 1024;
+ const int tile_width = 128;
+ int row_size = stride / tile_width * tile_size;
+
+ x *= cpp; /* convert to Byte offset */
+
+ /*
+ * Within a 4k Yf tile, the byte swizzling pattern is
+ * msb......lsb
+ * xyxyxyyyxxxx
+ * The tiles themselves are laid out in row major order.
+ */
+ return ptr +
+ ((x & 0xf) * 1) + /* 4x1 pixels(32bpp) = 16B */
+ ((y & 0x3) * 16) + /* 4x4 pixels = 64B */
+ (((y & 0x4) >> 2) * 64) + /* 1x2 64B blocks */
+ (((x & 0x10) >> 4) * 128) + /* 2x2 64B blocks = 256B block */
+ (((y & 0x8) >> 3) * 256) + /* 2x1 256B blocks */
+ (((x & 0x20) >> 5) * 512) + /* 2x2 256B blocks */
+ (((y & 0x10) >> 4) * 1024) + /* 4x2 256 blocks */
+ (((x & 0x40) >> 6) * 2048) + /* 4x4 256B blocks = 4k tile */
+ (((x & ~0x7f) >> 7) * tile_size) + /* row of tiles */
+ (((y & ~0x1f) >> 5) * row_size);
+}
+
+typedef void *(*tile_fn)(void *, unsigned int, unsigned int,
+ unsigned int, unsigned int);
+static tile_fn __get_tile_fn_ptr(int tiling)
+{
+ tile_fn fn = NULL;
+
+ switch (tiling) {
+ case I915_TILING_X:
+ fn = x_ptr;
+ break;
+ case I915_TILING_Y:
+ fn = y_ptr;
+ break;
+ case I915_TILING_Yf:
+ fn = yf_ptr;
+ break;
+ case I915_TILING_Ys:
+ /* To be implemented */
+ break;
+ }
+
+ igt_require_f(fn, "Can't find tile function for tiling: %d\n", tiling);
+ return fn;
+}
+
+static bool is_cache_coherent(int fd, uint32_t handle)
+{
+ return gem_get_caching(fd, handle) != I915_CACHING_NONE;
+}
+
+enum ccs_copy_direction {
+ CCS_LINEAR_TO_BUF,
+ CCS_BUF_TO_LINEAR,
+};
+
+static void __copy_ccs(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear, enum ccs_copy_direction dir)
+{
+ uint64_t size, offset, ccs_size;
+ void *map;
+ int gen;
+
+ if (!buf->compression)
+ return;
+
+ gen = bops->intel_gen;
+ offset = CCS_OFFSET(buf);
+ ccs_size = CCS_SIZE(gen, buf);
+ size = offset + ccs_size;
+
+ map = __gem_mmap_offset__wc(bops->fd, buf->handle, 0, size,
+ PROT_READ | PROT_WRITE);
+ if (!map)
+ map = gem_mmap__wc(bops->fd, buf->handle, 0, size,
+ PROT_READ | PROT_WRITE);
+
+ switch (dir) {
+ case CCS_LINEAR_TO_BUF:
+ gem_set_domain(bops->fd, buf->handle,
+ I915_GEM_DOMAIN_WC, I915_GEM_DOMAIN_WC);
+ igt_memcpy_from_wc(map + offset, (uint8_t *) linear + offset,
+ ccs_size);
+ case CCS_BUF_TO_LINEAR:
+ gem_set_domain(bops->fd, buf->handle, I915_GEM_DOMAIN_WC, 0);
+ igt_memcpy_from_wc((uint8_t *) linear + offset, map + offset,
+ ccs_size);
+ }
+
+ munmap(map, size);
+}
+
+static void *mmap_write(int fd, struct intel_buf *buf)
+{
+ void *map = NULL;
+
+ if (is_cache_coherent(fd, buf->handle)) {
+ map = __gem_mmap_offset__cpu(fd, buf->handle, 0, buf->size,
+ PROT_READ | PROT_WRITE);
+ if (!map)
+ map = __gem_mmap__cpu(fd, buf->handle, 0, buf->size,
+ PROT_READ | PROT_WRITE);
+
+ if (map)
+ gem_set_domain(fd, buf->handle,
+ I915_GEM_DOMAIN_CPU,
+ I915_GEM_DOMAIN_CPU);
+ }
+
+ if (!map) {
+ map = __gem_mmap_offset__wc(fd, buf->handle, 0, buf->size,
+ PROT_READ | PROT_WRITE);
+ if (!map)
+ map = gem_mmap__wc(fd, buf->handle, 0, buf->size,
+ PROT_READ | PROT_WRITE);
+
+ gem_set_domain(fd, buf->handle,
+ I915_GEM_DOMAIN_WC, I915_GEM_DOMAIN_WC);
+ }
+
+ return map;
+}
+
+static void *mmap_read(int fd, struct intel_buf *buf)
+{
+ void *map = NULL;
+
+ if (gem_has_llc(fd) || is_cache_coherent(fd, buf->handle)) {
+ map = __gem_mmap_offset__cpu(fd, buf->handle, 0,
+ buf->size, PROT_READ);
+ if (!map)
+ map = __gem_mmap__cpu(fd, buf->handle, 0, buf->size,
+ PROT_READ);
+
+ if (map)
+ gem_set_domain(fd, buf->handle, I915_GEM_DOMAIN_CPU, 0);
+ }
+
+ if (!map) {
+ map = __gem_mmap_offset__wc(fd, buf->handle, 0, buf->size,
+ PROT_READ);
+ if (!map)
+ map = gem_mmap__wc(fd, buf->handle, 0, buf->size,
+ PROT_READ);
+
+ gem_set_domain(fd, buf->handle, I915_GEM_DOMAIN_WC, 0);
+ }
+
+ return map;
+}
+
+static void __copy_linear_to(int fd, struct intel_buf *buf,
+ const uint32_t *linear,
+ int tiling, uint32_t swizzle)
+{
+ const tile_fn fn = __get_tile_fn_ptr(tiling);
+ int height = intel_buf_height(buf);
+ int width = intel_buf_width(buf);
+ void *map = mmap_write(fd, buf);
+
+ for (int y = 0; y < height; y++) {
+ for (int x = 0; x < width; x++) {
+ uint32_t *ptr = fn(map, x, y, buf->stride, buf->bpp/8);
+
+ if (swizzle)
+ ptr = from_user_pointer(swizzle_addr(ptr,
+ swizzle));
+ *ptr = linear[y * width + x];
+ }
+ }
+
+ munmap(map, buf->size);
+}
+
+static void copy_linear_to_x(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ DEBUGFN();
+ __copy_linear_to(bops->fd, buf, linear, I915_TILING_X, bops->swizzle_x);
+}
+
+static void copy_linear_to_y(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ DEBUGFN();
+ __copy_linear_to(bops->fd, buf, linear, I915_TILING_Y, bops->swizzle_y);
+}
+
+static void copy_linear_to_yf(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ DEBUGFN();
+ __copy_linear_to(bops->fd, buf, linear, I915_TILING_Yf, 0);
+}
+
+static void copy_linear_to_ys(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ DEBUGFN();
+ __copy_linear_to(bops->fd, buf, linear, I915_TILING_Ys, 0);
+}
+
+static void __copy_to_linear(int fd, struct intel_buf *buf,
+ uint32_t *linear, int tiling, uint32_t swizzle)
+{
+ const tile_fn fn = __get_tile_fn_ptr(tiling);
+ int height = intel_buf_height(buf);
+ int width = intel_buf_width(buf);
+ void *map = mmap_write(fd, buf);
+
+ for (int y = 0; y < height; y++) {
+ for (int x = 0; x < width; x++) {
+ uint32_t *ptr = fn(map, x, y, buf->stride, buf->bpp/8);
+
+ if (swizzle)
+ ptr = from_user_pointer(swizzle_addr(ptr,
+ swizzle));
+ linear[y * width + x] = *ptr;
+ }
+ }
+
+ munmap(map, buf->size);
+}
+
+static void copy_x_to_linear(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ DEBUGFN();
+ __copy_to_linear(bops->fd, buf, linear, I915_TILING_X, bops->swizzle_x);
+}
+
+static void copy_y_to_linear(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ DEBUGFN();
+ __copy_to_linear(bops->fd, buf, linear, I915_TILING_Y, bops->swizzle_y);
+}
+
+static void copy_yf_to_linear(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ DEBUGFN();
+ __copy_to_linear(bops->fd, buf, linear, I915_TILING_Yf, 0);
+}
+
+static void copy_ys_to_linear(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ DEBUGFN();
+ __copy_to_linear(bops->fd, buf, linear, I915_TILING_Ys, 0);
+}
+
+static void copy_linear_to_gtt(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ void *map;
+
+ DEBUGFN();
+
+ map = gem_mmap__gtt(bops->fd, buf->handle,
+ buf->size, PROT_READ | PROT_WRITE);
+
+ gem_set_domain(bops->fd, buf->handle,
+ I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT);
+
+ memcpy(map, linear, buf->size);
+
+ munmap(map, buf->size);
+}
+
+static void copy_gtt_to_linear(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ void *map;
+
+ DEBUGFN();
+
+ map = gem_mmap__gtt(bops->fd, buf->handle,
+ buf->size, PROT_READ);
+
+ gem_set_domain(bops->fd, buf->handle,
+ I915_GEM_DOMAIN_GTT, 0);
+
+ igt_memcpy_from_wc(linear, map, buf->size);
+
+ munmap(map, buf->size);
+}
+
+static void copy_linear_to_wc(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ void *map;
+
+ DEBUGFN();
+
+ map = mmap_write(bops->fd, buf);
+ memcpy(map, linear, buf->size);
+ munmap(map, buf->size);
+}
+
+static void copy_wc_to_linear(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ void *map;
+
+ DEBUGFN();
+
+ map = mmap_read(bops->fd, buf);
+ igt_memcpy_from_wc(linear, map, buf->size);
+ munmap(map, buf->size);
+}
+
+void intel_buf_to_linear(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ igt_assert(bops);
+
+ switch (buf->tiling) {
+ case I915_TILING_NONE:
+ igt_assert(bops->to_linear);
+ bops->to_linear(bops, buf, linear);
+ break;
+ case I915_TILING_X:
+ igt_assert(bops->x_to_linear);
+ bops->x_to_linear(bops, buf, linear);
+ break;
+ case I915_TILING_Y:
+ igt_assert(bops->y_to_linear);
+ bops->y_to_linear(bops, buf, linear);
+ break;
+ case I915_TILING_Yf:
+ igt_assert(bops->yf_to_linear);
+ bops->yf_to_linear(bops, buf, linear);
+ break;
+ case I915_TILING_Ys:
+ igt_assert(bops->ys_to_linear);
+ bops->ys_to_linear(bops, buf, linear);
+ break;
+ }
+
+ if (buf->compression)
+ __copy_ccs(bops, buf, linear, CCS_BUF_TO_LINEAR);
+}
+
+void linear_to_intel_buf(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear)
+{
+ igt_assert(bops);
+
+ switch (buf->tiling) {
+ case I915_TILING_NONE:
+ igt_assert(bops->linear_to);
+ bops->linear_to(bops, buf, linear);
+ break;
+ case I915_TILING_X:
+ igt_assert(bops->linear_to_x);
+ bops->linear_to_x(bops, buf, linear);
+ break;
+ case I915_TILING_Y:
+ igt_assert(bops->linear_to_y);
+ bops->linear_to_y(bops, buf, linear);
+ break;
+ case I915_TILING_Yf:
+ igt_assert(bops->linear_to_yf);
+ bops->linear_to_yf(bops, buf, linear);
+ break;
+ case I915_TILING_Ys:
+ igt_assert(bops->linear_to_ys);
+ bops->linear_to_ys(bops, buf, linear);
+ break;
+ }
+
+ if (buf->compression)
+ __copy_ccs(bops, buf, linear, CCS_LINEAR_TO_BUF);
+}
+
+static void __intel_buf_init(struct buf_ops *bops,
+ uint32_t handle,
+ struct intel_buf *buf,
+ int width, int height, int bpp,
+ uint32_t req_tiling, uint32_t compression)
+{
+ uint32_t tiling = req_tiling;
+ uint32_t size;
+
+ igt_assert(bops);
+ igt_assert(buf);
+ igt_assert(width > 0 && height > 0);
+ igt_assert(bpp == 8 || bpp == 16 || bpp == 32);
+
+ memset(buf, 0, sizeof(*buf));
+
+ if (compression) {
+ int aux_width, aux_height;
+
+ igt_require(bops->intel_gen >= 9);
+ igt_assert(req_tiling == I915_TILING_Y ||
+ req_tiling == I915_TILING_Yf);
+
+ /*
+ * On GEN12+ we align the main surface to 4 * 4 main surface
+ * tiles, which is 64kB. These 16 tiles are mapped by 4 AUX
+ * CCS units, that is 4 * 64 bytes. These 4 CCS units are in
+ * turn mapped by one L1 AUX page table entry.
+ */
+ if (bops->intel_gen >= 12)
+ buf->stride = ALIGN(width * (bpp / 8), 128 * 4);
+ else
+ buf->stride = ALIGN(width * (bpp / 8), 128);
+
+ if (bops->intel_gen >= 12)
+ height = ALIGN(height, 4 * 32);
+
+ buf->size = buf->stride * height;
+ buf->tiling = tiling;
+ buf->bpp = bpp;
+ buf->compression = compression;
+
+ aux_width = intel_buf_aux_width(bops->intel_gen, buf);
+ aux_height = intel_buf_aux_height(bops->intel_gen, buf);
+
+ buf->aux.offset = buf->stride * ALIGN(height, 32);
+ buf->aux.stride = aux_width;
+
+ size = buf->aux.offset + aux_width * aux_height;
+
+ } else {
+ buf->stride = ALIGN(width * (bpp / 8), 128);
+ buf->size = buf->stride * height;
+ buf->tiling = tiling;
+ buf->bpp = bpp;
+
+ size = buf->stride * ALIGN(height, 32);
+ }
+
+ if (handle)
+ buf->handle = handle;
+ else
+ buf->handle = gem_create(bops->fd, size);
+
+ set_hw_tiled(bops, buf);
+}
+
+/**
+ * intel_buf_init
+ * @bops: pointer to buf_ops
+ * @buf: pointer to intel_buf structure to be filled
+ * @width: surface width
+ * @height: surface height
+ * @bpp: bits-per-pixel (8 / 16 / 32)
+ * @tiling: surface tiling
+ * @compression: surface compression type
+ *
+ * Function creates new BO within intel_buf structure and fills all
+ * structure fields.
+ *
+ * Note. For X / Y if GPU supports fences HW tiling is configured.
+ */
+void intel_buf_init(struct buf_ops *bops,
+ struct intel_buf *buf,
+ int width, int height, int bpp,
+ uint32_t tiling, uint32_t compression)
+{
+ __intel_buf_init(bops, 0, buf, width, height, bpp, tiling,
+ compression);
+}
+
+/**
+ * intel_buf_close
+ * @bops: pointer to buf_ops
+ * @buf: pointer to intel_buf structure
+ *
+ * Function closes gem BO inside intel_buf.
+ */
+void intel_buf_close(struct buf_ops *bops, struct intel_buf *buf)
+{
+ igt_assert(bops);
+ igt_assert(buf);
+
+ gem_close(bops->fd, buf->handle);
+}
+
+/**
+ * intel_buf_init_using_handle
+ * @bops: pointer to buf_ops
+ * @handle: BO handle created by the caller
+ * @buf: pointer to intel_buf structure to be filled
+ * @width: surface width
+ * @height: surface height
+ * @bpp: bits-per-pixel (8 / 16 / 32)
+ * @tiling: surface tiling
+ * @compression: surface compression type
+ *
+ * Function configures BO handle within intel_buf structure passed by the caller
+ * (with all its metadata - width, height, ...). Useful if BO was created
+ * outside.
+ *
+ * Note: intel_buf_close() can be used to close the BO handle, but caller
+ * must be aware to not close the BO twice.
+ */
+void intel_buf_init_using_handle(struct buf_ops *bops,
+ uint32_t handle,
+ struct intel_buf *buf,
+ int width, int height, int bpp,
+ uint32_t req_tiling, uint32_t compression)
+{
+ __intel_buf_init(bops, handle, buf, width, height, bpp, req_tiling,
+ compression);
+}
+
+#define DEFAULT_BUFOPS(__gen_start, __gen_end) \
+ .gen_start = __gen_start, \
+ .gen_end = __gen_end, \
+ .supported_hw_tiles = TILE_X | TILE_Y, \
+ .linear_to = copy_linear_to_wc, \
+ .linear_to_x = copy_linear_to_gtt, \
+ .linear_to_y = copy_linear_to_gtt, \
+ .linear_to_yf = copy_linear_to_yf, \
+ .linear_to_ys = copy_linear_to_ys, \
+ .to_linear = copy_wc_to_linear, \
+ .x_to_linear = copy_gtt_to_linear, \
+ .y_to_linear = copy_gtt_to_linear, \
+ .yf_to_linear = copy_yf_to_linear, \
+ .ys_to_linear = copy_ys_to_linear
+
+struct buf_ops buf_ops_arr[] = {
+ {
+ DEFAULT_BUFOPS(2, 8),
+ .supported_tiles = TILE_NONE | TILE_X | TILE_Y,
+ },
+
+ {
+ DEFAULT_BUFOPS(9, 11),
+ .supported_tiles = TILE_NONE | TILE_X | TILE_Y | TILE_Yf,
+ },
+
+ {
+ DEFAULT_BUFOPS(12, 12),
+ .supported_tiles = TILE_NONE | TILE_X | TILE_Y | TILE_Yf | TILE_Ys,
+ },
+};
+
+static bool probe_hw_tiling(struct buf_ops *bops, uint32_t tiling)
+{
+ uint64_t size = 256 * 256;
+ uint32_t handle, buf_tiling, buf_swizzle;
+ uint32_t stride;
+ int ret;
+ bool is_set = false;
+
+ if (tiling == I915_TILING_X)
+ stride = 512;
+ else if (tiling == I915_TILING_Y)
+ stride = 128;
+ else
+ return false;
+
+ handle = gem_create(bops->fd, size);
+
+ /* Single shot, if no fences are available we fail immediately */
+ ret = __set_tiling(bops->fd, handle, tiling, stride);
+ if (ret)
+ goto end;
+
+ is_set = __get_tiling(bops->fd, handle, &buf_tiling, &buf_swizzle);
+ if (is_set) {
+ if (tiling == I915_TILING_X)
+ bops->swizzle_x = buf_swizzle;
+ else if (tiling == I915_TILING_Y)
+ bops->swizzle_y = buf_swizzle;
+ }
+end:
+ gem_close(bops->fd, handle);
+
+ return is_set;
+}
+
+static void *alloc_aligned(uint64_t size)
+{
+ void *p;
+
+ igt_assert_eq(posix_memalign(&p, 16, size), 0);
+
+ return p;
+}
+
+/*
+ * Simple idempotency test between HW -> SW and SW -> HW BO.
+ */
+static void idempotency_selftest(struct buf_ops *bops, uint32_t tiling)
+{
+ struct intel_buf buf;
+ uint8_t *linear_in, *linear_out, *map;
+ int i;
+ const int width = 512;
+ const int height = 512;
+ const int bpp = 32;
+ const int size = width * height * bpp / 8;
+ bool software_tiling = false;
+
+ if (!is_hw_tiling_supported(bops, tiling))
+ return;
+
+ linear_in = alloc_aligned(size);
+ linear_out = alloc_aligned(size);
+
+ for (i = 0; i < size; i++)
+ linear_in[i] = i % 253; /* prime chosen intentionally */
+
+ do {
+ igt_debug("Checking idempotency, SW: %s, HW: %s, tiling: %s\n",
+ bool_str(software_tiling),
+ bool_str(!software_tiling),
+ tiling_str(tiling));
+ intel_buf_init(bops, &buf, width, height, bpp, tiling, false);
+ buf_ops_set_software_tiling(bops, tiling, software_tiling);
+
+ linear_to_intel_buf(bops, &buf, (uint32_t *) linear_in);
+ map = __gem_mmap_offset__cpu(bops->fd, buf.handle, 0,
+ buf.size, PROT_READ);
+ if (!map)
+ map = gem_mmap__cpu(bops->fd, buf.handle, 0,
+ buf.size, PROT_READ);
+ gem_set_domain(bops->fd, buf.handle, I915_GEM_DOMAIN_CPU, 0);
+ igt_assert(memcmp(linear_in, map, size));
+ munmap(map, size);
+
+ buf_ops_set_software_tiling(bops, tiling, !software_tiling);
+ intel_buf_to_linear(bops, &buf, (uint32_t *) linear_out);
+ igt_assert(!memcmp(linear_in, linear_out, size));
+
+ intel_buf_close(bops, &buf);
+
+ software_tiling = !software_tiling;
+ } while (software_tiling);
+
+ igt_debug("Idempotency for %s tiling OK\n", tiling_str(tiling));
+ buf_ops_set_software_tiling(bops, tiling, false);
+}
+
+/**
+ * buf_ops_create
+ * @fd: device filedescriptor
+ *
+ * Create buf_ops structure depending on fd-device capabilities.
+ *
+ * Returns: opaque pointer to buf_ops.
+ *
+ */
+struct buf_ops *buf_ops_create(int fd)
+{
+ struct buf_ops *bops = calloc(1, sizeof(*bops));
+ uint32_t devid;
+ int generation;
+
+ igt_assert(bops);
+
+ devid = intel_get_drm_devid(fd);
+ generation = intel_gen(devid);
+
+ /* Predefined settings */
+ for (int i = 0; i < ARRAY_SIZE(buf_ops_arr); i++) {
+ if (generation >= buf_ops_arr[i].gen_start &&
+ generation <= buf_ops_arr[i].gen_end) {
+ memcpy(bops, &buf_ops_arr[i], sizeof(*bops));
+ bops->fd = fd;
+ bops->intel_gen = generation;
+ igt_debug("generation: %d, supported tiles: 0x%02x\n",
+ generation, bops->supported_tiles);
+ break;
+ }
+ }
+
+ igt_assert(bops->intel_gen);
+
+ /*
+ * Warning!
+ *
+ * Gen2 software tiling/detiling is not supported! (yet).
+ *
+ * If you are brave hero with an access to Gen2 you can save the world.
+ * Until then we're doomed to use only hardware (de)tiling.
+ *
+ * Ok, you have been warned.
+ */
+ if (bops->intel_gen == 2) {
+ igt_warn("Gen2 detected. HW (de)tiling support only.");
+ return bops;
+ }
+
+ /* Let's probe X and Y hw tiling support */
+ if (is_hw_tiling_supported(bops, I915_TILING_X)) {
+ bool supported = probe_hw_tiling(bops, I915_TILING_X);
+
+ igt_debug("X fence support: %s\n", bool_str(supported));
+ if (!supported) {
+ bops->supported_hw_tiles &= ~TILE_X;
+ bops->linear_to_x = copy_linear_to_x;
+ bops->x_to_linear = copy_x_to_linear;
+ }
+ }
+
+ if (is_hw_tiling_supported(bops, I915_TILING_Y)) {
+ bool supported = probe_hw_tiling(bops, I915_TILING_Y);
+
+ igt_debug("Y fence support: %s\n", bool_str(supported));
+ if (!supported) {
+ bops->supported_hw_tiles &= ~TILE_Y;
+ bops->linear_to_y = copy_linear_to_y;
+ bops->y_to_linear = copy_y_to_linear;
+ }
+ }
+
+ /* Disable other tiling format functions if not supported */
+ if (!is_tiling_supported(bops, I915_TILING_Yf)) {
+ igt_debug("Yf format not supported\n");
+ bops->linear_to_yf = NULL;
+ bops->yf_to_linear = NULL;
+ }
+
+ if (!is_tiling_supported(bops, I915_TILING_Ys)) {
+ igt_debug("Ys format not supported\n");
+ bops->linear_to_ys = NULL;
+ bops->ys_to_linear = NULL;
+ }
+
+ idempotency_selftest(bops, I915_TILING_X);
+ idempotency_selftest(bops, I915_TILING_Y);
+
+ return bops;
+}
+
+/**
+ * buf_ops_destroy
+ * @bops: pointer to buf_ops
+ *
+ * Function frees buf_ops structure.
+ */
+void buf_ops_destroy(struct buf_ops *bops)
+{
+ igt_assert(bops);
+ free(bops);
+}
+
+/**
+ * buf_ops_set_software_tiling
+ * @bops: pointer to buf_ops
+ * @tiling: surface tiling
+ * @use_software_tiling: if true use software copying methods, otherwise
+ * use hardware (via gtt)
+ *
+ * Function allows switch X / Y surfaces to software / hardware copying methods
+ * which honors tiling and swizzling.
+ *
+ * Returns:
+ * false - switch wasn't possible.
+ * true - switch to software / hardware method succeed.
+ */
+bool buf_ops_set_software_tiling(struct buf_ops *bops,
+ uint32_t tiling,
+ bool use_software_tiling)
+{
+ bool was_changed = true;
+
+ igt_assert(bops);
+
+ /* Until appropriate code is added we don't support SW tiling on Gen2 */
+ if (bops->intel_gen == 2) {
+ igt_warn("Change to software tiling on Gen2 is not supported!");
+ return false;
+ }
+
+ switch (tiling) {
+ case I915_TILING_X:
+ if (use_software_tiling) {
+ igt_debug("-> change X to SW\n");
+ bops->linear_to_x = copy_linear_to_x;
+ bops->x_to_linear = copy_x_to_linear;
+ } else {
+ if (is_hw_tiling_supported(bops, I915_TILING_X)) {
+ igt_debug("-> change X to HW\n");
+ bops->linear_to_x = copy_linear_to_gtt;
+ bops->x_to_linear = copy_gtt_to_linear;
+ } else {
+ igt_debug("-> X cannot be changed to HW\n");
+ was_changed = false;
+ }
+ }
+ break;
+
+ case I915_TILING_Y:
+ if (use_software_tiling) {
+ igt_debug("-> change Y to SW\n");
+ bops->linear_to_y = copy_linear_to_y;
+ bops->y_to_linear = copy_y_to_linear;
+ } else {
+ if (is_hw_tiling_supported(bops, I915_TILING_Y)) {
+ igt_debug("-> change Y to HW\n");
+ bops->linear_to_y = copy_linear_to_gtt;
+ bops->y_to_linear = copy_gtt_to_linear;
+ } else {
+ igt_debug("-> Y cannot be changed to HW\n");
+ was_changed = false;
+ }
+ }
+ break;
+
+ default:
+ igt_warn("Invalid tiling: %d\n", tiling);
+ was_changed = false;
+ }
+
+ return was_changed;
+}
+
+/**
+ * buf_ops_has_hw_fence
+ * @bops: pointer to buf_ops
+ * @tiling: surface tiling
+ *
+ * Function checks if surface with tiling has HW fences which can be used
+ * to copy it via gtt.
+ *
+ * Returns:
+ * false - fence for tiling is not supported.
+ * true - fence for tiling is supported.
+ */
+bool buf_ops_has_hw_fence(struct buf_ops *bops, uint32_t tiling)
+{
+ uint32_t tile_mask = TILE_DEF(tiling);
+
+ igt_assert(bops);
+
+ if (tile_mask & bops->supported_hw_tiles)
+ return true;
+
+ return false;
+}
+
+/**
+ * buf_ops_has_tiling_support
+ * @bops: pointer to buf_ops
+ * @tiling: surface tiling
+ *
+ * Function checks capabilities to handle surfaces with tiling in GPU.
+ *
+ * Returns:
+ * false - GPU does not support tiling.
+ * true - GPU supports tiling.
+ */
+bool buf_ops_has_tiling_support(struct buf_ops *bops, uint32_t tiling)
+{
+ uint32_t tile_mask = TILE_DEF(tiling);
+
+ igt_assert(bops);
+
+ if (tile_mask & bops->supported_tiles)
+ return true;
+
+ return false;
+}
diff --git a/lib/intel_bufops.h b/lib/intel_bufops.h
new file mode 100644
index 00000000..f3d6aed8
--- /dev/null
+++ b/lib/intel_bufops.h
@@ -0,0 +1,86 @@
+#ifndef __INTEL_BUFOPS_H__
+#define __INTEL_BUFOPS_H__
+
+#include <stdint.h>
+
+struct buf_ops;
+
+struct intel_buf {
+ uint32_t handle;
+ uint32_t stride;
+ uint32_t tiling;
+ uint32_t bpp;
+ uint32_t size;
+ uint32_t compression;
+ struct {
+ uint32_t offset;
+ uint32_t stride;
+ } aux;
+};
+
+static inline unsigned int intel_buf_width(const struct intel_buf *buf)
+{
+ return buf->stride / (buf->bpp / 8);
+}
+
+static inline unsigned int intel_buf_height(const struct intel_buf *buf)
+{
+ return buf->size / buf->stride;
+}
+
+static inline unsigned int
+intel_buf_aux_width(int gen, const struct intel_buf *buf)
+{
+ /*
+ * GEN12+: The AUX CCS unit size is 64 bytes mapping 4 main surface
+ * tiles. Thus the width of the CCS unit is 4*32=128 pixels on the
+ * main surface.
+ */
+ if (gen >= 12)
+ return DIV_ROUND_UP(intel_buf_width(buf), 128) * 64;
+
+ return DIV_ROUND_UP(intel_buf_width(buf), 1024) * 128;
+}
+
+static inline unsigned int
+intel_buf_aux_height(int gen, const struct intel_buf *buf)
+{
+ /*
+ * GEN12+: The AUX CCS unit size is 64 bytes mapping 4 main surface
+ * tiles. Thus the height of the CCS unit is 32 pixel rows on the main
+ * surface.
+ */
+ if (gen >= 12)
+ return DIV_ROUND_UP(intel_buf_height(buf), 32);
+
+ return DIV_ROUND_UP(intel_buf_height(buf), 512) * 32;
+}
+
+struct buf_ops *buf_ops_create(int fd);
+void buf_ops_destroy(struct buf_ops *bops);
+
+bool buf_ops_set_software_tiling(struct buf_ops *bops,
+ uint32_t tiling,
+ bool use_software_tiling);
+
+void intel_buf_to_linear(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear);
+
+void linear_to_intel_buf(struct buf_ops *bops, struct intel_buf *buf,
+ uint32_t *linear);
+
+bool buf_ops_has_hw_fence(struct buf_ops *bops, uint32_t tiling);
+bool buf_ops_has_tiling_support(struct buf_ops *bops, uint32_t tiling);
+
+void intel_buf_init(struct buf_ops *bops, struct intel_buf *buf,
+ int width, int height, int bpp,
+ uint32_t tiling, uint32_t compression);
+void intel_buf_close(struct buf_ops *bops, struct intel_buf *buf);
+
+void intel_buf_init_using_handle(struct buf_ops *bops,
+ uint32_t handle,
+ struct intel_buf *buf,
+ int width, int height, int bpp,
+ uint32_t req_tiling, uint32_t compression);
+
+#endif
diff --git a/lib/meson.build b/lib/meson.build
index 57eb7d93..cf3e35da 100644
--- a/lib/meson.build
+++ b/lib/meson.build
@@ -27,6 +27,7 @@ lib_sources = [
'igt_x86.c',
'instdone.c',
'intel_batchbuffer.c',
+ 'intel_bufops.c',
'intel_chipset.c',
'intel_device_info.c',
'intel_os.c',