summaryrefslogtreecommitdiff
path: root/tests
diff options
context:
space:
mode:
authorChris Wilson <chris@chris-wilson.co.uk>2020-08-10 13:32:31 +0100
committerChris Wilson <chris@chris-wilson.co.uk>2020-08-21 10:37:48 +0100
commitf1d0c240ea2e631dfb9f493f37f8fb61cb2b1cf2 (patch)
treec701cba54386588be5665923276430821024149a /tests
parent3d8c9e921e341b074f8ecb29872278cb4231a397 (diff)
i915/perf_pmu: Emit a semaphore to measure
Don't assume the kernel will emit a semaphore to synchronise between two engines, and emit the semaphore ourselves for the basis of our measurements. The purpose of the test is to try and ascertain the accuracy of the two sampling methods, semaphore busyness uses register polling, whereas the engine busyness may use ktime_t of the CS events. Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> Cc: Ramalingam C <ramalingam.c@intel.com> Reviewed-by: Ramalingam C <ramalingam.c@intel.com>
Diffstat (limited to 'tests')
-rw-r--r--tests/i915/perf_pmu.c95
1 files changed, 64 insertions, 31 deletions
diff --git a/tests/i915/perf_pmu.c b/tests/i915/perf_pmu.c
index 13e1bd93..a66b9a47 100644
--- a/tests/i915/perf_pmu.c
+++ b/tests/i915/perf_pmu.c
@@ -650,6 +650,7 @@ no_sema(int gem_fd, const struct intel_execution_engine2 *e, unsigned int flags)
#define MI_SEMAPHORE_WAIT MI_INSTR(0x1c, 2) /* GEN8+ */
#define MI_SEMAPHORE_POLL (1<<15)
#define MI_SEMAPHORE_SAD_GTE_SDD (1<<12)
+#define MI_SEMAPHORE_SAD_NEQ_SDD (5 << 12)
static void
sema_wait(int gem_fd, const struct intel_execution_engine2 *e,
@@ -751,10 +752,39 @@ sema_wait(int gem_fd, const struct intel_execution_engine2 *e,
assert_within_epsilon(val[1] - val[0], slept, tolerance);
}
+static uint32_t
+create_sema(int gem_fd, struct drm_i915_gem_relocation_entry *reloc)
+{
+ uint32_t cs[] = {
+ /* Reset our semaphore wait */
+ MI_STORE_DWORD_IMM,
+ 0,
+ 0,
+ 1,
+
+ /* Wait until the semaphore value is set to 0 [by caller] */
+ MI_SEMAPHORE_WAIT | MI_SEMAPHORE_POLL | MI_SEMAPHORE_SAD_NEQ_SDD,
+ 1,
+ 0,
+ 0,
+
+ MI_BATCH_BUFFER_END
+ };
+ uint32_t handle = gem_create(gem_fd, 4096);
+
+ memset(reloc, 0, 2 * sizeof(*reloc));
+ reloc[0].target_handle = handle;
+ reloc[0].offset = 64 + 1 * sizeof(uint32_t);
+ reloc[1].target_handle = handle;
+ reloc[1].offset = 64 + 6 * sizeof(uint32_t);
+
+ gem_write(gem_fd, handle, 64, cs, sizeof(cs));
+ return handle;
+}
+
static void
__sema_busy(int gem_fd, int pmu,
const struct intel_execution_engine2 *e,
- const struct intel_execution_engine2 *signal,
int sema_pct,
int busy_pct)
{
@@ -764,39 +794,54 @@ __sema_busy(int gem_fd, int pmu,
};
uint64_t total, sema, busy;
uint64_t start[2], val[2];
- igt_spin_t *spin[2];
+ struct drm_i915_gem_relocation_entry reloc[2];
+ struct drm_i915_gem_exec_object2 obj = {
+ .handle = create_sema(gem_fd, reloc),
+ .relocation_count = 2,
+ .relocs_ptr = to_user_pointer(reloc),
+ };
+ struct drm_i915_gem_execbuffer2 eb = {
+ .batch_start_offset = 64,
+ .buffer_count = 1,
+ .buffers_ptr = to_user_pointer(&obj),
+ .flags = e->flags,
+ };
+ igt_spin_t *spin;
+ uint32_t *map;
/* Time spent being busy includes time waiting on semaphores */
igt_assert(busy_pct >= sema_pct);
gem_quiescent_gpu(gem_fd);
- spin[0] = igt_spin_new(gem_fd,
- .engine = signal->flags,
- .flags = IGT_SPIN_FENCE_OUT | IGT_SPIN_POLL_RUN);
- spin[1] = igt_spin_new(gem_fd,
- .engine = e->flags,
- .fence = spin[0]->out_fence,
- .flags = IGT_SPIN_FENCE_IN);
+ map = gem_mmap__wc(gem_fd, obj.handle, 0, 4096, PROT_WRITE);
+ gem_execbuf(gem_fd, &eb);
+ spin = igt_spin_new(gem_fd, .engine = e->flags);
- igt_spin_busywait_until_started(spin[0]);
+ /* Wait until the batch is executed and the semaphore is busy-waiting */
+ while (!READ_ONCE(*map) && gem_bo_busy(gem_fd, obj.handle))
+ ;
+ igt_assert(gem_bo_busy(gem_fd, obj.handle));
+ gem_close(gem_fd, obj.handle);
total = pmu_read_multi(pmu, 2, start);
sema = measured_usleep(batch_duration_ns * sema_pct / 100 / 1000);
- igt_spin_end(spin[0]);
+ *map = 0; __sync_synchronize();
busy = measured_usleep(batch_duration_ns * (busy_pct - sema_pct) / 100 / 1000);
- igt_spin_end(spin[1]);
+ igt_spin_end(spin);
measured_usleep(batch_duration_ns * (100 - busy_pct) / 100 / 1000);
total = pmu_read_multi(pmu, 2, val) - total;
+ igt_spin_free(gem_fd, spin);
+ munmap(map, 4096);
busy += sema;
val[SEMA] -= start[SEMA];
val[BUSY] -= start[BUSY];
- igt_info("%s<-%s, target: {%.1f%% [%d], %.1f%% [%d]}, measured: {%.1f%%, %.1f%%}\n",
- e->name, signal->name,
+ igt_info("%s, target: {%.1f%% [%d], %.1f%% [%d]}, measured: {%.1f%%, %.1f%%}\n",
+ e->name,
sema * 100. / total, sema_pct,
busy * 100. / total, busy_pct,
val[SEMA] * 100. / total,
@@ -808,9 +853,6 @@ __sema_busy(int gem_fd, int pmu,
"Semaphore time (%.3fus, %.1f%%) greater than total time busy (%.3fus, %.1f%%)!\n",
val[SEMA] * 1e-3, val[SEMA] * 100. / total,
val[BUSY] * 1e-3, val[BUSY] * 100. / total);
-
- igt_spin_free(gem_fd, spin[1]);
- igt_spin_free(gem_fd, spin[0]);
}
static void
@@ -818,25 +860,16 @@ sema_busy(int gem_fd,
const struct intel_execution_engine2 *e,
unsigned int flags)
{
- const struct intel_execution_engine2 *signal;
int fd;
- igt_require(gem_scheduler_has_semaphores(gem_fd));
- igt_require(gem_scheduler_has_preemption(gem_fd));
+ igt_require(intel_gen(intel_get_drm_devid(gem_fd)) >= 8);
- fd = open_group(gem_fd,
- I915_PMU_ENGINE_SEMA(e->class, e->instance), -1);
+ fd = open_group(gem_fd, I915_PMU_ENGINE_SEMA(e->class, e->instance), -1);
open_group(gem_fd, I915_PMU_ENGINE_BUSY(e->class, e->instance), fd);
- __for_each_physical_engine(gem_fd, signal) {
- if (e->class == signal->class &&
- e->instance == signal->instance)
- continue;
-
- __sema_busy(gem_fd, fd, e, signal, 50, 100);
- __sema_busy(gem_fd, fd, e, signal, 25, 50);
- __sema_busy(gem_fd, fd, e, signal, 75, 75);
- }
+ __sema_busy(gem_fd, fd, e, 50, 100);
+ __sema_busy(gem_fd, fd, e, 25, 50);
+ __sema_busy(gem_fd, fd, e, 75, 75);
close(fd);
}