From: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com> To: igt-dev@lists.freedesktop.org Cc: Intel-gfx@lists.freedesktop.org Subject: [Intel-gfx] [PATCH i-g-t] i915/perf_pmu: Fix perf fd leak Date: Tue, 13 Oct 2020 10:46:12 +0100 [thread overview] Message-ID: <20201013094612.83843-1-tvrtko.ursulin@linux.intel.com> (raw) From: Tvrtko Ursulin <tvrtko.ursulin@intel.com> As it turns out opening the perf fd in group mode still produces separate file descriptors for all members of the group, which in turn need to be closed manually to avoid leaking them. Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com> --- tests/i915/perf_pmu.c | 130 +++++++++++++++++++++++++----------------- 1 file changed, 78 insertions(+), 52 deletions(-) diff --git a/tests/i915/perf_pmu.c b/tests/i915/perf_pmu.c index 873b275dca6b..6f8bec28d274 100644 --- a/tests/i915/perf_pmu.c +++ b/tests/i915/perf_pmu.c @@ -475,7 +475,8 @@ busy_check_all(int gem_fd, const struct intel_execution_engine2 *e, end_spin(gem_fd, spin, FLAG_SYNC); igt_spin_free(gem_fd, spin); - close(fd[0]); + for (i = 0; i < num_engines; i++) + close(fd[i]); for (i = 0; i < num_engines; i++) val[i] = tval[1][i] - tval[0][i]; @@ -546,7 +547,8 @@ most_busy_check_all(int gem_fd, const struct intel_execution_engine2 *e, end_spin(gem_fd, spin, FLAG_SYNC); igt_spin_free(gem_fd, spin); - close(fd[0]); + for (i = 0; i < num_engines; i++) + close(fd[i]); for (i = 0; i < num_engines; i++) val[i] = tval[1][i] - tval[0][i]; @@ -600,7 +602,8 @@ all_busy_check_all(int gem_fd, const unsigned int num_engines, end_spin(gem_fd, spin, FLAG_SYNC); igt_spin_free(gem_fd, spin); - close(fd[0]); + for (i = 0; i < num_engines; i++) + close(fd[i]); for (i = 0; i < num_engines; i++) val[i] = tval[1][i] - tval[0][i]; @@ -617,22 +620,23 @@ no_sema(int gem_fd, const struct intel_execution_engine2 *e, unsigned int flags) { igt_spin_t *spin; uint64_t val[2][2]; - int fd; + int fd[2]; - fd = open_group(gem_fd, - I915_PMU_ENGINE_SEMA(e->class, e->instance), -1); - open_group(gem_fd, I915_PMU_ENGINE_WAIT(e->class, e->instance), fd); + fd[0] = open_group(gem_fd, I915_PMU_ENGINE_SEMA(e->class, e->instance), + -1); + fd[1] = open_group(gem_fd, I915_PMU_ENGINE_WAIT(e->class, e->instance), + fd[0]); if (flags & TEST_BUSY) spin = spin_sync(gem_fd, 0, e); else spin = NULL; - pmu_read_multi(fd, 2, val[0]); + pmu_read_multi(fd[0], 2, val[0]); measured_usleep(batch_duration_ns / 1000); if (flags & TEST_TRAILING_IDLE) end_spin(gem_fd, spin, flags); - pmu_read_multi(fd, 2, val[1]); + pmu_read_multi(fd[0], 2, val[1]); val[0][0] = val[1][0] - val[0][0]; val[0][1] = val[1][1] - val[0][1]; @@ -641,7 +645,8 @@ no_sema(int gem_fd, const struct intel_execution_engine2 *e, unsigned int flags) end_spin(gem_fd, spin, FLAG_SYNC); igt_spin_free(gem_fd, spin); } - close(fd); + close(fd[0]); + close(fd[1]); assert_within_epsilon(val[0][0], 0.0f, tolerance); assert_within_epsilon(val[0][1], 0.0f, tolerance); @@ -861,18 +866,21 @@ sema_busy(int gem_fd, const struct intel_execution_engine2 *e, unsigned int flags) { - int fd; + int fd[2]; igt_require(intel_gen(intel_get_drm_devid(gem_fd)) >= 8); - fd = open_group(gem_fd, I915_PMU_ENGINE_SEMA(e->class, e->instance), -1); - open_group(gem_fd, I915_PMU_ENGINE_BUSY(e->class, e->instance), fd); + fd[0] = open_group(gem_fd, I915_PMU_ENGINE_SEMA(e->class, e->instance), + -1); + fd[1] = open_group(gem_fd, I915_PMU_ENGINE_BUSY(e->class, e->instance), + fd[0]); - __sema_busy(gem_fd, fd, e, 50, 100); - __sema_busy(gem_fd, fd, e, 25, 50); - __sema_busy(gem_fd, fd, e, 75, 75); + __sema_busy(gem_fd, fd[0], e, 50, 100); + __sema_busy(gem_fd, fd[0], e, 25, 50); + __sema_busy(gem_fd, fd[0], e, 75, 75); - close(fd); + close(fd[0]); + close(fd[1]); } #define MI_WAIT_FOR_PIPE_C_VBLANK (1<<21) @@ -1441,7 +1449,7 @@ test_frequency(int gem_fd) uint64_t val[2], start[2], slept; double min[2], max[2]; igt_spin_t *spin; - int fd, sysfs; + int fd[2], sysfs; sysfs = igt_sysfs_open(gem_fd); igt_require(sysfs >= 0); @@ -1455,8 +1463,8 @@ test_frequency(int gem_fd) igt_require(max_freq > min_freq); igt_require(boost_freq > min_freq); - fd = open_group(gem_fd, I915_PMU_REQUESTED_FREQUENCY, -1); - open_group(gem_fd, I915_PMU_ACTUAL_FREQUENCY, fd); + fd[0] = open_group(gem_fd, I915_PMU_REQUESTED_FREQUENCY, -1); + fd[1] = open_group(gem_fd, I915_PMU_ACTUAL_FREQUENCY, fd[0]); /* * Set GPU to min frequency and read PMU counters. @@ -1471,9 +1479,9 @@ test_frequency(int gem_fd) gem_quiescent_gpu(gem_fd); /* Idle to be sure the change takes effect */ spin = spin_sync_flags(gem_fd, 0, I915_EXEC_DEFAULT); - slept = pmu_read_multi(fd, 2, start); + slept = pmu_read_multi(fd[0], 2, start); measured_usleep(batch_duration_ns / 1000); - slept = pmu_read_multi(fd, 2, val) - slept; + slept = pmu_read_multi(fd[0], 2, val) - slept; min[0] = 1e9*(val[0] - start[0]) / slept; min[1] = 1e9*(val[1] - start[1]) / slept; @@ -1497,9 +1505,9 @@ test_frequency(int gem_fd) gem_quiescent_gpu(gem_fd); spin = spin_sync_flags(gem_fd, 0, I915_EXEC_DEFAULT); - slept = pmu_read_multi(fd, 2, start); + slept = pmu_read_multi(fd[0], 2, start); measured_usleep(batch_duration_ns / 1000); - slept = pmu_read_multi(fd, 2, val) - slept; + slept = pmu_read_multi(fd[0], 2, val) - slept; max[0] = 1e9*(val[0] - start[0]) / slept; max[1] = 1e9*(val[1] - start[1]) / slept; @@ -1514,7 +1522,8 @@ test_frequency(int gem_fd) if (igt_sysfs_get_u32(sysfs, "gt_min_freq_mhz") != min_freq) igt_warn("Unable to restore min frequency to saved value [%u MHz], now %u MHz\n", min_freq, igt_sysfs_get_u32(sysfs, "gt_min_freq_mhz")); - close(fd); + close(fd[0]); + close(fd[1]); igt_info("Min frequency: requested %.1f, actual %.1f\n", min[0], min[1]); @@ -1535,7 +1544,7 @@ test_frequency_idle(int gem_fd) uint32_t min_freq; uint64_t val[2], start[2], slept; double idle[2]; - int fd, sysfs; + int fd[2], sysfs; sysfs = igt_sysfs_open(gem_fd); igt_require(sysfs >= 0); @@ -1545,17 +1554,18 @@ test_frequency_idle(int gem_fd) /* While parked, our convention is to report the GPU at 0Hz */ - fd = open_group(gem_fd, I915_PMU_REQUESTED_FREQUENCY, -1); - open_group(gem_fd, I915_PMU_ACTUAL_FREQUENCY, fd); + fd[0] = open_group(gem_fd, I915_PMU_REQUESTED_FREQUENCY, -1); + fd[1] = open_group(gem_fd, I915_PMU_ACTUAL_FREQUENCY, fd[0]); gem_quiescent_gpu(gem_fd); /* Be idle! */ measured_usleep(2000); /* Wait for timers to cease */ - slept = pmu_read_multi(fd, 2, start); + slept = pmu_read_multi(fd[0], 2, start); measured_usleep(batch_duration_ns / 1000); - slept = pmu_read_multi(fd, 2, val) - slept; + slept = pmu_read_multi(fd[0], 2, val) - slept; - close(fd); + close(fd[0]); + close(fd[1]); idle[0] = 1e9*(val[0] - start[0]) / slept; idle[1] = 1e9*(val[1] - start[1]) / slept; @@ -1926,57 +1936,73 @@ static int unload_i915(void) return 0; } -static void test_unload(void) +static void test_unload(unsigned int num_engines) { igt_fork(child, 1) { const struct intel_execution_engine2 *e; + int fd[4 + num_engines], i; uint64_t *buf; - int count = 1; + int count = 0; int i915; - int fd; i915 = __drm_open_driver(DRIVER_INTEL); igt_debug("Opening perf events\n"); - fd = open_group(i915, I915_PMU_INTERRUPTS, -1); + fd[count] = open_group(i915, I915_PMU_INTERRUPTS, -1); + if (fd[count] != -1) + count++; - if (perf_i915_open_group(i915, I915_PMU_REQUESTED_FREQUENCY,fd) != -1) + fd[count] = perf_i915_open_group(i915, + I915_PMU_REQUESTED_FREQUENCY, + fd[count - 1]); + if (fd[count] != -1) count++; - if (perf_i915_open_group(i915, I915_PMU_ACTUAL_FREQUENCY, fd) != -1) + + fd[count] = perf_i915_open_group(i915, + I915_PMU_ACTUAL_FREQUENCY, + fd[count - 1]); + if (fd[count] != -1) count++; __for_each_physical_engine(i915, e) { - if (perf_i915_open_group(i915, - I915_PMU_ENGINE_BUSY(e->class, e->instance), - fd) != -1) + fd[count] = perf_i915_open_group(i915, + I915_PMU_ENGINE_BUSY(e->class, e->instance), + fd[count - 1]); + if (fd[count] != -1) count++; - if (perf_i915_open_group(i915, - I915_PMU_ENGINE_SEMA(e->class, e->instance), - fd) != -1) + fd[count] = perf_i915_open_group(i915, + I915_PMU_ENGINE_SEMA(e->class, e->instance), + fd[count - 1]); + if (fd[count] != -1) count++; - if (perf_i915_open_group(i915, - I915_PMU_ENGINE_WAIT(e->class, e->instance), - fd) != -1) + fd[count] = perf_i915_open_group(i915, + I915_PMU_ENGINE_WAIT(e->class, e->instance), + fd[count - 1]); + if (fd[count] != -1) count++; } - if (perf_i915_open_group(i915, I915_PMU_RC6_RESIDENCY,fd) != -1) + fd[count] = perf_i915_open_group(i915, I915_PMU_RC6_RESIDENCY, + fd[count - 1]); + if (fd[count] != -1) count++; close(i915); - buf = calloc(count + 1, sizeof(uint64_t)); + buf = calloc(count, sizeof(uint64_t)); igt_assert(buf); igt_debug("Read %d events from perf and trial unload\n", count); - pmu_read_multi(fd, count, buf); + pmu_read_multi(fd[0], count, buf); igt_assert_eq(unload_i915(), -EBUSY); - pmu_read_multi(fd, count, buf); + pmu_read_multi(fd[0], count, buf); igt_debug("Close perf\n"); - close(fd); + + for (i = 0; i < count; i++) + close(fd[i]); free(buf); } @@ -2357,6 +2383,6 @@ igt_main igt_subtest("module-unload") { igt_require(unload_i915() == 0); for (int pass = 0; pass < 3; pass++) - test_unload(); + test_unload(num_engines); } } -- 2.25.1 _______________________________________________ Intel-gfx mailing list Intel-gfx@lists.freedesktop.org https://lists.freedesktop.org/mailman/listinfo/intel-gfx
WARNING: multiple messages have this Message-ID (diff)
From: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com> To: igt-dev@lists.freedesktop.org Cc: Intel-gfx@lists.freedesktop.org, Tvrtko Ursulin <tvrtko.ursulin@intel.com> Subject: [igt-dev] [PATCH i-g-t] i915/perf_pmu: Fix perf fd leak Date: Tue, 13 Oct 2020 10:46:12 +0100 [thread overview] Message-ID: <20201013094612.83843-1-tvrtko.ursulin@linux.intel.com> (raw) From: Tvrtko Ursulin <tvrtko.ursulin@intel.com> As it turns out opening the perf fd in group mode still produces separate file descriptors for all members of the group, which in turn need to be closed manually to avoid leaking them. Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com> --- tests/i915/perf_pmu.c | 130 +++++++++++++++++++++++++----------------- 1 file changed, 78 insertions(+), 52 deletions(-) diff --git a/tests/i915/perf_pmu.c b/tests/i915/perf_pmu.c index 873b275dca6b..6f8bec28d274 100644 --- a/tests/i915/perf_pmu.c +++ b/tests/i915/perf_pmu.c @@ -475,7 +475,8 @@ busy_check_all(int gem_fd, const struct intel_execution_engine2 *e, end_spin(gem_fd, spin, FLAG_SYNC); igt_spin_free(gem_fd, spin); - close(fd[0]); + for (i = 0; i < num_engines; i++) + close(fd[i]); for (i = 0; i < num_engines; i++) val[i] = tval[1][i] - tval[0][i]; @@ -546,7 +547,8 @@ most_busy_check_all(int gem_fd, const struct intel_execution_engine2 *e, end_spin(gem_fd, spin, FLAG_SYNC); igt_spin_free(gem_fd, spin); - close(fd[0]); + for (i = 0; i < num_engines; i++) + close(fd[i]); for (i = 0; i < num_engines; i++) val[i] = tval[1][i] - tval[0][i]; @@ -600,7 +602,8 @@ all_busy_check_all(int gem_fd, const unsigned int num_engines, end_spin(gem_fd, spin, FLAG_SYNC); igt_spin_free(gem_fd, spin); - close(fd[0]); + for (i = 0; i < num_engines; i++) + close(fd[i]); for (i = 0; i < num_engines; i++) val[i] = tval[1][i] - tval[0][i]; @@ -617,22 +620,23 @@ no_sema(int gem_fd, const struct intel_execution_engine2 *e, unsigned int flags) { igt_spin_t *spin; uint64_t val[2][2]; - int fd; + int fd[2]; - fd = open_group(gem_fd, - I915_PMU_ENGINE_SEMA(e->class, e->instance), -1); - open_group(gem_fd, I915_PMU_ENGINE_WAIT(e->class, e->instance), fd); + fd[0] = open_group(gem_fd, I915_PMU_ENGINE_SEMA(e->class, e->instance), + -1); + fd[1] = open_group(gem_fd, I915_PMU_ENGINE_WAIT(e->class, e->instance), + fd[0]); if (flags & TEST_BUSY) spin = spin_sync(gem_fd, 0, e); else spin = NULL; - pmu_read_multi(fd, 2, val[0]); + pmu_read_multi(fd[0], 2, val[0]); measured_usleep(batch_duration_ns / 1000); if (flags & TEST_TRAILING_IDLE) end_spin(gem_fd, spin, flags); - pmu_read_multi(fd, 2, val[1]); + pmu_read_multi(fd[0], 2, val[1]); val[0][0] = val[1][0] - val[0][0]; val[0][1] = val[1][1] - val[0][1]; @@ -641,7 +645,8 @@ no_sema(int gem_fd, const struct intel_execution_engine2 *e, unsigned int flags) end_spin(gem_fd, spin, FLAG_SYNC); igt_spin_free(gem_fd, spin); } - close(fd); + close(fd[0]); + close(fd[1]); assert_within_epsilon(val[0][0], 0.0f, tolerance); assert_within_epsilon(val[0][1], 0.0f, tolerance); @@ -861,18 +866,21 @@ sema_busy(int gem_fd, const struct intel_execution_engine2 *e, unsigned int flags) { - int fd; + int fd[2]; igt_require(intel_gen(intel_get_drm_devid(gem_fd)) >= 8); - fd = open_group(gem_fd, I915_PMU_ENGINE_SEMA(e->class, e->instance), -1); - open_group(gem_fd, I915_PMU_ENGINE_BUSY(e->class, e->instance), fd); + fd[0] = open_group(gem_fd, I915_PMU_ENGINE_SEMA(e->class, e->instance), + -1); + fd[1] = open_group(gem_fd, I915_PMU_ENGINE_BUSY(e->class, e->instance), + fd[0]); - __sema_busy(gem_fd, fd, e, 50, 100); - __sema_busy(gem_fd, fd, e, 25, 50); - __sema_busy(gem_fd, fd, e, 75, 75); + __sema_busy(gem_fd, fd[0], e, 50, 100); + __sema_busy(gem_fd, fd[0], e, 25, 50); + __sema_busy(gem_fd, fd[0], e, 75, 75); - close(fd); + close(fd[0]); + close(fd[1]); } #define MI_WAIT_FOR_PIPE_C_VBLANK (1<<21) @@ -1441,7 +1449,7 @@ test_frequency(int gem_fd) uint64_t val[2], start[2], slept; double min[2], max[2]; igt_spin_t *spin; - int fd, sysfs; + int fd[2], sysfs; sysfs = igt_sysfs_open(gem_fd); igt_require(sysfs >= 0); @@ -1455,8 +1463,8 @@ test_frequency(int gem_fd) igt_require(max_freq > min_freq); igt_require(boost_freq > min_freq); - fd = open_group(gem_fd, I915_PMU_REQUESTED_FREQUENCY, -1); - open_group(gem_fd, I915_PMU_ACTUAL_FREQUENCY, fd); + fd[0] = open_group(gem_fd, I915_PMU_REQUESTED_FREQUENCY, -1); + fd[1] = open_group(gem_fd, I915_PMU_ACTUAL_FREQUENCY, fd[0]); /* * Set GPU to min frequency and read PMU counters. @@ -1471,9 +1479,9 @@ test_frequency(int gem_fd) gem_quiescent_gpu(gem_fd); /* Idle to be sure the change takes effect */ spin = spin_sync_flags(gem_fd, 0, I915_EXEC_DEFAULT); - slept = pmu_read_multi(fd, 2, start); + slept = pmu_read_multi(fd[0], 2, start); measured_usleep(batch_duration_ns / 1000); - slept = pmu_read_multi(fd, 2, val) - slept; + slept = pmu_read_multi(fd[0], 2, val) - slept; min[0] = 1e9*(val[0] - start[0]) / slept; min[1] = 1e9*(val[1] - start[1]) / slept; @@ -1497,9 +1505,9 @@ test_frequency(int gem_fd) gem_quiescent_gpu(gem_fd); spin = spin_sync_flags(gem_fd, 0, I915_EXEC_DEFAULT); - slept = pmu_read_multi(fd, 2, start); + slept = pmu_read_multi(fd[0], 2, start); measured_usleep(batch_duration_ns / 1000); - slept = pmu_read_multi(fd, 2, val) - slept; + slept = pmu_read_multi(fd[0], 2, val) - slept; max[0] = 1e9*(val[0] - start[0]) / slept; max[1] = 1e9*(val[1] - start[1]) / slept; @@ -1514,7 +1522,8 @@ test_frequency(int gem_fd) if (igt_sysfs_get_u32(sysfs, "gt_min_freq_mhz") != min_freq) igt_warn("Unable to restore min frequency to saved value [%u MHz], now %u MHz\n", min_freq, igt_sysfs_get_u32(sysfs, "gt_min_freq_mhz")); - close(fd); + close(fd[0]); + close(fd[1]); igt_info("Min frequency: requested %.1f, actual %.1f\n", min[0], min[1]); @@ -1535,7 +1544,7 @@ test_frequency_idle(int gem_fd) uint32_t min_freq; uint64_t val[2], start[2], slept; double idle[2]; - int fd, sysfs; + int fd[2], sysfs; sysfs = igt_sysfs_open(gem_fd); igt_require(sysfs >= 0); @@ -1545,17 +1554,18 @@ test_frequency_idle(int gem_fd) /* While parked, our convention is to report the GPU at 0Hz */ - fd = open_group(gem_fd, I915_PMU_REQUESTED_FREQUENCY, -1); - open_group(gem_fd, I915_PMU_ACTUAL_FREQUENCY, fd); + fd[0] = open_group(gem_fd, I915_PMU_REQUESTED_FREQUENCY, -1); + fd[1] = open_group(gem_fd, I915_PMU_ACTUAL_FREQUENCY, fd[0]); gem_quiescent_gpu(gem_fd); /* Be idle! */ measured_usleep(2000); /* Wait for timers to cease */ - slept = pmu_read_multi(fd, 2, start); + slept = pmu_read_multi(fd[0], 2, start); measured_usleep(batch_duration_ns / 1000); - slept = pmu_read_multi(fd, 2, val) - slept; + slept = pmu_read_multi(fd[0], 2, val) - slept; - close(fd); + close(fd[0]); + close(fd[1]); idle[0] = 1e9*(val[0] - start[0]) / slept; idle[1] = 1e9*(val[1] - start[1]) / slept; @@ -1926,57 +1936,73 @@ static int unload_i915(void) return 0; } -static void test_unload(void) +static void test_unload(unsigned int num_engines) { igt_fork(child, 1) { const struct intel_execution_engine2 *e; + int fd[4 + num_engines], i; uint64_t *buf; - int count = 1; + int count = 0; int i915; - int fd; i915 = __drm_open_driver(DRIVER_INTEL); igt_debug("Opening perf events\n"); - fd = open_group(i915, I915_PMU_INTERRUPTS, -1); + fd[count] = open_group(i915, I915_PMU_INTERRUPTS, -1); + if (fd[count] != -1) + count++; - if (perf_i915_open_group(i915, I915_PMU_REQUESTED_FREQUENCY,fd) != -1) + fd[count] = perf_i915_open_group(i915, + I915_PMU_REQUESTED_FREQUENCY, + fd[count - 1]); + if (fd[count] != -1) count++; - if (perf_i915_open_group(i915, I915_PMU_ACTUAL_FREQUENCY, fd) != -1) + + fd[count] = perf_i915_open_group(i915, + I915_PMU_ACTUAL_FREQUENCY, + fd[count - 1]); + if (fd[count] != -1) count++; __for_each_physical_engine(i915, e) { - if (perf_i915_open_group(i915, - I915_PMU_ENGINE_BUSY(e->class, e->instance), - fd) != -1) + fd[count] = perf_i915_open_group(i915, + I915_PMU_ENGINE_BUSY(e->class, e->instance), + fd[count - 1]); + if (fd[count] != -1) count++; - if (perf_i915_open_group(i915, - I915_PMU_ENGINE_SEMA(e->class, e->instance), - fd) != -1) + fd[count] = perf_i915_open_group(i915, + I915_PMU_ENGINE_SEMA(e->class, e->instance), + fd[count - 1]); + if (fd[count] != -1) count++; - if (perf_i915_open_group(i915, - I915_PMU_ENGINE_WAIT(e->class, e->instance), - fd) != -1) + fd[count] = perf_i915_open_group(i915, + I915_PMU_ENGINE_WAIT(e->class, e->instance), + fd[count - 1]); + if (fd[count] != -1) count++; } - if (perf_i915_open_group(i915, I915_PMU_RC6_RESIDENCY,fd) != -1) + fd[count] = perf_i915_open_group(i915, I915_PMU_RC6_RESIDENCY, + fd[count - 1]); + if (fd[count] != -1) count++; close(i915); - buf = calloc(count + 1, sizeof(uint64_t)); + buf = calloc(count, sizeof(uint64_t)); igt_assert(buf); igt_debug("Read %d events from perf and trial unload\n", count); - pmu_read_multi(fd, count, buf); + pmu_read_multi(fd[0], count, buf); igt_assert_eq(unload_i915(), -EBUSY); - pmu_read_multi(fd, count, buf); + pmu_read_multi(fd[0], count, buf); igt_debug("Close perf\n"); - close(fd); + + for (i = 0; i < count; i++) + close(fd[i]); free(buf); } @@ -2357,6 +2383,6 @@ igt_main igt_subtest("module-unload") { igt_require(unload_i915() == 0); for (int pass = 0; pass < 3; pass++) - test_unload(); + test_unload(num_engines); } } -- 2.25.1 _______________________________________________ igt-dev mailing list igt-dev@lists.freedesktop.org https://lists.freedesktop.org/mailman/listinfo/igt-dev
next reply other threads:[~2020-10-13 9:46 UTC|newest] Thread overview: 13+ messages / expand[flat|nested] mbox.gz Atom feed top 2020-10-13 9:46 Tvrtko Ursulin [this message] 2020-10-13 9:46 ` [igt-dev] [PATCH i-g-t] i915/perf_pmu: Fix perf fd leak Tvrtko Ursulin 2020-10-13 9:52 ` [Intel-gfx] " Tvrtko Ursulin 2020-10-13 9:52 ` [igt-dev] " Tvrtko Ursulin 2020-10-13 9:54 ` [Intel-gfx] [PATCH i-g-t v2] " Tvrtko Ursulin 2020-10-13 9:54 ` [igt-dev] " Tvrtko Ursulin 2020-10-13 10:04 ` [Intel-gfx] [igt-dev] [PATCH i-g-t] " Chris Wilson 2020-10-13 10:34 ` Tvrtko Ursulin 2020-10-13 10:34 ` Tvrtko Ursulin 2020-10-13 10:40 ` [Intel-gfx] " Chris Wilson 2020-10-13 10:40 ` Chris Wilson 2020-10-13 11:03 ` [igt-dev] ✓ Fi.CI.BAT: success for i915/perf_pmu: Fix perf fd leak (rev2) Patchwork 2020-10-13 23:22 ` [igt-dev] ✗ Fi.CI.IGT: failure " Patchwork
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=20201013094612.83843-1-tvrtko.ursulin@linux.intel.com \ --to=tvrtko.ursulin@linux.intel.com \ --cc=Intel-gfx@lists.freedesktop.org \ --cc=igt-dev@lists.freedesktop.org \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.