Message ID | 1384279096-10886-1-git-send-email-mika.kuoppala@intel.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On Tue, Nov 12, 2013 at 07:58:16PM +0200, Mika Kuoppala wrote: > v2: check the ioctl pad and flag parameters > > Signed-off-by: Mika Kuoppala <mika.kuoppala@intel.com> I've merged this to igt, but there are a few small fixups to do on top: - We now have the igt_main macro to cut down a bit on boilerplate. - I haven't tested it, but I guess inject_hang will cause some *ERROR* noise in dmesg. Rule is that igt testcases should only cause info/debug level messages, everything else is considered a failure. I think we could fix this by setting the stop_rings debugfs value right _after_ the hang is injected, to tell the kernel that the hang it'll see is actually fake. - The userspace interface checking has two missing spots: a) checking that lookup for an invalid ctx id fails with ENOENT b) checking that non-root can't read out the default context. For the later it's probably simples to fork a 2nd process and drop the CAP_SYS_ADMIN capability in there. Thanks, Daniel > --- > tests/Makefile.sources | 1 + > tests/gem_reset_stats.c | 770 +++++++++++++++++++++++++++++++++++++++++++++++ > 2 files changed, 771 insertions(+) > create mode 100644 tests/gem_reset_stats.c > > diff --git a/tests/Makefile.sources b/tests/Makefile.sources > index 5f4d282..a02b93d 100644 > --- a/tests/Makefile.sources > +++ b/tests/Makefile.sources > @@ -182,6 +182,7 @@ HANG = \ > gem_bad_blit \ > gem_bad_address \ > gem_non_secure_batch \ > + gem_reset_stats \ > $(NULL) > > scripts = \ > diff --git a/tests/gem_reset_stats.c b/tests/gem_reset_stats.c > new file mode 100644 > index 0000000..2040294 > --- /dev/null > +++ b/tests/gem_reset_stats.c > @@ -0,0 +1,770 @@ > +/* > + * Copyright (c) 2013 Intel Corporation > + * > + * Permission is hereby granted, free of charge, to any person obtaining a > + * copy of this software and associated documentation files (the "Software"), > + * to deal in the Software without restriction, including without limitation > + * the rights to use, copy, modify, merge, publish, distribute, sublicense, > + * and/or sell copies of the Software, and to permit persons to whom the > + * Software is furnished to do so, subject to the following conditions: > + * > + * The above copyright notice and this permission notice (including the next > + * paragraph) shall be included in all copies or substantial portions of the > + * Software. > + * > + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR > + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, > + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL > + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER > + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING > + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS > + * IN THE SOFTWARE. > + * > + * Authors: > + * Mika Kuoppala <mika.kuoppala@intel.com> > + * > + */ > + > +#include <unistd.h> > +#include <stdlib.h> > +#include <stdio.h> > +#include <string.h> > +#include <fcntl.h> > +#include <inttypes.h> > +#include <errno.h> > +#include <sys/stat.h> > +#include <sys/ioctl.h> > +#include <sys/mman.h> > +#include <time.h> > + > +#include "i915_drm.h" > +#include "intel_bufmgr.h" > +#include "intel_batchbuffer.h" > +#include "intel_gpu_tools.h" > +#include "rendercopy.h" > + > +#define RS_NO_ERROR 0 > +#define RS_BATCH_ACTIVE (1 << 0) > +#define RS_BATCH_PENDING (1 << 1) > +#define RS_UNKNOWN (1 << 2) > + > +struct local_drm_i915_reset_stats { > + __u32 ctx_id; > + __u32 flags; > + __u32 reset_count; > + __u32 batch_active; > + __u32 batch_pending; > + __u32 pad; > +}; > + > +struct local_drm_i915_gem_context_create { > + __u32 ctx_id; > + __u32 pad; > +}; > + > +struct local_drm_i915_gem_context_destroy { > + __u32 ctx_id; > + __u32 pad; > +}; > + > +#define MAX_FD 32 > + > +#define CONTEXT_CREATE_IOCTL DRM_IOWR(DRM_COMMAND_BASE + 0x2d, struct local_drm_i915_gem_context_create) > +#define CONTEXT_DESTROY_IOCTL DRM_IOWR(DRM_COMMAND_BASE + 0x2e, struct local_drm_i915_gem_context_destroy) > +#define GET_RESET_STATS_IOCTL DRM_IOWR(DRM_COMMAND_BASE + 0x32, struct local_drm_i915_reset_stats) > + > +static uint32_t context_create(int fd) > +{ > + struct local_drm_i915_gem_context_create create; > + int ret; > + > + create.ctx_id = rand(); > + create.pad = rand(); > + > + ret = drmIoctl(fd, CONTEXT_CREATE_IOCTL, &create); > + igt_assert(ret == 0); > + > + return create.ctx_id; > +} > + > +static int context_destroy(int fd, uint32_t ctx_id) > +{ > + int ret; > + struct local_drm_i915_gem_context_destroy destroy; > + > + destroy.ctx_id = ctx_id; > + destroy.pad = rand(); > + > + ret = drmIoctl(fd, CONTEXT_DESTROY_IOCTL, &destroy); > + if (ret != 0) > + return -errno; > + > + return 0; > +} > + > +static int gem_reset_stats(int fd, int ctx_id, > + struct local_drm_i915_reset_stats *rs) > +{ > + int ret; > + > + rs->ctx_id = ctx_id; > + rs->flags = 0; > + rs->reset_count = rand(); > + rs->batch_active = rand(); > + rs->batch_pending = rand(); > + rs->pad = 0; > + > + do { > + ret = ioctl(fd, GET_RESET_STATS_IOCTL, rs); > + } while (ret == -1 && (errno == EINTR || errno == EAGAIN)); > + > + if (ret < 0) > + return -errno; > + > + return 0; > +} > + > +static int gem_reset_status(int fd, int ctx_id) > +{ > + int ret; > + struct local_drm_i915_reset_stats rs; > + > + ret = gem_reset_stats(fd, ctx_id, &rs); > + if (ret) > + return ret; > + > + if (rs.batch_active) > + return RS_BATCH_ACTIVE; > + if (rs.batch_pending) > + return RS_BATCH_PENDING; > + > + return RS_NO_ERROR; > +} > + > +static int gem_exec(int fd, struct drm_i915_gem_execbuffer2 *execbuf) > +{ > + int ret; > + > + ret = ioctl(fd, > + DRM_IOCTL_I915_GEM_EXECBUFFER2, > + execbuf); > + > + if (ret < 0) > + return -errno; > + > + return 0; > +} > + > +static int exec_valid(int fd, int ctx) > +{ > + struct drm_i915_gem_execbuffer2 execbuf; > + struct drm_i915_gem_exec_object2 exec; > + int ret; > + > + uint32_t buf[2] = { MI_BATCH_BUFFER_END, 0 }; > + > + exec.handle = gem_create(fd, 4096); > + gem_write(fd, exec.handle, 0, buf, sizeof(buf)); > + exec.relocation_count = 0; > + exec.relocs_ptr = 0; > + exec.alignment = 0; > + exec.offset = 0; > + exec.flags = 0; > + exec.rsvd1 = 0; > + exec.rsvd2 = 0; > + > + execbuf.buffers_ptr = (uintptr_t)&exec; > + execbuf.buffer_count = 1; > + execbuf.batch_start_offset = 0; > + execbuf.batch_len = sizeof(buf); > + execbuf.cliprects_ptr = 0; > + execbuf.num_cliprects = 0; > + execbuf.DR1 = 0; > + execbuf.DR4 = 0; > + execbuf.flags = 0; > + i915_execbuffer2_set_context_id(execbuf, ctx); > + execbuf.rsvd2 = 0; > + > + ret = gem_exec(fd, &execbuf); > + if (ret < 0) > + return ret; > + > + return exec.handle; > +} > + > +#define BUFSIZE (4 * 1024) > +#define ITEMS (BUFSIZE >> 2) > + > +static int inject_hang(int fd, int ctx) > +{ > + struct drm_i915_gem_execbuffer2 execbuf; > + struct drm_i915_gem_exec_object2 exec; > + uint64_t gtt_off; > + uint32_t *buf; > + int roff, i; > + > + srandom(time(NULL)); > + > + buf = malloc(BUFSIZE); > + igt_assert(buf != NULL); > + > + buf[0] = MI_BATCH_BUFFER_END; > + buf[1] = MI_NOOP; > + > + exec.handle = gem_create(fd, BUFSIZE); > + gem_write(fd, exec.handle, 0, buf, BUFSIZE); > + exec.relocation_count = 0; > + exec.relocs_ptr = 0; > + exec.alignment = 0; > + exec.offset = 0; > + exec.flags = 0; > + exec.rsvd1 = 0; > + exec.rsvd2 = 0; > + > + execbuf.buffers_ptr = (uintptr_t)&exec; > + execbuf.buffer_count = 1; > + execbuf.batch_start_offset = 0; > + execbuf.batch_len = BUFSIZE; > + execbuf.cliprects_ptr = 0; > + execbuf.num_cliprects = 0; > + execbuf.DR1 = 0; > + execbuf.DR4 = 0; > + execbuf.flags = 0; > + i915_execbuffer2_set_context_id(execbuf, ctx); > + execbuf.rsvd2 = 0; > + > + igt_assert(gem_exec(fd, &execbuf) == 0); > + > + gtt_off = exec.offset; > + > + for (i = 0; i < ITEMS; i++) > + buf[i] = MI_NOOP; > + > + roff = random() % (ITEMS - 2); > + buf[roff] = MI_BATCH_BUFFER_START; > + buf[roff + 1] = gtt_off + (roff << 2); > + > +#ifdef VERBOSE > + printf("loop injected at 0x%lx (off 0x%x, bo_start 0x%lx, bo_end 0x%lx)\n", > + (long unsigned int)((roff << 2) + gtt_off), > + roff << 2, (long unsigned int)gtt_off, > + (long unsigned int)(gtt_off + BUFSIZE - 1)); > +#endif > + gem_write(fd, exec.handle, 0, buf, BUFSIZE); > + > + exec.relocation_count = 0; > + exec.relocs_ptr = 0; > + exec.alignment = 0; > + exec.offset = 0; > + exec.flags = 0; > + exec.rsvd1 = 0; > + exec.rsvd2 = 0; > + > + execbuf.buffers_ptr = (uintptr_t)&exec; > + execbuf.buffer_count = 1; > + execbuf.batch_start_offset = 0; > + execbuf.batch_len = BUFSIZE; > + execbuf.cliprects_ptr = 0; > + execbuf.num_cliprects = 0; > + execbuf.DR1 = 0; > + execbuf.DR4 = 0; > + execbuf.flags = 0; > + i915_execbuffer2_set_context_id(execbuf, ctx); > + execbuf.rsvd2 = 0; > + > + igt_assert(gem_exec(fd, &execbuf) == 0); > + > + igt_assert(gtt_off == exec.offset); > + > + free(buf); > + > + return exec.handle; > +} > + > +static int _assert_reset_status(int fd, int ctx, int status) > +{ > + int rs; > + > + rs = gem_reset_status(fd, ctx); > + if (rs < 0) { > + printf("reset status for %d ctx %d returned %d\n", > + fd, ctx, rs); > + return rs; > + } > + > + if (rs != status) { > + printf("%d:%d reset status %d differs from assumed %d\n", > + fd, ctx, rs, status); > + > + return 1; > + } > + > + return 0; > +} > + > +#define assert_reset_status(fd, ctx, status) \ > + igt_assert(_assert_reset_status(fd, ctx, status) == 0) > + > +static void test_rs(int num_fds, int hang_index, int rs_assumed_no_hang) > +{ > + int i; > + int fd[MAX_FD]; > + int h[MAX_FD]; > + > + igt_assert (num_fds <= MAX_FD); > + igt_assert (hang_index < MAX_FD); > + > + for (i = 0; i < num_fds; i++) { > + fd[i] = drm_open_any(); > + igt_assert(fd[i]); > + } > + > + for (i = 0; i < num_fds; i++) > + assert_reset_status(fd[i], 0, RS_NO_ERROR); > + > + for (i = 0; i < num_fds; i++) { > + if (i == hang_index) > + h[i] = inject_hang(fd[i], 0); > + else > + h[i] = exec_valid(fd[i], 0); > + } > + > + gem_sync(fd[num_fds - 1], h[num_fds - 1]); > + > + for (i = 0; i < num_fds; i++) { > + if (hang_index < 0) { > + assert_reset_status(fd[i], 0, rs_assumed_no_hang); > + continue; > + } > + > + if (i < hang_index) > + assert_reset_status(fd[i], 0, RS_NO_ERROR); > + if (i == hang_index) > + assert_reset_status(fd[i], 0, RS_BATCH_ACTIVE); > + if (i > hang_index) > + assert_reset_status(fd[i], 0, RS_BATCH_PENDING); > + } > + > + for (i = 0; i < num_fds; i++) { > + gem_close(fd[i], h[i]); > + close(fd[i]); > + } > +} > + > +#define MAX_CTX 100 > +static void test_rs_ctx(int num_fds, int num_ctx, int hang_index, > + int hang_context) > +{ > + int i, j; > + int fd[MAX_FD]; > + int h[MAX_FD][MAX_CTX]; > + int ctx[MAX_FD][MAX_CTX]; > + > + igt_assert (num_fds <= MAX_FD); > + igt_assert (hang_index < MAX_FD); > + > + igt_assert (num_ctx <= MAX_CTX); > + igt_assert (hang_context < MAX_CTX); > + > + test_rs(num_fds, -1, RS_NO_ERROR); > + > + for (i = 0; i < num_fds; i++) { > + fd[i] = drm_open_any(); > + igt_assert(fd[i]); > + assert_reset_status(fd[i], 0, RS_NO_ERROR); > + > + for (j = 0; j < num_ctx; j++) { > + ctx[i][j] = context_create(fd[i]); > + > + } > + > + assert_reset_status(fd[i], 0, RS_NO_ERROR); > + } > + > + for (i = 0; i < num_fds; i++) { > + > + assert_reset_status(fd[i], 0, RS_NO_ERROR); > + > + for (j = 0; j < num_ctx; j++) > + assert_reset_status(fd[i], ctx[i][j], RS_NO_ERROR); > + > + assert_reset_status(fd[i], 0, RS_NO_ERROR); > + } > + > + for (i = 0; i < num_fds; i++) { > + for (j = 0; j < num_ctx; j++) { > + if (i == hang_index && j == hang_context) > + h[i][j] = inject_hang(fd[i], ctx[i][j]); > + else > + h[i][j] = exec_valid(fd[i], ctx[i][j]); > + } > + } > + > + gem_sync(fd[num_fds - 1], ctx[num_fds - 1][num_ctx - 1]); > + > + for (i = 0; i < num_fds; i++) > + assert_reset_status(fd[i], 0, RS_NO_ERROR); > + > + for (i = 0; i < num_fds; i++) { > + for (j = 0; j < num_ctx; j++) { > + if (i < hang_index) > + assert_reset_status(fd[i], ctx[i][j], RS_NO_ERROR); > + if (i == hang_index && j < hang_context) > + assert_reset_status(fd[i], ctx[i][j], RS_NO_ERROR); > + if (i == hang_index && j == hang_context) > + assert_reset_status(fd[i], ctx[i][j], > + RS_BATCH_ACTIVE); > + if (i == hang_index && j > hang_context) > + assert_reset_status(fd[i], ctx[i][j], > + RS_BATCH_PENDING); > + if (i > hang_index) > + assert_reset_status(fd[i], ctx[i][j], > + RS_BATCH_PENDING); > + } > + } > + > + for (i = 0; i < num_fds; i++) { > + for (j = 0; j < num_ctx; j++) { > + gem_close(fd[i], h[i][j]); > + igt_assert(context_destroy(fd[i], ctx[i][j]) == 0); > + } > + > + assert_reset_status(fd[i], 0, RS_NO_ERROR); > + > + close(fd[i]); > + } > +} > + > +static void test_ban(void) > +{ > + int h1,h2,h3,h4,h5,h6,h7; > + int ctx_good, ctx_bad; > + int fd; > + int retry = 10; > + int active_count = 0, pending_count = 0; > + struct local_drm_i915_reset_stats rs_bad, rs_good; > + > + fd = drm_open_any(); > + igt_assert(fd >= 0); > + > + assert_reset_status(fd, 0, RS_NO_ERROR); > + > + ctx_good = context_create(fd); > + ctx_bad = context_create(fd); > + > + assert_reset_status(fd, 0, RS_NO_ERROR); > + assert_reset_status(fd, ctx_good, RS_NO_ERROR); > + assert_reset_status(fd, ctx_bad, RS_NO_ERROR); > + > + h1 = exec_valid(fd, ctx_bad); > + igt_assert(h1 >= 0); > + h5 = exec_valid(fd, ctx_good); > + igt_assert(h5 >= 0); > + > + assert_reset_status(fd, ctx_good, RS_NO_ERROR); > + assert_reset_status(fd, ctx_bad, RS_NO_ERROR); > + > + h2 = inject_hang(fd, ctx_bad); > + igt_assert(h2 >= 0); > + active_count++; > + /* Second hang will be pending for this */ > + pending_count++; > + > + h6 = exec_valid(fd, ctx_good); > + h7 = exec_valid(fd, ctx_good); > + > + while (retry--) { > + h3 = inject_hang(fd, ctx_bad); > + igt_assert(h3 >= 0); > + gem_sync(fd, h3); > + active_count++; > + /* This second hand will count as pending */ > + assert_reset_status(fd, ctx_bad, RS_BATCH_ACTIVE); > + > + h4 = exec_valid(fd, ctx_bad); > + if (h4 == -EIO) { > + gem_close(fd, h3); > + break; > + } > + > + /* Should not happen often but sometimes hang is declared too slow > + * due to our way of faking hang using loop */ > + > + igt_assert(h4 >= 0); > + gem_close(fd, h3); > + gem_close(fd, h4); > + > + printf("retrying for ban (%d)\n", retry); > + } > + > + igt_assert(h4 == -EIO); > + assert_reset_status(fd, ctx_bad, RS_BATCH_ACTIVE); > + > + gem_sync(fd, h7); > + assert_reset_status(fd, ctx_good, RS_BATCH_PENDING); > + > + igt_assert(gem_reset_stats(fd, ctx_good, &rs_good) == 0); > + igt_assert(gem_reset_stats(fd, ctx_bad, &rs_bad) == 0); > + > + igt_assert(rs_bad.batch_active == active_count); > + igt_assert(rs_bad.batch_pending == pending_count); > + igt_assert(rs_good.batch_active == 0); > + igt_assert(rs_good.batch_pending == 2); > + > + gem_close(fd, h1); > + gem_close(fd, h2); > + gem_close(fd, h6); > + gem_close(fd, h7); > + > + h1 = exec_valid(fd, ctx_good); > + igt_assert(h1 >= 0); > + gem_close(fd, h1); > + > + igt_assert(context_destroy(fd, ctx_good) == 0); > + igt_assert(context_destroy(fd, ctx_bad) == 0); > + igt_assert(gem_reset_status(fd, ctx_good) < 0); > + igt_assert(gem_reset_status(fd, ctx_bad) < 0); > + igt_assert(exec_valid(fd, ctx_good) < 0); > + igt_assert(exec_valid(fd, ctx_bad) < 0); > + > + close(fd); > +} > + > +static void test_nonrelated_hang(void) > +{ > + int h1,h2; > + int fd1,fd2; > + int ctx_guilty, ctx_unrelated; > + > + fd1 = drm_open_any(); > + fd2 = drm_open_any(); > + assert_reset_status(fd1, 0, RS_NO_ERROR); > + assert_reset_status(fd2, 0, RS_NO_ERROR); > + ctx_guilty = context_create(fd1); > + ctx_unrelated = context_create(fd2); > + > + assert_reset_status(fd1, ctx_guilty, RS_NO_ERROR); > + assert_reset_status(fd2, ctx_unrelated, RS_NO_ERROR); > + > + h1 = inject_hang(fd1, ctx_guilty); > + igt_assert(h1 >= 0); > + gem_sync(fd1, h1); > + assert_reset_status(fd1, ctx_guilty, RS_BATCH_ACTIVE); > + assert_reset_status(fd2, ctx_unrelated, RS_NO_ERROR); > + > + h2 = exec_valid(fd2, ctx_unrelated); > + igt_assert(h2 >= 0); > + gem_sync(fd2, h2); > + assert_reset_status(fd1, ctx_guilty, RS_BATCH_ACTIVE); > + assert_reset_status(fd2, ctx_unrelated, RS_NO_ERROR); > + gem_close(fd1, h1); > + gem_close(fd2, h2); > + > + igt_assert(context_destroy(fd1, ctx_guilty) == 0); > + igt_assert(context_destroy(fd2, ctx_unrelated) == 0); > + > + close(fd1); > + close(fd2); > +} > + > +static int get_reset_count(int fd, int ctx) > +{ > + int ret; > + struct local_drm_i915_reset_stats rs; > + > + ret = gem_reset_stats(fd, ctx, &rs); > + if (ret) > + return ret; > + > + return rs.reset_count; > +} > + > +static void test_double_destroy_pending(void) > +{ > + int fd, h; > + uint32_t ctx; > + > + fd = drm_open_any(); > + igt_assert(fd >= 0); > + ctx = context_create(fd); > + > + assert_reset_status(fd, ctx, RS_NO_ERROR); > + > + h = inject_hang(fd, ctx); > + igt_assert(h >= 0); > + igt_assert(context_destroy(fd, ctx) == 0); > + igt_assert(context_destroy(fd, ctx) == -ENOENT); > + > + gem_close(fd, h); > + close(fd); > +} > + > +static void test_close_pending(void) > +{ > + int fd, h; > + > + fd = drm_open_any(); > + igt_assert(fd >= 0); > + > + assert_reset_status(fd, 0, RS_NO_ERROR); > + > + h = inject_hang(fd, 0); > + igt_assert(h >= 0); > + > + gem_close(fd, h); > + close(fd); > +} > + > +static void __test_count(const bool create_ctx) > +{ > + int fd, h, ctx; > + long c1, c2; > + > + fd = drm_open_any(); > + igt_assert(fd >= 0); > + if (create_ctx) > + ctx = context_create(fd); > + else > + ctx = 0; > + > + assert_reset_status(fd, ctx, RS_NO_ERROR); > + > + c1 = get_reset_count(fd, ctx); > + igt_assert(c1 >= 0); > + > + h = inject_hang(fd, ctx); > + igt_assert (h >= 0); > + gem_sync(fd, h); > + > + assert_reset_status(fd, ctx, RS_BATCH_ACTIVE); > + c2 = get_reset_count(fd, ctx); > + igt_assert(c2 >= 0); > + > + igt_assert(c2 == (c1 + 1)); > + > + gem_close(fd, h); > + > + if (create_ctx) > + context_destroy(fd, ctx); > + > + close(fd); > +} > + > +static void test_count(void) > +{ > + return __test_count(false); > +} > + > +static void test_count_context(void) > +{ > + return __test_count(true); > +} > + > +static void test_global_reset_count(void) > +{ > + test_count(); > + test_count_context(); > +} > + > +static int _test_params(int fd, int ctx, uint32_t flags, uint32_t pad) > +{ > + struct local_drm_i915_reset_stats rs; > + int ret; > + > + rs.ctx_id = ctx; > + rs.flags = flags; > + rs.reset_count = rand(); > + rs.batch_active = rand(); > + rs.batch_pending = rand(); > + rs.pad = pad; > + > + do { > + ret = ioctl(fd, GET_RESET_STATS_IOCTL, &rs); > + } while (ret == -1 && (errno == EINTR || errno == EAGAIN)); > + > + if (ret < 0) > + return -errno; > + > + return 0; > +} > + > +static void test_param_ctx(int fd, int ctx) > +{ > + const uint32_t bad = rand() + 1; > + > + igt_assert(_test_params(fd, ctx, 0, 0) == 0); > + igt_assert(_test_params(fd, ctx, 0, bad) == -EINVAL); > + igt_assert(_test_params(fd, ctx, bad, 0) == -EINVAL); > + igt_assert(_test_params(fd, ctx, bad, bad) == -EINVAL); > +} > + > +static void test_params(void) > +{ > + int fd, ctx; > + > + fd = drm_open_any(); > + igt_assert(fd >= 0); > + ctx = context_create(fd); > + > + igt_assert(ioctl(fd, GET_RESET_STATS_IOCTL, 0) == -1); > + > + igt_assert(_test_params(fd, 0xbadbad, 0, 0) == -ENOENT); > + > + test_param_ctx(fd, 0); > + test_param_ctx(fd, ctx); > + > + close(fd); > +} > + > + > +int main(int argc, char **argv) > +{ > + struct local_drm_i915_gem_context_create create; > + uint32_t devid; > + int fd; > + int ret; > + > + igt_skip_on_simulation(); > + > + igt_subtest_init(argc, argv); > + > + igt_fixture { > + fd = drm_open_any(); > + devid = intel_get_drm_devid(fd); > + if (intel_gen(devid) < 4) > + igt_skip("Architecture %d too old\n", intel_gen(devid)); > + > + ret = drmIoctl(fd, CONTEXT_CREATE_IOCTL, &create); > + if (ret != 0 && (errno == ENODEV || errno == EINVAL)) > + igt_skip("Kernel is too old, or contexts not supported: %s\n", > + strerror(errno)); > + > + close(fd); > + } > + > + igt_subtest("basic-reset-status") > + test_rs(4, 1, 0); > + > + igt_subtest("context-reset-status") > + test_rs_ctx(4, 4, 1, 2); > + > + igt_subtest("ban") > + test_ban(); > + > + igt_subtest("ctx-unrelated") > + test_nonrelated_hang(); > + > + igt_subtest("global-count") > + test_global_reset_count(); > + > + igt_subtest("double-destroy-pending") > + test_double_destroy_pending(); > + > + igt_subtest("close-pending") > + test_close_pending(); > + > + igt_subtest("params") > + test_params(); > + > + igt_exit(); > +} > -- > 1.7.9.5 > > _______________________________________________ > Intel-gfx mailing list > Intel-gfx@lists.freedesktop.org > http://lists.freedesktop.org/mailman/listinfo/intel-gfx
Daniel Vetter <daniel@ffwll.ch> writes: > On Tue, Nov 12, 2013 at 07:58:16PM +0200, Mika Kuoppala wrote: >> v2: check the ioctl pad and flag parameters >> >> Signed-off-by: Mika Kuoppala <mika.kuoppala@intel.com> > > I've merged this to igt, but there are a few small fixups to do on top: > - We now have the igt_main macro to cut down a bit on boilerplate. > - I haven't tested it, but I guess inject_hang will cause some *ERROR* > noise in dmesg. Rule is that igt testcases should only cause info/debug > level messages, everything else is considered a failure. I think we > could fix this by setting the stop_rings debugfs value right _after_ the > hang is injected, to tell the kernel that the hang it'll see is actually > fake. I will take a look what stop_rings do. I just have a feeling that with this kind of trickery we shoot ourselves in foot some day. Would be good that we get a proper error messages when we submit a real hanging batches. Would be also good that our test checks that the ERROR msg was really emitted. Having whitelist of expected '*ERROR*' messages for these kind of tests is not an option? > - The userspace interface checking has two missing spots: a) checking that > lookup for an invalid ctx id fails with ENOENT b) checking that non-root > can't read out the default context. For the later it's probably simples > to fork a 2nd process and drop the CAP_SYS_ADMIN capability in there. I check for ENOENT on submitting bad context id there. Did you miss it or it is not enough? I will resubmit when I have b) in place. -Mika > Thanks, Daniel >> --- >> tests/Makefile.sources | 1 + >> tests/gem_reset_stats.c | 770 +++++++++++++++++++++++++++++++++++++++++++++++ >> 2 files changed, 771 insertions(+) >> create mode 100644 tests/gem_reset_stats.c >> >> diff --git a/tests/Makefile.sources b/tests/Makefile.sources >> index 5f4d282..a02b93d 100644 >> --- a/tests/Makefile.sources >> +++ b/tests/Makefile.sources >> @@ -182,6 +182,7 @@ HANG = \ >> gem_bad_blit \ >> gem_bad_address \ >> gem_non_secure_batch \ >> + gem_reset_stats \ >> $(NULL) >> >> scripts = \ >> diff --git a/tests/gem_reset_stats.c b/tests/gem_reset_stats.c >> new file mode 100644 >> index 0000000..2040294 >> --- /dev/null >> +++ b/tests/gem_reset_stats.c >> @@ -0,0 +1,770 @@ >> +/* >> + * Copyright (c) 2013 Intel Corporation >> + * >> + * Permission is hereby granted, free of charge, to any person obtaining a >> + * copy of this software and associated documentation files (the "Software"), >> + * to deal in the Software without restriction, including without limitation >> + * the rights to use, copy, modify, merge, publish, distribute, sublicense, >> + * and/or sell copies of the Software, and to permit persons to whom the >> + * Software is furnished to do so, subject to the following conditions: >> + * >> + * The above copyright notice and this permission notice (including the next >> + * paragraph) shall be included in all copies or substantial portions of the >> + * Software. >> + * >> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR >> + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, >> + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL >> + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER >> + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING >> + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS >> + * IN THE SOFTWARE. >> + * >> + * Authors: >> + * Mika Kuoppala <mika.kuoppala@intel.com> >> + * >> + */ >> + >> +#include <unistd.h> >> +#include <stdlib.h> >> +#include <stdio.h> >> +#include <string.h> >> +#include <fcntl.h> >> +#include <inttypes.h> >> +#include <errno.h> >> +#include <sys/stat.h> >> +#include <sys/ioctl.h> >> +#include <sys/mman.h> >> +#include <time.h> >> + >> +#include "i915_drm.h" >> +#include "intel_bufmgr.h" >> +#include "intel_batchbuffer.h" >> +#include "intel_gpu_tools.h" >> +#include "rendercopy.h" >> + >> +#define RS_NO_ERROR 0 >> +#define RS_BATCH_ACTIVE (1 << 0) >> +#define RS_BATCH_PENDING (1 << 1) >> +#define RS_UNKNOWN (1 << 2) >> + >> +struct local_drm_i915_reset_stats { >> + __u32 ctx_id; >> + __u32 flags; >> + __u32 reset_count; >> + __u32 batch_active; >> + __u32 batch_pending; >> + __u32 pad; >> +}; >> + >> +struct local_drm_i915_gem_context_create { >> + __u32 ctx_id; >> + __u32 pad; >> +}; >> + >> +struct local_drm_i915_gem_context_destroy { >> + __u32 ctx_id; >> + __u32 pad; >> +}; >> + >> +#define MAX_FD 32 >> + >> +#define CONTEXT_CREATE_IOCTL DRM_IOWR(DRM_COMMAND_BASE + 0x2d, struct local_drm_i915_gem_context_create) >> +#define CONTEXT_DESTROY_IOCTL DRM_IOWR(DRM_COMMAND_BASE + 0x2e, struct local_drm_i915_gem_context_destroy) >> +#define GET_RESET_STATS_IOCTL DRM_IOWR(DRM_COMMAND_BASE + 0x32, struct local_drm_i915_reset_stats) >> + >> +static uint32_t context_create(int fd) >> +{ >> + struct local_drm_i915_gem_context_create create; >> + int ret; >> + >> + create.ctx_id = rand(); >> + create.pad = rand(); >> + >> + ret = drmIoctl(fd, CONTEXT_CREATE_IOCTL, &create); >> + igt_assert(ret == 0); >> + >> + return create.ctx_id; >> +} >> + >> +static int context_destroy(int fd, uint32_t ctx_id) >> +{ >> + int ret; >> + struct local_drm_i915_gem_context_destroy destroy; >> + >> + destroy.ctx_id = ctx_id; >> + destroy.pad = rand(); >> + >> + ret = drmIoctl(fd, CONTEXT_DESTROY_IOCTL, &destroy); >> + if (ret != 0) >> + return -errno; >> + >> + return 0; >> +} >> + >> +static int gem_reset_stats(int fd, int ctx_id, >> + struct local_drm_i915_reset_stats *rs) >> +{ >> + int ret; >> + >> + rs->ctx_id = ctx_id; >> + rs->flags = 0; >> + rs->reset_count = rand(); >> + rs->batch_active = rand(); >> + rs->batch_pending = rand(); >> + rs->pad = 0; >> + >> + do { >> + ret = ioctl(fd, GET_RESET_STATS_IOCTL, rs); >> + } while (ret == -1 && (errno == EINTR || errno == EAGAIN)); >> + >> + if (ret < 0) >> + return -errno; >> + >> + return 0; >> +} >> + >> +static int gem_reset_status(int fd, int ctx_id) >> +{ >> + int ret; >> + struct local_drm_i915_reset_stats rs; >> + >> + ret = gem_reset_stats(fd, ctx_id, &rs); >> + if (ret) >> + return ret; >> + >> + if (rs.batch_active) >> + return RS_BATCH_ACTIVE; >> + if (rs.batch_pending) >> + return RS_BATCH_PENDING; >> + >> + return RS_NO_ERROR; >> +} >> + >> +static int gem_exec(int fd, struct drm_i915_gem_execbuffer2 *execbuf) >> +{ >> + int ret; >> + >> + ret = ioctl(fd, >> + DRM_IOCTL_I915_GEM_EXECBUFFER2, >> + execbuf); >> + >> + if (ret < 0) >> + return -errno; >> + >> + return 0; >> +} >> + >> +static int exec_valid(int fd, int ctx) >> +{ >> + struct drm_i915_gem_execbuffer2 execbuf; >> + struct drm_i915_gem_exec_object2 exec; >> + int ret; >> + >> + uint32_t buf[2] = { MI_BATCH_BUFFER_END, 0 }; >> + >> + exec.handle = gem_create(fd, 4096); >> + gem_write(fd, exec.handle, 0, buf, sizeof(buf)); >> + exec.relocation_count = 0; >> + exec.relocs_ptr = 0; >> + exec.alignment = 0; >> + exec.offset = 0; >> + exec.flags = 0; >> + exec.rsvd1 = 0; >> + exec.rsvd2 = 0; >> + >> + execbuf.buffers_ptr = (uintptr_t)&exec; >> + execbuf.buffer_count = 1; >> + execbuf.batch_start_offset = 0; >> + execbuf.batch_len = sizeof(buf); >> + execbuf.cliprects_ptr = 0; >> + execbuf.num_cliprects = 0; >> + execbuf.DR1 = 0; >> + execbuf.DR4 = 0; >> + execbuf.flags = 0; >> + i915_execbuffer2_set_context_id(execbuf, ctx); >> + execbuf.rsvd2 = 0; >> + >> + ret = gem_exec(fd, &execbuf); >> + if (ret < 0) >> + return ret; >> + >> + return exec.handle; >> +} >> + >> +#define BUFSIZE (4 * 1024) >> +#define ITEMS (BUFSIZE >> 2) >> + >> +static int inject_hang(int fd, int ctx) >> +{ >> + struct drm_i915_gem_execbuffer2 execbuf; >> + struct drm_i915_gem_exec_object2 exec; >> + uint64_t gtt_off; >> + uint32_t *buf; >> + int roff, i; >> + >> + srandom(time(NULL)); >> + >> + buf = malloc(BUFSIZE); >> + igt_assert(buf != NULL); >> + >> + buf[0] = MI_BATCH_BUFFER_END; >> + buf[1] = MI_NOOP; >> + >> + exec.handle = gem_create(fd, BUFSIZE); >> + gem_write(fd, exec.handle, 0, buf, BUFSIZE); >> + exec.relocation_count = 0; >> + exec.relocs_ptr = 0; >> + exec.alignment = 0; >> + exec.offset = 0; >> + exec.flags = 0; >> + exec.rsvd1 = 0; >> + exec.rsvd2 = 0; >> + >> + execbuf.buffers_ptr = (uintptr_t)&exec; >> + execbuf.buffer_count = 1; >> + execbuf.batch_start_offset = 0; >> + execbuf.batch_len = BUFSIZE; >> + execbuf.cliprects_ptr = 0; >> + execbuf.num_cliprects = 0; >> + execbuf.DR1 = 0; >> + execbuf.DR4 = 0; >> + execbuf.flags = 0; >> + i915_execbuffer2_set_context_id(execbuf, ctx); >> + execbuf.rsvd2 = 0; >> + >> + igt_assert(gem_exec(fd, &execbuf) == 0); >> + >> + gtt_off = exec.offset; >> + >> + for (i = 0; i < ITEMS; i++) >> + buf[i] = MI_NOOP; >> + >> + roff = random() % (ITEMS - 2); >> + buf[roff] = MI_BATCH_BUFFER_START; >> + buf[roff + 1] = gtt_off + (roff << 2); >> + >> +#ifdef VERBOSE >> + printf("loop injected at 0x%lx (off 0x%x, bo_start 0x%lx, bo_end 0x%lx)\n", >> + (long unsigned int)((roff << 2) + gtt_off), >> + roff << 2, (long unsigned int)gtt_off, >> + (long unsigned int)(gtt_off + BUFSIZE - 1)); >> +#endif >> + gem_write(fd, exec.handle, 0, buf, BUFSIZE); >> + >> + exec.relocation_count = 0; >> + exec.relocs_ptr = 0; >> + exec.alignment = 0; >> + exec.offset = 0; >> + exec.flags = 0; >> + exec.rsvd1 = 0; >> + exec.rsvd2 = 0; >> + >> + execbuf.buffers_ptr = (uintptr_t)&exec; >> + execbuf.buffer_count = 1; >> + execbuf.batch_start_offset = 0; >> + execbuf.batch_len = BUFSIZE; >> + execbuf.cliprects_ptr = 0; >> + execbuf.num_cliprects = 0; >> + execbuf.DR1 = 0; >> + execbuf.DR4 = 0; >> + execbuf.flags = 0; >> + i915_execbuffer2_set_context_id(execbuf, ctx); >> + execbuf.rsvd2 = 0; >> + >> + igt_assert(gem_exec(fd, &execbuf) == 0); >> + >> + igt_assert(gtt_off == exec.offset); >> + >> + free(buf); >> + >> + return exec.handle; >> +} >> + >> +static int _assert_reset_status(int fd, int ctx, int status) >> +{ >> + int rs; >> + >> + rs = gem_reset_status(fd, ctx); >> + if (rs < 0) { >> + printf("reset status for %d ctx %d returned %d\n", >> + fd, ctx, rs); >> + return rs; >> + } >> + >> + if (rs != status) { >> + printf("%d:%d reset status %d differs from assumed %d\n", >> + fd, ctx, rs, status); >> + >> + return 1; >> + } >> + >> + return 0; >> +} >> + >> +#define assert_reset_status(fd, ctx, status) \ >> + igt_assert(_assert_reset_status(fd, ctx, status) == 0) >> + >> +static void test_rs(int num_fds, int hang_index, int rs_assumed_no_hang) >> +{ >> + int i; >> + int fd[MAX_FD]; >> + int h[MAX_FD]; >> + >> + igt_assert (num_fds <= MAX_FD); >> + igt_assert (hang_index < MAX_FD); >> + >> + for (i = 0; i < num_fds; i++) { >> + fd[i] = drm_open_any(); >> + igt_assert(fd[i]); >> + } >> + >> + for (i = 0; i < num_fds; i++) >> + assert_reset_status(fd[i], 0, RS_NO_ERROR); >> + >> + for (i = 0; i < num_fds; i++) { >> + if (i == hang_index) >> + h[i] = inject_hang(fd[i], 0); >> + else >> + h[i] = exec_valid(fd[i], 0); >> + } >> + >> + gem_sync(fd[num_fds - 1], h[num_fds - 1]); >> + >> + for (i = 0; i < num_fds; i++) { >> + if (hang_index < 0) { >> + assert_reset_status(fd[i], 0, rs_assumed_no_hang); >> + continue; >> + } >> + >> + if (i < hang_index) >> + assert_reset_status(fd[i], 0, RS_NO_ERROR); >> + if (i == hang_index) >> + assert_reset_status(fd[i], 0, RS_BATCH_ACTIVE); >> + if (i > hang_index) >> + assert_reset_status(fd[i], 0, RS_BATCH_PENDING); >> + } >> + >> + for (i = 0; i < num_fds; i++) { >> + gem_close(fd[i], h[i]); >> + close(fd[i]); >> + } >> +} >> + >> +#define MAX_CTX 100 >> +static void test_rs_ctx(int num_fds, int num_ctx, int hang_index, >> + int hang_context) >> +{ >> + int i, j; >> + int fd[MAX_FD]; >> + int h[MAX_FD][MAX_CTX]; >> + int ctx[MAX_FD][MAX_CTX]; >> + >> + igt_assert (num_fds <= MAX_FD); >> + igt_assert (hang_index < MAX_FD); >> + >> + igt_assert (num_ctx <= MAX_CTX); >> + igt_assert (hang_context < MAX_CTX); >> + >> + test_rs(num_fds, -1, RS_NO_ERROR); >> + >> + for (i = 0; i < num_fds; i++) { >> + fd[i] = drm_open_any(); >> + igt_assert(fd[i]); >> + assert_reset_status(fd[i], 0, RS_NO_ERROR); >> + >> + for (j = 0; j < num_ctx; j++) { >> + ctx[i][j] = context_create(fd[i]); >> + >> + } >> + >> + assert_reset_status(fd[i], 0, RS_NO_ERROR); >> + } >> + >> + for (i = 0; i < num_fds; i++) { >> + >> + assert_reset_status(fd[i], 0, RS_NO_ERROR); >> + >> + for (j = 0; j < num_ctx; j++) >> + assert_reset_status(fd[i], ctx[i][j], RS_NO_ERROR); >> + >> + assert_reset_status(fd[i], 0, RS_NO_ERROR); >> + } >> + >> + for (i = 0; i < num_fds; i++) { >> + for (j = 0; j < num_ctx; j++) { >> + if (i == hang_index && j == hang_context) >> + h[i][j] = inject_hang(fd[i], ctx[i][j]); >> + else >> + h[i][j] = exec_valid(fd[i], ctx[i][j]); >> + } >> + } >> + >> + gem_sync(fd[num_fds - 1], ctx[num_fds - 1][num_ctx - 1]); >> + >> + for (i = 0; i < num_fds; i++) >> + assert_reset_status(fd[i], 0, RS_NO_ERROR); >> + >> + for (i = 0; i < num_fds; i++) { >> + for (j = 0; j < num_ctx; j++) { >> + if (i < hang_index) >> + assert_reset_status(fd[i], ctx[i][j], RS_NO_ERROR); >> + if (i == hang_index && j < hang_context) >> + assert_reset_status(fd[i], ctx[i][j], RS_NO_ERROR); >> + if (i == hang_index && j == hang_context) >> + assert_reset_status(fd[i], ctx[i][j], >> + RS_BATCH_ACTIVE); >> + if (i == hang_index && j > hang_context) >> + assert_reset_status(fd[i], ctx[i][j], >> + RS_BATCH_PENDING); >> + if (i > hang_index) >> + assert_reset_status(fd[i], ctx[i][j], >> + RS_BATCH_PENDING); >> + } >> + } >> + >> + for (i = 0; i < num_fds; i++) { >> + for (j = 0; j < num_ctx; j++) { >> + gem_close(fd[i], h[i][j]); >> + igt_assert(context_destroy(fd[i], ctx[i][j]) == 0); >> + } >> + >> + assert_reset_status(fd[i], 0, RS_NO_ERROR); >> + >> + close(fd[i]); >> + } >> +} >> + >> +static void test_ban(void) >> +{ >> + int h1,h2,h3,h4,h5,h6,h7; >> + int ctx_good, ctx_bad; >> + int fd; >> + int retry = 10; >> + int active_count = 0, pending_count = 0; >> + struct local_drm_i915_reset_stats rs_bad, rs_good; >> + >> + fd = drm_open_any(); >> + igt_assert(fd >= 0); >> + >> + assert_reset_status(fd, 0, RS_NO_ERROR); >> + >> + ctx_good = context_create(fd); >> + ctx_bad = context_create(fd); >> + >> + assert_reset_status(fd, 0, RS_NO_ERROR); >> + assert_reset_status(fd, ctx_good, RS_NO_ERROR); >> + assert_reset_status(fd, ctx_bad, RS_NO_ERROR); >> + >> + h1 = exec_valid(fd, ctx_bad); >> + igt_assert(h1 >= 0); >> + h5 = exec_valid(fd, ctx_good); >> + igt_assert(h5 >= 0); >> + >> + assert_reset_status(fd, ctx_good, RS_NO_ERROR); >> + assert_reset_status(fd, ctx_bad, RS_NO_ERROR); >> + >> + h2 = inject_hang(fd, ctx_bad); >> + igt_assert(h2 >= 0); >> + active_count++; >> + /* Second hang will be pending for this */ >> + pending_count++; >> + >> + h6 = exec_valid(fd, ctx_good); >> + h7 = exec_valid(fd, ctx_good); >> + >> + while (retry--) { >> + h3 = inject_hang(fd, ctx_bad); >> + igt_assert(h3 >= 0); >> + gem_sync(fd, h3); >> + active_count++; >> + /* This second hand will count as pending */ >> + assert_reset_status(fd, ctx_bad, RS_BATCH_ACTIVE); >> + >> + h4 = exec_valid(fd, ctx_bad); >> + if (h4 == -EIO) { >> + gem_close(fd, h3); >> + break; >> + } >> + >> + /* Should not happen often but sometimes hang is declared too slow >> + * due to our way of faking hang using loop */ >> + >> + igt_assert(h4 >= 0); >> + gem_close(fd, h3); >> + gem_close(fd, h4); >> + >> + printf("retrying for ban (%d)\n", retry); >> + } >> + >> + igt_assert(h4 == -EIO); >> + assert_reset_status(fd, ctx_bad, RS_BATCH_ACTIVE); >> + >> + gem_sync(fd, h7); >> + assert_reset_status(fd, ctx_good, RS_BATCH_PENDING); >> + >> + igt_assert(gem_reset_stats(fd, ctx_good, &rs_good) == 0); >> + igt_assert(gem_reset_stats(fd, ctx_bad, &rs_bad) == 0); >> + >> + igt_assert(rs_bad.batch_active == active_count); >> + igt_assert(rs_bad.batch_pending == pending_count); >> + igt_assert(rs_good.batch_active == 0); >> + igt_assert(rs_good.batch_pending == 2); >> + >> + gem_close(fd, h1); >> + gem_close(fd, h2); >> + gem_close(fd, h6); >> + gem_close(fd, h7); >> + >> + h1 = exec_valid(fd, ctx_good); >> + igt_assert(h1 >= 0); >> + gem_close(fd, h1); >> + >> + igt_assert(context_destroy(fd, ctx_good) == 0); >> + igt_assert(context_destroy(fd, ctx_bad) == 0); >> + igt_assert(gem_reset_status(fd, ctx_good) < 0); >> + igt_assert(gem_reset_status(fd, ctx_bad) < 0); >> + igt_assert(exec_valid(fd, ctx_good) < 0); >> + igt_assert(exec_valid(fd, ctx_bad) < 0); >> + >> + close(fd); >> +} >> + >> +static void test_nonrelated_hang(void) >> +{ >> + int h1,h2; >> + int fd1,fd2; >> + int ctx_guilty, ctx_unrelated; >> + >> + fd1 = drm_open_any(); >> + fd2 = drm_open_any(); >> + assert_reset_status(fd1, 0, RS_NO_ERROR); >> + assert_reset_status(fd2, 0, RS_NO_ERROR); >> + ctx_guilty = context_create(fd1); >> + ctx_unrelated = context_create(fd2); >> + >> + assert_reset_status(fd1, ctx_guilty, RS_NO_ERROR); >> + assert_reset_status(fd2, ctx_unrelated, RS_NO_ERROR); >> + >> + h1 = inject_hang(fd1, ctx_guilty); >> + igt_assert(h1 >= 0); >> + gem_sync(fd1, h1); >> + assert_reset_status(fd1, ctx_guilty, RS_BATCH_ACTIVE); >> + assert_reset_status(fd2, ctx_unrelated, RS_NO_ERROR); >> + >> + h2 = exec_valid(fd2, ctx_unrelated); >> + igt_assert(h2 >= 0); >> + gem_sync(fd2, h2); >> + assert_reset_status(fd1, ctx_guilty, RS_BATCH_ACTIVE); >> + assert_reset_status(fd2, ctx_unrelated, RS_NO_ERROR); >> + gem_close(fd1, h1); >> + gem_close(fd2, h2); >> + >> + igt_assert(context_destroy(fd1, ctx_guilty) == 0); >> + igt_assert(context_destroy(fd2, ctx_unrelated) == 0); >> + >> + close(fd1); >> + close(fd2); >> +} >> + >> +static int get_reset_count(int fd, int ctx) >> +{ >> + int ret; >> + struct local_drm_i915_reset_stats rs; >> + >> + ret = gem_reset_stats(fd, ctx, &rs); >> + if (ret) >> + return ret; >> + >> + return rs.reset_count; >> +} >> + >> +static void test_double_destroy_pending(void) >> +{ >> + int fd, h; >> + uint32_t ctx; >> + >> + fd = drm_open_any(); >> + igt_assert(fd >= 0); >> + ctx = context_create(fd); >> + >> + assert_reset_status(fd, ctx, RS_NO_ERROR); >> + >> + h = inject_hang(fd, ctx); >> + igt_assert(h >= 0); >> + igt_assert(context_destroy(fd, ctx) == 0); >> + igt_assert(context_destroy(fd, ctx) == -ENOENT); >> + >> + gem_close(fd, h); >> + close(fd); >> +} >> + >> +static void test_close_pending(void) >> +{ >> + int fd, h; >> + >> + fd = drm_open_any(); >> + igt_assert(fd >= 0); >> + >> + assert_reset_status(fd, 0, RS_NO_ERROR); >> + >> + h = inject_hang(fd, 0); >> + igt_assert(h >= 0); >> + >> + gem_close(fd, h); >> + close(fd); >> +} >> + >> +static void __test_count(const bool create_ctx) >> +{ >> + int fd, h, ctx; >> + long c1, c2; >> + >> + fd = drm_open_any(); >> + igt_assert(fd >= 0); >> + if (create_ctx) >> + ctx = context_create(fd); >> + else >> + ctx = 0; >> + >> + assert_reset_status(fd, ctx, RS_NO_ERROR); >> + >> + c1 = get_reset_count(fd, ctx); >> + igt_assert(c1 >= 0); >> + >> + h = inject_hang(fd, ctx); >> + igt_assert (h >= 0); >> + gem_sync(fd, h); >> + >> + assert_reset_status(fd, ctx, RS_BATCH_ACTIVE); >> + c2 = get_reset_count(fd, ctx); >> + igt_assert(c2 >= 0); >> + >> + igt_assert(c2 == (c1 + 1)); >> + >> + gem_close(fd, h); >> + >> + if (create_ctx) >> + context_destroy(fd, ctx); >> + >> + close(fd); >> +} >> + >> +static void test_count(void) >> +{ >> + return __test_count(false); >> +} >> + >> +static void test_count_context(void) >> +{ >> + return __test_count(true); >> +} >> + >> +static void test_global_reset_count(void) >> +{ >> + test_count(); >> + test_count_context(); >> +} >> + >> +static int _test_params(int fd, int ctx, uint32_t flags, uint32_t pad) >> +{ >> + struct local_drm_i915_reset_stats rs; >> + int ret; >> + >> + rs.ctx_id = ctx; >> + rs.flags = flags; >> + rs.reset_count = rand(); >> + rs.batch_active = rand(); >> + rs.batch_pending = rand(); >> + rs.pad = pad; >> + >> + do { >> + ret = ioctl(fd, GET_RESET_STATS_IOCTL, &rs); >> + } while (ret == -1 && (errno == EINTR || errno == EAGAIN)); >> + >> + if (ret < 0) >> + return -errno; >> + >> + return 0; >> +} >> + >> +static void test_param_ctx(int fd, int ctx) >> +{ >> + const uint32_t bad = rand() + 1; >> + >> + igt_assert(_test_params(fd, ctx, 0, 0) == 0); >> + igt_assert(_test_params(fd, ctx, 0, bad) == -EINVAL); >> + igt_assert(_test_params(fd, ctx, bad, 0) == -EINVAL); >> + igt_assert(_test_params(fd, ctx, bad, bad) == -EINVAL); >> +} >> + >> +static void test_params(void) >> +{ >> + int fd, ctx; >> + >> + fd = drm_open_any(); >> + igt_assert(fd >= 0); >> + ctx = context_create(fd); >> + >> + igt_assert(ioctl(fd, GET_RESET_STATS_IOCTL, 0) == -1); >> + >> + igt_assert(_test_params(fd, 0xbadbad, 0, 0) == -ENOENT); >> + >> + test_param_ctx(fd, 0); >> + test_param_ctx(fd, ctx); >> + >> + close(fd); >> +} >> + >> + >> +int main(int argc, char **argv) >> +{ >> + struct local_drm_i915_gem_context_create create; >> + uint32_t devid; >> + int fd; >> + int ret; >> + >> + igt_skip_on_simulation(); >> + >> + igt_subtest_init(argc, argv); >> + >> + igt_fixture { >> + fd = drm_open_any(); >> + devid = intel_get_drm_devid(fd); >> + if (intel_gen(devid) < 4) >> + igt_skip("Architecture %d too old\n", intel_gen(devid)); >> + >> + ret = drmIoctl(fd, CONTEXT_CREATE_IOCTL, &create); >> + if (ret != 0 && (errno == ENODEV || errno == EINVAL)) >> + igt_skip("Kernel is too old, or contexts not supported: %s\n", >> + strerror(errno)); >> + >> + close(fd); >> + } >> + >> + igt_subtest("basic-reset-status") >> + test_rs(4, 1, 0); >> + >> + igt_subtest("context-reset-status") >> + test_rs_ctx(4, 4, 1, 2); >> + >> + igt_subtest("ban") >> + test_ban(); >> + >> + igt_subtest("ctx-unrelated") >> + test_nonrelated_hang(); >> + >> + igt_subtest("global-count") >> + test_global_reset_count(); >> + >> + igt_subtest("double-destroy-pending") >> + test_double_destroy_pending(); >> + >> + igt_subtest("close-pending") >> + test_close_pending(); >> + >> + igt_subtest("params") >> + test_params(); >> + >> + igt_exit(); >> +} >> -- >> 1.7.9.5 >> >> _______________________________________________ >> Intel-gfx mailing list >> Intel-gfx@lists.freedesktop.org >> http://lists.freedesktop.org/mailman/listinfo/intel-gfx > > -- > Daniel Vetter > Software Engineer, Intel Corporation > +41 (0) 79 365 57 48 - http://blog.ffwll.ch
On Tue, Nov 12, 2013 at 07:58:16PM +0200, Mika Kuoppala wrote: > + buf[roff] = MI_BATCH_BUFFER_START; > + buf[roff + 1] = gtt_off + (roff << 2); I was thinking that this probably need to be adapted for BDW if we want to loop at the right address (MI_BATCH_BUFFER_START has an extra dword, and I believe one needs to set the right command length. The high bits should be 0 as the batch is filled with nops, but it's still unclear to me if the correct lengh is needed).
On Wed, Nov 13, 2013 at 3:13 PM, Mika Kuoppala <mika.kuoppala@linux.intel.com> wrote: > Daniel Vetter <daniel@ffwll.ch> writes: > >> On Tue, Nov 12, 2013 at 07:58:16PM +0200, Mika Kuoppala wrote: >>> v2: check the ioctl pad and flag parameters >>> >>> Signed-off-by: Mika Kuoppala <mika.kuoppala@intel.com> >> >> I've merged this to igt, but there are a few small fixups to do on top: >> - We now have the igt_main macro to cut down a bit on boilerplate. >> - I haven't tested it, but I guess inject_hang will cause some *ERROR* >> noise in dmesg. Rule is that igt testcases should only cause info/debug >> level messages, everything else is considered a failure. I think we >> could fix this by setting the stop_rings debugfs value right _after_ the >> hang is injected, to tell the kernel that the hang it'll see is actually >> fake. > > I will take a look what stop_rings do. I just have a feeling > that with this kind of trickery we shoot ourselves in foot some day. Yeah, it's a bit trickery. But as long as we only have on special testcase I think it's ok. If we grow more injected hang tests which aren't simulated with the stop_ring stuff then maybe a new debugfs variable to tell the kernel to expect a fake hang could be useful. But smells like not worth it right now. > Would be good that we get a proper error messages when we submit a real > hanging batches. Would be also good that our test checks that > the ERROR msg was really emitted. I think the check for the reset count is good enough to make sure the hangcheck code works. If we start to check for specific dmesg lines I fear we'll implicitly make them abi. We have uevents as a general signal to userspace that a hang happened (and now the reset_stat ioctl). > Having whitelist of expected '*ERROR*' messages for these kind of tests > is not an option? Thus far we've just tuned down the message to info level for fake hangs. It does make things a bit easier for the test runner since we can just filter for any dmesg noise with a level >= warn and then fail the test. This is what QA does, and with the latest piglit patches I've just pushed also what the piglit testrunner does. >> - The userspace interface checking has two missing spots: a) checking that >> lookup for an invalid ctx id fails with ENOENT b) checking that non-root >> can't read out the default context. For the later it's probably simples >> to fork a 2nd process and drop the CAP_SYS_ADMIN capability in there. > > I check for ENOENT on submitting bad context id there. Did you miss it > or it is not enough? Oops, missed that. I didn't reread the latest version too carefully :( > I will resubmit when I have b) in place. Thanks. Btw I've already pushed your current patch to igt, so just a follow-up patch is required. -Daniel
diff --git a/tests/Makefile.sources b/tests/Makefile.sources index 5f4d282..a02b93d 100644 --- a/tests/Makefile.sources +++ b/tests/Makefile.sources @@ -182,6 +182,7 @@ HANG = \ gem_bad_blit \ gem_bad_address \ gem_non_secure_batch \ + gem_reset_stats \ $(NULL) scripts = \ diff --git a/tests/gem_reset_stats.c b/tests/gem_reset_stats.c new file mode 100644 index 0000000..2040294 --- /dev/null +++ b/tests/gem_reset_stats.c @@ -0,0 +1,770 @@ +/* + * Copyright (c) 2013 Intel Corporation + * + * Permission is hereby granted, free of charge, to any person obtaining a + * copy of this software and associated documentation files (the "Software"), + * to deal in the Software without restriction, including without limitation + * the rights to use, copy, modify, merge, publish, distribute, sublicense, + * and/or sell copies of the Software, and to permit persons to whom the + * Software is furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice (including the next + * paragraph) shall be included in all copies or substantial portions of the + * Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS + * IN THE SOFTWARE. + * + * Authors: + * Mika Kuoppala <mika.kuoppala@intel.com> + * + */ + +#include <unistd.h> +#include <stdlib.h> +#include <stdio.h> +#include <string.h> +#include <fcntl.h> +#include <inttypes.h> +#include <errno.h> +#include <sys/stat.h> +#include <sys/ioctl.h> +#include <sys/mman.h> +#include <time.h> + +#include "i915_drm.h" +#include "intel_bufmgr.h" +#include "intel_batchbuffer.h" +#include "intel_gpu_tools.h" +#include "rendercopy.h" + +#define RS_NO_ERROR 0 +#define RS_BATCH_ACTIVE (1 << 0) +#define RS_BATCH_PENDING (1 << 1) +#define RS_UNKNOWN (1 << 2) + +struct local_drm_i915_reset_stats { + __u32 ctx_id; + __u32 flags; + __u32 reset_count; + __u32 batch_active; + __u32 batch_pending; + __u32 pad; +}; + +struct local_drm_i915_gem_context_create { + __u32 ctx_id; + __u32 pad; +}; + +struct local_drm_i915_gem_context_destroy { + __u32 ctx_id; + __u32 pad; +}; + +#define MAX_FD 32 + +#define CONTEXT_CREATE_IOCTL DRM_IOWR(DRM_COMMAND_BASE + 0x2d, struct local_drm_i915_gem_context_create) +#define CONTEXT_DESTROY_IOCTL DRM_IOWR(DRM_COMMAND_BASE + 0x2e, struct local_drm_i915_gem_context_destroy) +#define GET_RESET_STATS_IOCTL DRM_IOWR(DRM_COMMAND_BASE + 0x32, struct local_drm_i915_reset_stats) + +static uint32_t context_create(int fd) +{ + struct local_drm_i915_gem_context_create create; + int ret; + + create.ctx_id = rand(); + create.pad = rand(); + + ret = drmIoctl(fd, CONTEXT_CREATE_IOCTL, &create); + igt_assert(ret == 0); + + return create.ctx_id; +} + +static int context_destroy(int fd, uint32_t ctx_id) +{ + int ret; + struct local_drm_i915_gem_context_destroy destroy; + + destroy.ctx_id = ctx_id; + destroy.pad = rand(); + + ret = drmIoctl(fd, CONTEXT_DESTROY_IOCTL, &destroy); + if (ret != 0) + return -errno; + + return 0; +} + +static int gem_reset_stats(int fd, int ctx_id, + struct local_drm_i915_reset_stats *rs) +{ + int ret; + + rs->ctx_id = ctx_id; + rs->flags = 0; + rs->reset_count = rand(); + rs->batch_active = rand(); + rs->batch_pending = rand(); + rs->pad = 0; + + do { + ret = ioctl(fd, GET_RESET_STATS_IOCTL, rs); + } while (ret == -1 && (errno == EINTR || errno == EAGAIN)); + + if (ret < 0) + return -errno; + + return 0; +} + +static int gem_reset_status(int fd, int ctx_id) +{ + int ret; + struct local_drm_i915_reset_stats rs; + + ret = gem_reset_stats(fd, ctx_id, &rs); + if (ret) + return ret; + + if (rs.batch_active) + return RS_BATCH_ACTIVE; + if (rs.batch_pending) + return RS_BATCH_PENDING; + + return RS_NO_ERROR; +} + +static int gem_exec(int fd, struct drm_i915_gem_execbuffer2 *execbuf) +{ + int ret; + + ret = ioctl(fd, + DRM_IOCTL_I915_GEM_EXECBUFFER2, + execbuf); + + if (ret < 0) + return -errno; + + return 0; +} + +static int exec_valid(int fd, int ctx) +{ + struct drm_i915_gem_execbuffer2 execbuf; + struct drm_i915_gem_exec_object2 exec; + int ret; + + uint32_t buf[2] = { MI_BATCH_BUFFER_END, 0 }; + + exec.handle = gem_create(fd, 4096); + gem_write(fd, exec.handle, 0, buf, sizeof(buf)); + exec.relocation_count = 0; + exec.relocs_ptr = 0; + exec.alignment = 0; + exec.offset = 0; + exec.flags = 0; + exec.rsvd1 = 0; + exec.rsvd2 = 0; + + execbuf.buffers_ptr = (uintptr_t)&exec; + execbuf.buffer_count = 1; + execbuf.batch_start_offset = 0; + execbuf.batch_len = sizeof(buf); + execbuf.cliprects_ptr = 0; + execbuf.num_cliprects = 0; + execbuf.DR1 = 0; + execbuf.DR4 = 0; + execbuf.flags = 0; + i915_execbuffer2_set_context_id(execbuf, ctx); + execbuf.rsvd2 = 0; + + ret = gem_exec(fd, &execbuf); + if (ret < 0) + return ret; + + return exec.handle; +} + +#define BUFSIZE (4 * 1024) +#define ITEMS (BUFSIZE >> 2) + +static int inject_hang(int fd, int ctx) +{ + struct drm_i915_gem_execbuffer2 execbuf; + struct drm_i915_gem_exec_object2 exec; + uint64_t gtt_off; + uint32_t *buf; + int roff, i; + + srandom(time(NULL)); + + buf = malloc(BUFSIZE); + igt_assert(buf != NULL); + + buf[0] = MI_BATCH_BUFFER_END; + buf[1] = MI_NOOP; + + exec.handle = gem_create(fd, BUFSIZE); + gem_write(fd, exec.handle, 0, buf, BUFSIZE); + exec.relocation_count = 0; + exec.relocs_ptr = 0; + exec.alignment = 0; + exec.offset = 0; + exec.flags = 0; + exec.rsvd1 = 0; + exec.rsvd2 = 0; + + execbuf.buffers_ptr = (uintptr_t)&exec; + execbuf.buffer_count = 1; + execbuf.batch_start_offset = 0; + execbuf.batch_len = BUFSIZE; + execbuf.cliprects_ptr = 0; + execbuf.num_cliprects = 0; + execbuf.DR1 = 0; + execbuf.DR4 = 0; + execbuf.flags = 0; + i915_execbuffer2_set_context_id(execbuf, ctx); + execbuf.rsvd2 = 0; + + igt_assert(gem_exec(fd, &execbuf) == 0); + + gtt_off = exec.offset; + + for (i = 0; i < ITEMS; i++) + buf[i] = MI_NOOP; + + roff = random() % (ITEMS - 2); + buf[roff] = MI_BATCH_BUFFER_START; + buf[roff + 1] = gtt_off + (roff << 2); + +#ifdef VERBOSE + printf("loop injected at 0x%lx (off 0x%x, bo_start 0x%lx, bo_end 0x%lx)\n", + (long unsigned int)((roff << 2) + gtt_off), + roff << 2, (long unsigned int)gtt_off, + (long unsigned int)(gtt_off + BUFSIZE - 1)); +#endif + gem_write(fd, exec.handle, 0, buf, BUFSIZE); + + exec.relocation_count = 0; + exec.relocs_ptr = 0; + exec.alignment = 0; + exec.offset = 0; + exec.flags = 0; + exec.rsvd1 = 0; + exec.rsvd2 = 0; + + execbuf.buffers_ptr = (uintptr_t)&exec; + execbuf.buffer_count = 1; + execbuf.batch_start_offset = 0; + execbuf.batch_len = BUFSIZE; + execbuf.cliprects_ptr = 0; + execbuf.num_cliprects = 0; + execbuf.DR1 = 0; + execbuf.DR4 = 0; + execbuf.flags = 0; + i915_execbuffer2_set_context_id(execbuf, ctx); + execbuf.rsvd2 = 0; + + igt_assert(gem_exec(fd, &execbuf) == 0); + + igt_assert(gtt_off == exec.offset); + + free(buf); + + return exec.handle; +} + +static int _assert_reset_status(int fd, int ctx, int status) +{ + int rs; + + rs = gem_reset_status(fd, ctx); + if (rs < 0) { + printf("reset status for %d ctx %d returned %d\n", + fd, ctx, rs); + return rs; + } + + if (rs != status) { + printf("%d:%d reset status %d differs from assumed %d\n", + fd, ctx, rs, status); + + return 1; + } + + return 0; +} + +#define assert_reset_status(fd, ctx, status) \ + igt_assert(_assert_reset_status(fd, ctx, status) == 0) + +static void test_rs(int num_fds, int hang_index, int rs_assumed_no_hang) +{ + int i; + int fd[MAX_FD]; + int h[MAX_FD]; + + igt_assert (num_fds <= MAX_FD); + igt_assert (hang_index < MAX_FD); + + for (i = 0; i < num_fds; i++) { + fd[i] = drm_open_any(); + igt_assert(fd[i]); + } + + for (i = 0; i < num_fds; i++) + assert_reset_status(fd[i], 0, RS_NO_ERROR); + + for (i = 0; i < num_fds; i++) { + if (i == hang_index) + h[i] = inject_hang(fd[i], 0); + else + h[i] = exec_valid(fd[i], 0); + } + + gem_sync(fd[num_fds - 1], h[num_fds - 1]); + + for (i = 0; i < num_fds; i++) { + if (hang_index < 0) { + assert_reset_status(fd[i], 0, rs_assumed_no_hang); + continue; + } + + if (i < hang_index) + assert_reset_status(fd[i], 0, RS_NO_ERROR); + if (i == hang_index) + assert_reset_status(fd[i], 0, RS_BATCH_ACTIVE); + if (i > hang_index) + assert_reset_status(fd[i], 0, RS_BATCH_PENDING); + } + + for (i = 0; i < num_fds; i++) { + gem_close(fd[i], h[i]); + close(fd[i]); + } +} + +#define MAX_CTX 100 +static void test_rs_ctx(int num_fds, int num_ctx, int hang_index, + int hang_context) +{ + int i, j; + int fd[MAX_FD]; + int h[MAX_FD][MAX_CTX]; + int ctx[MAX_FD][MAX_CTX]; + + igt_assert (num_fds <= MAX_FD); + igt_assert (hang_index < MAX_FD); + + igt_assert (num_ctx <= MAX_CTX); + igt_assert (hang_context < MAX_CTX); + + test_rs(num_fds, -1, RS_NO_ERROR); + + for (i = 0; i < num_fds; i++) { + fd[i] = drm_open_any(); + igt_assert(fd[i]); + assert_reset_status(fd[i], 0, RS_NO_ERROR); + + for (j = 0; j < num_ctx; j++) { + ctx[i][j] = context_create(fd[i]); + + } + + assert_reset_status(fd[i], 0, RS_NO_ERROR); + } + + for (i = 0; i < num_fds; i++) { + + assert_reset_status(fd[i], 0, RS_NO_ERROR); + + for (j = 0; j < num_ctx; j++) + assert_reset_status(fd[i], ctx[i][j], RS_NO_ERROR); + + assert_reset_status(fd[i], 0, RS_NO_ERROR); + } + + for (i = 0; i < num_fds; i++) { + for (j = 0; j < num_ctx; j++) { + if (i == hang_index && j == hang_context) + h[i][j] = inject_hang(fd[i], ctx[i][j]); + else + h[i][j] = exec_valid(fd[i], ctx[i][j]); + } + } + + gem_sync(fd[num_fds - 1], ctx[num_fds - 1][num_ctx - 1]); + + for (i = 0; i < num_fds; i++) + assert_reset_status(fd[i], 0, RS_NO_ERROR); + + for (i = 0; i < num_fds; i++) { + for (j = 0; j < num_ctx; j++) { + if (i < hang_index) + assert_reset_status(fd[i], ctx[i][j], RS_NO_ERROR); + if (i == hang_index && j < hang_context) + assert_reset_status(fd[i], ctx[i][j], RS_NO_ERROR); + if (i == hang_index && j == hang_context) + assert_reset_status(fd[i], ctx[i][j], + RS_BATCH_ACTIVE); + if (i == hang_index && j > hang_context) + assert_reset_status(fd[i], ctx[i][j], + RS_BATCH_PENDING); + if (i > hang_index) + assert_reset_status(fd[i], ctx[i][j], + RS_BATCH_PENDING); + } + } + + for (i = 0; i < num_fds; i++) { + for (j = 0; j < num_ctx; j++) { + gem_close(fd[i], h[i][j]); + igt_assert(context_destroy(fd[i], ctx[i][j]) == 0); + } + + assert_reset_status(fd[i], 0, RS_NO_ERROR); + + close(fd[i]); + } +} + +static void test_ban(void) +{ + int h1,h2,h3,h4,h5,h6,h7; + int ctx_good, ctx_bad; + int fd; + int retry = 10; + int active_count = 0, pending_count = 0; + struct local_drm_i915_reset_stats rs_bad, rs_good; + + fd = drm_open_any(); + igt_assert(fd >= 0); + + assert_reset_status(fd, 0, RS_NO_ERROR); + + ctx_good = context_create(fd); + ctx_bad = context_create(fd); + + assert_reset_status(fd, 0, RS_NO_ERROR); + assert_reset_status(fd, ctx_good, RS_NO_ERROR); + assert_reset_status(fd, ctx_bad, RS_NO_ERROR); + + h1 = exec_valid(fd, ctx_bad); + igt_assert(h1 >= 0); + h5 = exec_valid(fd, ctx_good); + igt_assert(h5 >= 0); + + assert_reset_status(fd, ctx_good, RS_NO_ERROR); + assert_reset_status(fd, ctx_bad, RS_NO_ERROR); + + h2 = inject_hang(fd, ctx_bad); + igt_assert(h2 >= 0); + active_count++; + /* Second hang will be pending for this */ + pending_count++; + + h6 = exec_valid(fd, ctx_good); + h7 = exec_valid(fd, ctx_good); + + while (retry--) { + h3 = inject_hang(fd, ctx_bad); + igt_assert(h3 >= 0); + gem_sync(fd, h3); + active_count++; + /* This second hand will count as pending */ + assert_reset_status(fd, ctx_bad, RS_BATCH_ACTIVE); + + h4 = exec_valid(fd, ctx_bad); + if (h4 == -EIO) { + gem_close(fd, h3); + break; + } + + /* Should not happen often but sometimes hang is declared too slow + * due to our way of faking hang using loop */ + + igt_assert(h4 >= 0); + gem_close(fd, h3); + gem_close(fd, h4); + + printf("retrying for ban (%d)\n", retry); + } + + igt_assert(h4 == -EIO); + assert_reset_status(fd, ctx_bad, RS_BATCH_ACTIVE); + + gem_sync(fd, h7); + assert_reset_status(fd, ctx_good, RS_BATCH_PENDING); + + igt_assert(gem_reset_stats(fd, ctx_good, &rs_good) == 0); + igt_assert(gem_reset_stats(fd, ctx_bad, &rs_bad) == 0); + + igt_assert(rs_bad.batch_active == active_count); + igt_assert(rs_bad.batch_pending == pending_count); + igt_assert(rs_good.batch_active == 0); + igt_assert(rs_good.batch_pending == 2); + + gem_close(fd, h1); + gem_close(fd, h2); + gem_close(fd, h6); + gem_close(fd, h7); + + h1 = exec_valid(fd, ctx_good); + igt_assert(h1 >= 0); + gem_close(fd, h1); + + igt_assert(context_destroy(fd, ctx_good) == 0); + igt_assert(context_destroy(fd, ctx_bad) == 0); + igt_assert(gem_reset_status(fd, ctx_good) < 0); + igt_assert(gem_reset_status(fd, ctx_bad) < 0); + igt_assert(exec_valid(fd, ctx_good) < 0); + igt_assert(exec_valid(fd, ctx_bad) < 0); + + close(fd); +} + +static void test_nonrelated_hang(void) +{ + int h1,h2; + int fd1,fd2; + int ctx_guilty, ctx_unrelated; + + fd1 = drm_open_any(); + fd2 = drm_open_any(); + assert_reset_status(fd1, 0, RS_NO_ERROR); + assert_reset_status(fd2, 0, RS_NO_ERROR); + ctx_guilty = context_create(fd1); + ctx_unrelated = context_create(fd2); + + assert_reset_status(fd1, ctx_guilty, RS_NO_ERROR); + assert_reset_status(fd2, ctx_unrelated, RS_NO_ERROR); + + h1 = inject_hang(fd1, ctx_guilty); + igt_assert(h1 >= 0); + gem_sync(fd1, h1); + assert_reset_status(fd1, ctx_guilty, RS_BATCH_ACTIVE); + assert_reset_status(fd2, ctx_unrelated, RS_NO_ERROR); + + h2 = exec_valid(fd2, ctx_unrelated); + igt_assert(h2 >= 0); + gem_sync(fd2, h2); + assert_reset_status(fd1, ctx_guilty, RS_BATCH_ACTIVE); + assert_reset_status(fd2, ctx_unrelated, RS_NO_ERROR); + gem_close(fd1, h1); + gem_close(fd2, h2); + + igt_assert(context_destroy(fd1, ctx_guilty) == 0); + igt_assert(context_destroy(fd2, ctx_unrelated) == 0); + + close(fd1); + close(fd2); +} + +static int get_reset_count(int fd, int ctx) +{ + int ret; + struct local_drm_i915_reset_stats rs; + + ret = gem_reset_stats(fd, ctx, &rs); + if (ret) + return ret; + + return rs.reset_count; +} + +static void test_double_destroy_pending(void) +{ + int fd, h; + uint32_t ctx; + + fd = drm_open_any(); + igt_assert(fd >= 0); + ctx = context_create(fd); + + assert_reset_status(fd, ctx, RS_NO_ERROR); + + h = inject_hang(fd, ctx); + igt_assert(h >= 0); + igt_assert(context_destroy(fd, ctx) == 0); + igt_assert(context_destroy(fd, ctx) == -ENOENT); + + gem_close(fd, h); + close(fd); +} + +static void test_close_pending(void) +{ + int fd, h; + + fd = drm_open_any(); + igt_assert(fd >= 0); + + assert_reset_status(fd, 0, RS_NO_ERROR); + + h = inject_hang(fd, 0); + igt_assert(h >= 0); + + gem_close(fd, h); + close(fd); +} + +static void __test_count(const bool create_ctx) +{ + int fd, h, ctx; + long c1, c2; + + fd = drm_open_any(); + igt_assert(fd >= 0); + if (create_ctx) + ctx = context_create(fd); + else + ctx = 0; + + assert_reset_status(fd, ctx, RS_NO_ERROR); + + c1 = get_reset_count(fd, ctx); + igt_assert(c1 >= 0); + + h = inject_hang(fd, ctx); + igt_assert (h >= 0); + gem_sync(fd, h); + + assert_reset_status(fd, ctx, RS_BATCH_ACTIVE); + c2 = get_reset_count(fd, ctx); + igt_assert(c2 >= 0); + + igt_assert(c2 == (c1 + 1)); + + gem_close(fd, h); + + if (create_ctx) + context_destroy(fd, ctx); + + close(fd); +} + +static void test_count(void) +{ + return __test_count(false); +} + +static void test_count_context(void) +{ + return __test_count(true); +} + +static void test_global_reset_count(void) +{ + test_count(); + test_count_context(); +} + +static int _test_params(int fd, int ctx, uint32_t flags, uint32_t pad) +{ + struct local_drm_i915_reset_stats rs; + int ret; + + rs.ctx_id = ctx; + rs.flags = flags; + rs.reset_count = rand(); + rs.batch_active = rand(); + rs.batch_pending = rand(); + rs.pad = pad; + + do { + ret = ioctl(fd, GET_RESET_STATS_IOCTL, &rs); + } while (ret == -1 && (errno == EINTR || errno == EAGAIN)); + + if (ret < 0) + return -errno; + + return 0; +} + +static void test_param_ctx(int fd, int ctx) +{ + const uint32_t bad = rand() + 1; + + igt_assert(_test_params(fd, ctx, 0, 0) == 0); + igt_assert(_test_params(fd, ctx, 0, bad) == -EINVAL); + igt_assert(_test_params(fd, ctx, bad, 0) == -EINVAL); + igt_assert(_test_params(fd, ctx, bad, bad) == -EINVAL); +} + +static void test_params(void) +{ + int fd, ctx; + + fd = drm_open_any(); + igt_assert(fd >= 0); + ctx = context_create(fd); + + igt_assert(ioctl(fd, GET_RESET_STATS_IOCTL, 0) == -1); + + igt_assert(_test_params(fd, 0xbadbad, 0, 0) == -ENOENT); + + test_param_ctx(fd, 0); + test_param_ctx(fd, ctx); + + close(fd); +} + + +int main(int argc, char **argv) +{ + struct local_drm_i915_gem_context_create create; + uint32_t devid; + int fd; + int ret; + + igt_skip_on_simulation(); + + igt_subtest_init(argc, argv); + + igt_fixture { + fd = drm_open_any(); + devid = intel_get_drm_devid(fd); + if (intel_gen(devid) < 4) + igt_skip("Architecture %d too old\n", intel_gen(devid)); + + ret = drmIoctl(fd, CONTEXT_CREATE_IOCTL, &create); + if (ret != 0 && (errno == ENODEV || errno == EINVAL)) + igt_skip("Kernel is too old, or contexts not supported: %s\n", + strerror(errno)); + + close(fd); + } + + igt_subtest("basic-reset-status") + test_rs(4, 1, 0); + + igt_subtest("context-reset-status") + test_rs_ctx(4, 4, 1, 2); + + igt_subtest("ban") + test_ban(); + + igt_subtest("ctx-unrelated") + test_nonrelated_hang(); + + igt_subtest("global-count") + test_global_reset_count(); + + igt_subtest("double-destroy-pending") + test_double_destroy_pending(); + + igt_subtest("close-pending") + test_close_pending(); + + igt_subtest("params") + test_params(); + + igt_exit(); +}
v2: check the ioctl pad and flag parameters Signed-off-by: Mika Kuoppala <mika.kuoppala@intel.com> --- tests/Makefile.sources | 1 + tests/gem_reset_stats.c | 770 +++++++++++++++++++++++++++++++++++++++++++++++ 2 files changed, 771 insertions(+) create mode 100644 tests/gem_reset_stats.c