/* * Copyright © 2009 Intel Corporation * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. * * Authors: * Eric Anholt * */ /** @file gem_ringfill.c * * This is a test of doing many tiny batchbuffer operations, in the hope of * catching failure to manage the ring properly near full. */ #include #include #include #include #include #include #include #include #include #include #include "ioctl_wrappers.h" #include "drmtest.h" #include "intel_chipset.h" #include "intel_io.h" #include "igt_aux.h" struct bo { const char *ring; drm_intel_bo *src, *dst, *tmp; }; static const int width = 512, height = 512; int fd; static void create_bo(drm_intel_bufmgr *bufmgr, struct bo *b, const char *ring) { int size = 4 * width * height, i; uint32_t *map; b->ring = ring; b->src = drm_intel_bo_alloc(bufmgr, "src", size, 4096); b->dst = drm_intel_bo_alloc(bufmgr, "dst", size, 4096); b->tmp = drm_intel_bo_alloc(bufmgr, "tmp", size, 4096); /* Fill the src with indexes of the pixels */ drm_intel_bo_map(b->src, true); map = b->src->virtual; for (i = 0; i < width * height; i++) map[i] = i; drm_intel_bo_unmap(b->src); /* Fill the dst with garbage. */ drm_intel_bo_map(b->dst, true); map = b->dst->virtual; for (i = 0; i < width * height; i++) map[i] = 0xd0d0d0d0; drm_intel_bo_unmap(b->dst); } static int check_bo(struct bo *b) { const uint32_t *map; int i, fails = 0; igt_debug("verifying\n"); do_or_die(drm_intel_bo_map(b->dst, false)); map = b->dst->virtual; for (i = 0; i < width*height; i++) { if (map[i] != i && ++fails <= 9) { int x = i % width; int y = i / width; igt_info("%s: copy #%d at %d,%d failed: read 0x%08x\n", b->ring, i, x, y, map[i]); } } drm_intel_bo_unmap(b->dst); return fails; } static void destroy_bo(struct bo *b) { drm_intel_bo_unreference(b->src); drm_intel_bo_unreference(b->tmp); drm_intel_bo_unreference(b->dst); } static void fill_ring(drm_intel_bufmgr *bufmgr, const char *ring, igt_render_copyfunc_t copy) { struct intel_batchbuffer *batch; struct igt_buf src, tmp, dst; struct bo bo; int i; batch = intel_batchbuffer_alloc(bufmgr, intel_get_drm_devid(fd)); igt_assert(batch); create_bo(bufmgr, &bo, ring); src.stride = 4 * width; src.tiling = 0; src.size = 4 * width * height; src.num_tiles = 4 * width * height; dst = tmp = src; src.bo = bo.src; tmp.bo = bo.tmp; dst.bo = bo.dst; /* The ring we've been using is 128k, and each rendering op * will use at least 8 dwords: * * BATCH_START * BATCH_START offset * MI_FLUSH * STORE_DATA_INDEX * STORE_DATA_INDEX offset * STORE_DATA_INDEX value * MI_USER_INTERRUPT * (padding) * * So iterate just a little more than that -- if we don't fill the ring * doing this, we aren't likely to with this test. */ for (i = 0; i < width * height; i++) { int x = i % width; int y = i / width; igt_assert_lt(y, height); /* Dummy load to fill the ring */ copy(batch, NULL, &src, 0, 0, width, height, &tmp, 0, 0); /* And copy the src into dst, pixel by pixel */ copy(batch, NULL, &src, x, y, 1, 1, &dst, x, y); } /* verify */ igt_assert_eq(check_bo(&bo), 0); destroy_bo(&bo); intel_batchbuffer_free(batch); } static void blt_copy(struct intel_batchbuffer *batch, drm_intel_context *context, struct igt_buf *src, unsigned src_x, unsigned src_y, unsigned w, unsigned h, struct igt_buf *dst, unsigned dst_x, unsigned dst_y) { BLIT_COPY_BATCH_START(0); OUT_BATCH((3 << 24) | /* 32 bits */ (0xcc << 16) | /* copy ROP */ dst->stride); OUT_BATCH((dst_y << 16) | dst_x); /* dst x1,y1 */ OUT_BATCH(((dst_y + h) << 16) | (dst_x + w)); /* dst x2,y2 */ OUT_RELOC_FENCED(dst->bo, I915_GEM_DOMAIN_RENDER, I915_GEM_DOMAIN_RENDER, 0); OUT_BATCH((src_y << 16) | src_x); /* src x1,y1 */ OUT_BATCH(src->stride); OUT_RELOC_FENCED(src->bo, I915_GEM_DOMAIN_RENDER, 0, 0); ADVANCE_BATCH(); intel_batchbuffer_flush(batch); } static void run_test(int ring, bool interruptible, int nchild) { drm_intel_bufmgr *bufmgr; igt_render_copyfunc_t copy; const char* ring_name; bufmgr = drm_intel_bufmgr_gem_init(fd, 4096); igt_assert(bufmgr); drm_intel_bufmgr_gem_enable_reuse(bufmgr); if (ring == I915_EXEC_RENDER) { copy = igt_get_render_copyfunc(intel_get_drm_devid(fd)); ring_name = "render"; } else if (ring == I915_EXEC_BLT) { copy = blt_copy; ring_name = "blt"; } else { igt_fail_on_f(true, "Unsupported ring."); } /* Not all platforms have dedicated render ring. */ igt_require(copy); if (interruptible) { igt_fork_signal_helper(); } if (nchild) { igt_fork(child, nchild) { fill_ring(bufmgr, ring_name, copy); } igt_waitchildren(); } else { fill_ring(bufmgr, ring_name, copy); } if (interruptible) { igt_stop_signal_helper(); } drm_intel_bufmgr_destroy(bufmgr); } igt_main { igt_skip_on_simulation(); igt_fixture { fd = drm_open_any(); } igt_subtest("blitter") run_test(I915_EXEC_BLT, false, 0); igt_subtest("render") run_test(I915_EXEC_RENDER, false, 0); igt_subtest("blitter-interruptible") run_test(I915_EXEC_BLT, true, 0); igt_subtest("render-interruptible") run_test(I915_EXEC_RENDER, true, 0); igt_subtest("blitter-forked-1") run_test(I915_EXEC_BLT, false, 1); igt_subtest("render-forked-1") run_test(I915_EXEC_RENDER, false, 1); igt_subtest("blitter-forked-4") run_test(I915_EXEC_BLT, false, 4); igt_subtest("render-forked-4") run_test(I915_EXEC_RENDER, false, 4); igt_fixture { close(fd); } }