You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
376 lines
9.1 KiB
376 lines
9.1 KiB
/*
|
|
* Copyright © 2013 Intel Corporation
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
|
* IN THE SOFTWARE.
|
|
*
|
|
* Authors:
|
|
* Damien Lespiau <damien.lespiau@intel.com>
|
|
*/
|
|
|
|
/*
|
|
* The goal of this test is to ensure that we respect inter ring dependencies
|
|
*
|
|
* For each pair of rings R1, R2 where we have copy support (i.e. blt,
|
|
* rendercpy and mediafill) do:
|
|
* - Throw a busy load onto R1. gem_concurrent_blt just uses lots of buffers
|
|
* for this effect.
|
|
* - Fill three buffers A, B, C with unique data.
|
|
* - Copy A to B on ring R1
|
|
*
|
|
* Then come the three different variants.
|
|
* - Copy B to C on ring R2, check that C now contains what A originally
|
|
* contained. This is the write->read hazard. gem_concurrent_blt calls this
|
|
* early read.
|
|
* - Copy C to A on ring R2, check that B now contains what A originally
|
|
* contained. This is the read->write hazard, gem_concurrent_blt calls it
|
|
* overwrite_source.
|
|
* - Copy C to B on ring R2 and check that B contains what C originally
|
|
* contained. This is the write/write hazard. gem_concurrent_blt doesn't
|
|
* have that since for the cpu case it's too boring.
|
|
*
|
|
*/
|
|
|
|
#include "igt.h"
|
|
#include <stdlib.h>
|
|
#include <stdbool.h>
|
|
|
|
|
|
IGT_TEST_DESCRIPTION("Ensure inter-ring dependencies are respected.");
|
|
|
|
#define WIDTH 512
|
|
#define HEIGHT 512
|
|
#define NUM_BUSY_BUFFERS 32
|
|
|
|
typedef struct {
|
|
int drm_fd;
|
|
uint32_t devid;
|
|
drm_intel_bufmgr *bufmgr;
|
|
struct intel_batchbuffer *batch;
|
|
|
|
/* number of buffers to keep the ring busy for a while */
|
|
unsigned int n_buffers_load;
|
|
|
|
uint32_t linear[WIDTH * HEIGHT];
|
|
|
|
struct {
|
|
igt_render_copyfunc_t copy;
|
|
struct igt_buf *srcs;
|
|
struct igt_buf *dsts;
|
|
} render;
|
|
|
|
struct {
|
|
drm_intel_bo **srcs;
|
|
drm_intel_bo **dsts;
|
|
} blitter;
|
|
|
|
} data_t;
|
|
|
|
enum ring {
|
|
RENDER,
|
|
BLITTER,
|
|
};
|
|
|
|
enum test {
|
|
TEST_WRITE_READ,
|
|
TEST_READ_WRITE,
|
|
TEST_WRITE_WRITE,
|
|
};
|
|
|
|
static const char *ring_name(enum ring ring)
|
|
{
|
|
const char *names[] = {
|
|
"render",
|
|
"blitter",
|
|
};
|
|
|
|
return names[ring];
|
|
}
|
|
|
|
static drm_intel_bo *bo_create(data_t *data, int width, int height, int val)
|
|
{
|
|
drm_intel_bo *bo;
|
|
int i;
|
|
|
|
bo = drm_intel_bo_alloc(data->bufmgr, "", 4 * width * height, 4096);
|
|
igt_assert(bo);
|
|
|
|
for (i = 0; i < width * height; i++)
|
|
data->linear[i] = val;
|
|
gem_write(data->drm_fd, bo->handle, 0, data->linear,
|
|
sizeof(data->linear));
|
|
|
|
return bo;
|
|
}
|
|
|
|
static void bo_check(data_t *data, drm_intel_bo *bo, uint32_t val)
|
|
{
|
|
int i;
|
|
|
|
gem_read(data->drm_fd, bo->handle, 0,
|
|
data->linear, sizeof(data->linear));
|
|
for (i = 0; i < WIDTH * HEIGHT; i++)
|
|
igt_assert_eq_u32(data->linear[i], val);
|
|
}
|
|
|
|
static void scratch_buf_init_from_bo(struct igt_buf *buf, drm_intel_bo *bo)
|
|
{
|
|
memset(buf, 0, sizeof(*buf));
|
|
|
|
buf->bo = bo;
|
|
buf->stride = 4 * WIDTH;
|
|
buf->tiling = I915_TILING_NONE;
|
|
buf->size = 4 * WIDTH * HEIGHT;
|
|
buf->bpp = 32;
|
|
}
|
|
|
|
static void scratch_buf_init(data_t *data, struct igt_buf *buf,
|
|
int width, int height, uint32_t color)
|
|
{
|
|
drm_intel_bo *bo;
|
|
|
|
bo = bo_create(data, width, height, color);
|
|
scratch_buf_init_from_bo(buf, bo);
|
|
}
|
|
|
|
/*
|
|
* Provide a few ring specific vfuncs for run_test().
|
|
*
|
|
* busy() Queue a n_buffers_load workloads onto the ring to keep it busy
|
|
* busy_fini() Clean up after busy
|
|
* copy() Copy one BO to another
|
|
*/
|
|
|
|
/*
|
|
* Render ring
|
|
*/
|
|
|
|
static void render_busy(data_t *data)
|
|
{
|
|
size_t array_size;
|
|
int i;
|
|
|
|
/* allocate 32 buffer objects and re-use them as needed */
|
|
array_size = NUM_BUSY_BUFFERS * sizeof(struct igt_buf);
|
|
|
|
data->render.srcs = malloc(array_size);
|
|
data->render.dsts = malloc(array_size);
|
|
|
|
for (i = 0; i < NUM_BUSY_BUFFERS; i++) {
|
|
scratch_buf_init(data, &data->render.srcs[i], WIDTH, HEIGHT,
|
|
0xdeadbeef);
|
|
scratch_buf_init(data, &data->render.dsts[i], WIDTH, HEIGHT,
|
|
0xdeadbeef);
|
|
}
|
|
|
|
for (i = 0; i < data->n_buffers_load; i++) {
|
|
data->render.copy(data->batch,
|
|
NULL, /* context */
|
|
&data->render.srcs[i % NUM_BUSY_BUFFERS],
|
|
0, 0, /* src_x, src_y */
|
|
WIDTH, HEIGHT,
|
|
&data->render.dsts[i % NUM_BUSY_BUFFERS],
|
|
0, 0 /* dst_x, dst_y */);
|
|
}
|
|
}
|
|
|
|
static void render_busy_fini(data_t *data)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < NUM_BUSY_BUFFERS; i++) {
|
|
drm_intel_bo_unreference(data->render.srcs[i].bo);
|
|
drm_intel_bo_unreference(data->render.dsts[i].bo);
|
|
}
|
|
|
|
free(data->render.srcs);
|
|
free(data->render.dsts);
|
|
data->render.srcs = NULL;
|
|
data->render.dsts = NULL;
|
|
}
|
|
|
|
static void render_copy(data_t *data, drm_intel_bo *src, drm_intel_bo *dst)
|
|
{
|
|
struct igt_buf src_buf, dst_buf;
|
|
|
|
scratch_buf_init_from_bo(&src_buf, src);
|
|
scratch_buf_init_from_bo(&dst_buf, dst);
|
|
|
|
data->render.copy(data->batch,
|
|
NULL, /* context */
|
|
&src_buf,
|
|
0, 0, /* src_x, src_y */
|
|
WIDTH, HEIGHT,
|
|
&dst_buf,
|
|
0, 0 /* dst_x, dst_y */);
|
|
}
|
|
|
|
/*
|
|
* Blitter ring
|
|
*/
|
|
|
|
static void blitter_busy(data_t *data)
|
|
{
|
|
size_t array_size;
|
|
int i;
|
|
|
|
/* allocate 32 buffer objects and re-use them as needed */
|
|
array_size = NUM_BUSY_BUFFERS * sizeof(drm_intel_bo *);
|
|
|
|
data->blitter.srcs = malloc(array_size);
|
|
data->blitter.dsts = malloc(array_size);
|
|
|
|
for (i = 0; i < NUM_BUSY_BUFFERS; i++) {
|
|
data->blitter.srcs[i] = bo_create(data,
|
|
WIDTH, HEIGHT,
|
|
0xdeadbeef);
|
|
data->blitter.dsts[i] = bo_create(data,
|
|
WIDTH, HEIGHT,
|
|
0xdeadbeef);
|
|
}
|
|
|
|
for (i = 0; i < data->n_buffers_load; i++) {
|
|
intel_copy_bo(data->batch,
|
|
data->blitter.srcs[i % NUM_BUSY_BUFFERS],
|
|
data->blitter.dsts[i % NUM_BUSY_BUFFERS],
|
|
WIDTH*HEIGHT*4);
|
|
}
|
|
}
|
|
|
|
static void blitter_busy_fini(data_t *data)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < NUM_BUSY_BUFFERS; i++) {
|
|
drm_intel_bo_unreference(data->blitter.srcs[i]);
|
|
drm_intel_bo_unreference(data->blitter.dsts[i]);
|
|
}
|
|
|
|
free(data->blitter.srcs);
|
|
free(data->blitter.dsts);
|
|
data->blitter.srcs = NULL;
|
|
data->blitter.dsts = NULL;
|
|
}
|
|
|
|
static void blitter_copy(data_t *data, drm_intel_bo *src, drm_intel_bo *dst)
|
|
{
|
|
intel_copy_bo(data->batch, dst, src, WIDTH*HEIGHT*4);
|
|
}
|
|
|
|
struct ring_ops {
|
|
void (*busy)(data_t *data);
|
|
void (*busy_fini)(data_t *data);
|
|
void (*copy)(data_t *data, drm_intel_bo *src, drm_intel_bo *dst);
|
|
} ops [] = {
|
|
{
|
|
.busy = render_busy,
|
|
.busy_fini = render_busy_fini,
|
|
.copy = render_copy,
|
|
},
|
|
{
|
|
.busy = blitter_busy,
|
|
.busy_fini = blitter_busy_fini,
|
|
.copy = blitter_copy,
|
|
},
|
|
};
|
|
|
|
static void run_test(data_t *data, enum ring r1, enum ring r2, enum test test)
|
|
{
|
|
struct ring_ops *r1_ops = &ops[r1];
|
|
struct ring_ops *r2_ops = &ops[r2];
|
|
drm_intel_bo *a, *b, *c;
|
|
|
|
a = bo_create(data, WIDTH, HEIGHT, 0xa);
|
|
b = bo_create(data, WIDTH, HEIGHT, 0xb);
|
|
c = bo_create(data, WIDTH, HEIGHT, 0xc);
|
|
|
|
r1_ops->busy(data);
|
|
r1_ops->copy(data, a, b);
|
|
|
|
switch (test) {
|
|
case TEST_WRITE_READ:
|
|
r2_ops->copy(data, b, c);
|
|
bo_check(data, c, 0xa);
|
|
break;
|
|
case TEST_READ_WRITE:
|
|
r2_ops->copy(data, c, a);
|
|
bo_check(data, b, 0xa);
|
|
break;
|
|
case TEST_WRITE_WRITE:
|
|
r2_ops->copy(data, c, b);
|
|
bo_check(data, b, 0xc);
|
|
break;
|
|
default:
|
|
igt_fail(IGT_EXIT_FAILURE);
|
|
}
|
|
|
|
r1_ops->busy_fini(data);
|
|
}
|
|
|
|
igt_main
|
|
{
|
|
data_t data = {0, };
|
|
int i;
|
|
struct combination {
|
|
int r1, r2;
|
|
} ring_combinations [] = {
|
|
{ RENDER, BLITTER },
|
|
{ BLITTER, RENDER },
|
|
};
|
|
|
|
igt_fixture {
|
|
data.drm_fd = drm_open_driver_render(DRIVER_INTEL);
|
|
data.devid = intel_get_drm_devid(data.drm_fd);
|
|
|
|
data.n_buffers_load = 1000;
|
|
|
|
data.bufmgr = drm_intel_bufmgr_gem_init(data.drm_fd, 4096);
|
|
igt_assert(data.bufmgr);
|
|
drm_intel_bufmgr_gem_enable_reuse(data.bufmgr);
|
|
|
|
data.render.copy = igt_get_render_copyfunc(data.devid);
|
|
igt_require_f(data.render.copy,
|
|
"no render-copy function\n");
|
|
|
|
data.batch = intel_batchbuffer_alloc(data.bufmgr, data.devid);
|
|
igt_assert(data.batch);
|
|
}
|
|
|
|
for (i = 0; i < ARRAY_SIZE(ring_combinations); i++) {
|
|
struct combination *c = &ring_combinations[i];
|
|
|
|
igt_subtest_f("sync-%s-%s-write-read",
|
|
ring_name(c->r1), ring_name(c->r2))
|
|
run_test(&data, c->r1, c->r2, TEST_WRITE_READ);
|
|
|
|
igt_subtest_f("sync-%s-%s-read-write",
|
|
ring_name(c->r1), ring_name(c->r2))
|
|
run_test(&data, c->r1, c->r2, TEST_READ_WRITE);
|
|
igt_subtest_f("sync-%s-%s-write-write",
|
|
ring_name(c->r1), ring_name(c->r2))
|
|
run_test(&data, c->r1, c->r2, TEST_WRITE_WRITE);
|
|
}
|
|
|
|
igt_fixture {
|
|
intel_batchbuffer_free(data.batch);
|
|
drm_intel_bufmgr_destroy(data.bufmgr);
|
|
close(data.drm_fd);
|
|
}
|
|
}
|