You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
906 lines
24 KiB
906 lines
24 KiB
/* basic set of prime tests between intel and nouveau */
|
|
|
|
/* test list -
|
|
1. share buffer from intel -> nouveau.
|
|
2. share buffer from nouveau -> intel
|
|
3. share intel->nouveau, map on both, write intel, read nouveau
|
|
4. share intel->nouveau, blit intel fill, readback on nouveau
|
|
test 1 + map buffer, read/write, map other size.
|
|
do some hw actions on the buffer
|
|
some illegal operations -
|
|
close prime fd try and map
|
|
|
|
TODO add some nouveau rendering tests
|
|
*/
|
|
|
|
|
|
#include "igt.h"
|
|
#include <inttypes.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <unistd.h>
|
|
#include <fcntl.h>
|
|
#include <string.h>
|
|
#include <sys/stat.h>
|
|
#include <sys/ioctl.h>
|
|
#include <errno.h>
|
|
|
|
#include "intel_bufmgr.h"
|
|
#include "nouveau.h"
|
|
|
|
static int intel_fd = -1, nouveau_fd = -1;
|
|
static drm_intel_bufmgr *bufmgr;
|
|
static struct nouveau_device *ndev;
|
|
static struct nouveau_client *nclient;
|
|
static uint32_t devid;
|
|
static struct intel_batchbuffer *batch;
|
|
static struct nouveau_object *nchannel, *pcopy;
|
|
static struct nouveau_bufctx *nbufctx;
|
|
static struct nouveau_pushbuf *npush;
|
|
|
|
static struct nouveau_bo *query_bo;
|
|
static uint32_t query_counter;
|
|
static volatile uint32_t *query;
|
|
static uint32_t memtype_intel, tile_intel_y, tile_intel_x;
|
|
|
|
#define SUBC_COPY(x) 6, (x)
|
|
#define NV01_SUBCHAN_OBJECT 0
|
|
|
|
#define NV01_SUBC(subc, mthd) SUBC_##subc((NV01_SUBCHAN_##mthd))
|
|
|
|
typedef struct {
|
|
uint32_t w, h;
|
|
uint32_t pitch, lines;
|
|
} rect;
|
|
|
|
static void nv_bo_alloc(struct nouveau_bo **bo, rect *r,
|
|
uint32_t w, uint32_t h, uint32_t tile_mode,
|
|
int handle, uint32_t dom)
|
|
{
|
|
uint32_t size;
|
|
uint32_t dx = 1, dy = 1, memtype = 0;
|
|
|
|
*bo = NULL;
|
|
if (tile_mode) {
|
|
uint32_t tile_y;
|
|
uint32_t tile_x;
|
|
|
|
/* Y major tiling */
|
|
if ((tile_mode & 0xf) == 0xe)
|
|
/* but the internal layout is different */
|
|
tile_x = 7;
|
|
else
|
|
tile_x = 6 + (tile_mode & 0xf);
|
|
if (ndev->chipset < 0xc0) {
|
|
memtype = 0x70;
|
|
tile_y = 2;
|
|
} else {
|
|
memtype = 0xfe;
|
|
tile_y = 3;
|
|
}
|
|
if ((tile_mode & 0xf) == 0xe)
|
|
memtype = memtype_intel;
|
|
tile_y += ((tile_mode & 0xf0)>>4);
|
|
|
|
dx = 1 << tile_x;
|
|
dy = 1 << tile_y;
|
|
igt_debug("Tiling requirements: x y %u %u\n", dx, dy);
|
|
}
|
|
|
|
r->w = w;
|
|
r->h = h;
|
|
|
|
r->pitch = w = (w + dx-1) & ~(dx-1);
|
|
r->lines = h = (h + dy-1) & ~(dy-1);
|
|
size = w*h;
|
|
|
|
if (handle < 0) {
|
|
union nouveau_bo_config cfg;
|
|
cfg.nv50.memtype = memtype;
|
|
cfg.nv50.tile_mode = tile_mode;
|
|
if (dom == NOUVEAU_BO_GART)
|
|
dom |= NOUVEAU_BO_MAP;
|
|
igt_assert(nouveau_bo_new(ndev, dom, 4096, size, &cfg, bo) == 0);
|
|
igt_assert(nouveau_bo_map(*bo, NOUVEAU_BO_RDWR, nclient) == 0);
|
|
|
|
igt_debug("new flags %08x memtype %08x tile %08x\n",
|
|
(*bo)->flags, (*bo)->config.nv50.memtype,
|
|
(*bo)->config.nv50.tile_mode);
|
|
if (tile_mode == tile_intel_y || tile_mode == tile_intel_x) {
|
|
igt_debug("tile mode was: %02x, now: %02x\n",
|
|
(*bo)->config.nv50.tile_mode, tile_mode);
|
|
/* Doesn't like intel tiling much.. */
|
|
(*bo)->config.nv50.tile_mode = tile_mode;
|
|
}
|
|
} else {
|
|
igt_assert(nouveau_bo_prime_handle_ref(ndev, handle, bo) == 0);
|
|
close(handle);
|
|
igt_assert_f((*bo)->size >= size,
|
|
"expected bo size to be at least %u,"
|
|
"but received %"PRIu64"\n", size, (*bo)->size);
|
|
igt_debug("prime flags %08x memtype %08x tile %08x\n",
|
|
(*bo)->flags, (*bo)->config.nv50.memtype,
|
|
(*bo)->config.nv50.tile_mode);
|
|
(*bo)->config.nv50.memtype = memtype;
|
|
(*bo)->config.nv50.tile_mode = tile_mode;
|
|
}
|
|
igt_debug("size: %"PRIu64"\n", (*bo)->size);
|
|
}
|
|
|
|
static inline void
|
|
PUSH_DATA(struct nouveau_pushbuf *push, uint32_t data)
|
|
{
|
|
*push->cur++ = data;
|
|
}
|
|
|
|
static inline void
|
|
BEGIN_NV04(struct nouveau_pushbuf *push, int subc, int mthd, int size)
|
|
{
|
|
PUSH_DATA (push, 0x00000000 | (size << 18) | (subc << 13) | mthd);
|
|
}
|
|
|
|
static inline void
|
|
BEGIN_NI04(struct nouveau_pushbuf *push, int subc, int mthd, int size)
|
|
{
|
|
PUSH_DATA (push, 0x40000000 | (size << 18) | (subc << 13) | mthd);
|
|
}
|
|
|
|
static inline void
|
|
BEGIN_NVC0(struct nouveau_pushbuf *push, int subc, int mthd, int size)
|
|
{
|
|
PUSH_DATA (push, 0x20000000 | (size << 16) | (subc << 13) | (mthd / 4));
|
|
}
|
|
|
|
static inline void
|
|
BEGIN_NVXX(struct nouveau_pushbuf *push, int subc, int mthd, int size)
|
|
{
|
|
if (ndev->chipset < 0xc0)
|
|
BEGIN_NV04(push, subc, mthd, size);
|
|
else
|
|
BEGIN_NVC0(push, subc, mthd, size);
|
|
}
|
|
|
|
static void
|
|
noop_intel(drm_intel_bo *bo)
|
|
{
|
|
BEGIN_BATCH(3, 1);
|
|
OUT_BATCH(MI_NOOP);
|
|
OUT_BATCH(MI_BATCH_BUFFER_END);
|
|
OUT_RELOC(bo, I915_GEM_DOMAIN_RENDER,
|
|
I915_GEM_DOMAIN_RENDER, 0);
|
|
ADVANCE_BATCH();
|
|
|
|
intel_batchbuffer_flush(batch);
|
|
}
|
|
|
|
static void find_and_open_devices(void)
|
|
{
|
|
int i;
|
|
char path[80], *unused;
|
|
struct stat buf;
|
|
FILE *fl;
|
|
char vendor_id[8] = {};
|
|
int venid;
|
|
for (i = 0; i < 9; i++) {
|
|
sprintf(path, "/sys/class/drm/card%d/device/vendor", i);
|
|
if (stat(path, &buf))
|
|
break;
|
|
|
|
fl = fopen(path, "r");
|
|
if (!fl)
|
|
break;
|
|
|
|
unused = fgets(vendor_id, sizeof(vendor_id)-1, fl);
|
|
(void)unused;
|
|
fclose(fl);
|
|
|
|
venid = strtoul(vendor_id, NULL, 16);
|
|
sprintf(path, "/dev/dri/card%d", i);
|
|
if (venid == 0x8086) {
|
|
intel_fd = open(path, O_RDWR);
|
|
igt_assert(intel_fd);
|
|
} else if (venid == 0x10de) {
|
|
nouveau_fd = open(path, O_RDWR);
|
|
igt_assert(nouveau_fd);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void init_nouveau(void)
|
|
{
|
|
struct nv04_fifo nv04_data = { .vram = 0xbeef0201,
|
|
.gart = 0xbeef0202 };
|
|
struct nvc0_fifo nvc0_data = { };
|
|
struct nouveau_fifo *fifo;
|
|
int size;
|
|
uint32_t class;
|
|
void *data;
|
|
|
|
igt_assert(nouveau_device_wrap(nouveau_fd, 0, &ndev) == 0);
|
|
|
|
igt_assert(nouveau_client_new(ndev, &nclient) == 0);
|
|
|
|
igt_skip_on_f(ndev->chipset < 0xa3 || ndev->chipset == 0xaa || ndev->chipset == 0xac,
|
|
"Your card doesn't support PCOPY\n");
|
|
|
|
// TODO: Get a kepler and add support for it
|
|
igt_skip_on_f(ndev->chipset >= 0xe0,
|
|
"Unsure how kepler works!\n");
|
|
igt_assert(nouveau_bo_new(ndev, NOUVEAU_BO_GART | NOUVEAU_BO_MAP,
|
|
4096, 4096, NULL, &query_bo) == 0);
|
|
igt_assert(nouveau_bo_map(query_bo, NOUVEAU_BO_RDWR, nclient) == 0);
|
|
query = query_bo->map;
|
|
*query = query_counter;
|
|
|
|
if (ndev->chipset < 0xc0) {
|
|
class = 0x85b5;
|
|
data = &nv04_data;
|
|
size = sizeof(nv04_data);
|
|
} else {
|
|
class = ndev->chipset < 0xe0 ? 0x490b5 : 0xa0b5;
|
|
data = &nvc0_data;
|
|
size = sizeof(nvc0_data);
|
|
}
|
|
|
|
igt_assert(nouveau_object_new(&ndev->object, 0, NOUVEAU_FIFO_CHANNEL_CLASS,
|
|
data, size, &nchannel) == 0);
|
|
|
|
fifo = nchannel->data;
|
|
|
|
igt_assert(nouveau_pushbuf_new(nclient, nchannel, 4, 32 * 1024,
|
|
true, &npush) == 0);
|
|
|
|
igt_assert(nouveau_bufctx_new(nclient, 1, &nbufctx) == 0);
|
|
|
|
npush->user_priv = nbufctx;
|
|
|
|
/* Hope this is enough init for PCOPY */
|
|
igt_assert(nouveau_object_new(nchannel, class, class & 0xffff, NULL, 0, &pcopy) == 0);
|
|
igt_assert(nouveau_pushbuf_space(npush, 512, 0, 0) == 0);
|
|
|
|
if (ndev->chipset < 0xc0) {
|
|
struct nv04_fifo *nv04_fifo = (struct nv04_fifo*)fifo;
|
|
tile_intel_y = 0x3e;
|
|
tile_intel_x = 0x13;
|
|
|
|
BEGIN_NV04(npush, NV01_SUBC(COPY, OBJECT), 1);
|
|
PUSH_DATA(npush, pcopy->handle);
|
|
BEGIN_NV04(npush, SUBC_COPY(0x0180), 3);
|
|
PUSH_DATA(npush, nv04_fifo->vram);
|
|
PUSH_DATA(npush, nv04_fifo->vram);
|
|
PUSH_DATA(npush, nv04_fifo->vram);
|
|
} else {
|
|
tile_intel_y = 0x2e;
|
|
tile_intel_x = 0x03;
|
|
BEGIN_NVC0(npush, NV01_SUBC(COPY, OBJECT), 1);
|
|
PUSH_DATA(npush, pcopy->handle);
|
|
}
|
|
nouveau_pushbuf_kick(npush, npush->channel);
|
|
}
|
|
|
|
static void fill16(void *ptr, uint32_t val)
|
|
{
|
|
uint32_t *p = ptr;
|
|
val = (val) | (val << 8) | (val << 16) | (val << 24);
|
|
p[0] = p[1] = p[2] = p[3] = val;
|
|
}
|
|
|
|
#define TILE_SIZE 4096
|
|
|
|
static void swtile_y(uint8_t *out, const uint8_t *in, int w, int h)
|
|
{
|
|
uint32_t x, y, dx, dy;
|
|
uint8_t *endptr = out + w * h;
|
|
igt_assert(!(w % 128));
|
|
igt_assert(!(h % 32));
|
|
|
|
for (y = 0; y < h; y += 32) {
|
|
for (x = 0; x < w; x += 128, out += TILE_SIZE) {
|
|
for (dx = 0; dx < 8; ++dx) {
|
|
for (dy = 0; dy < 32; ++dy) {
|
|
uint32_t out_ofs = (dx * 32 + dy) * 16;
|
|
uint32_t in_ofs = (y + dy) * w + (x + 16 * dx);
|
|
igt_assert(out_ofs < TILE_SIZE);
|
|
igt_assert(in_ofs < w*h);
|
|
|
|
// To do the Y tiling quirk:
|
|
// out_ofs = out_ofs ^ (((out_ofs >> 9) & 1) << 6);
|
|
memcpy(&out[out_ofs], &in[in_ofs], 16);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
igt_assert(out == endptr);
|
|
}
|
|
|
|
static void swtile_x(uint8_t *out, const uint8_t *in, int w, int h)
|
|
{
|
|
uint32_t x, y, dy;
|
|
uint8_t *endptr = out + w * h;
|
|
igt_assert(!(w % 512));
|
|
igt_assert(!(h % 8));
|
|
|
|
for (y = 0; y < h; y += 8) {
|
|
for (x = 0; x < w; x += 512, out += TILE_SIZE) {
|
|
for (dy = 0; dy < 8; ++dy) {
|
|
uint32_t out_ofs = 512 * dy;
|
|
uint32_t in_ofs = (y + dy) * w + x;
|
|
igt_assert(out_ofs < TILE_SIZE);
|
|
igt_assert(in_ofs < w*h);
|
|
memcpy(&out[out_ofs], &in[in_ofs], 512);
|
|
}
|
|
}
|
|
}
|
|
igt_assert(out == endptr);
|
|
}
|
|
|
|
static void perform_copy(struct nouveau_bo *nvbo, const rect *dst,
|
|
uint32_t dst_x, uint32_t dst_y,
|
|
struct nouveau_bo *nvbi, const rect *src,
|
|
uint32_t src_x, uint32_t src_y,
|
|
uint32_t w, uint32_t h)
|
|
{
|
|
struct nouveau_pushbuf_refn refs[] = {
|
|
{ nvbi, (nvbi->flags & NOUVEAU_BO_APER) | NOUVEAU_BO_RD },
|
|
{ nvbo, (nvbo->flags & NOUVEAU_BO_APER) | NOUVEAU_BO_WR },
|
|
{ query_bo, NOUVEAU_BO_GART | NOUVEAU_BO_RDWR }
|
|
};
|
|
uint32_t cpp = 1, exec = 0x00003000; /* QUERY|QUERY_SHORT|FORMAT */
|
|
uint32_t src_off = 0, dst_off = 0;
|
|
struct nouveau_pushbuf *push = npush;
|
|
int ret;
|
|
|
|
if (nvbi->config.nv50.tile_mode == tile_intel_y)
|
|
igt_debug("src is y-tiled\n");
|
|
if (nvbo->config.nv50.tile_mode == tile_intel_y)
|
|
igt_debug("dst is y-tiled\n");
|
|
|
|
igt_assert(nouveau_pushbuf_space(push, 64, 0, 0) == 0);
|
|
igt_assert(nouveau_pushbuf_refn(push, refs, 3) == 0);
|
|
|
|
if (!nvbi->config.nv50.tile_mode) {
|
|
src_off = src_y * src->pitch + src_x;
|
|
exec |= 0x00000010;
|
|
}
|
|
|
|
if (!nvbo->config.nv50.tile_mode) {
|
|
dst_off = dst_y * dst->pitch + dst_x;
|
|
exec |= 0x00000100;
|
|
}
|
|
|
|
BEGIN_NVXX(push, SUBC_COPY(0x0200), 7);
|
|
PUSH_DATA (push, nvbi->config.nv50.tile_mode);
|
|
PUSH_DATA (push, src->pitch / cpp);
|
|
PUSH_DATA (push, src->h);
|
|
PUSH_DATA (push, 1);
|
|
PUSH_DATA (push, 0);
|
|
PUSH_DATA (push, src_x / cpp);
|
|
PUSH_DATA (push, src_y);
|
|
|
|
BEGIN_NVXX(push, SUBC_COPY(0x0220), 7);
|
|
PUSH_DATA (push, nvbo->config.nv50.tile_mode);
|
|
PUSH_DATA (push, dst->pitch / cpp);
|
|
PUSH_DATA (push, dst->h);
|
|
PUSH_DATA (push, 1);
|
|
PUSH_DATA (push, 0);
|
|
PUSH_DATA (push, dst_x / cpp);
|
|
PUSH_DATA (push, dst_y);
|
|
|
|
BEGIN_NVXX(push, SUBC_COPY(0x030c), 9);
|
|
PUSH_DATA (push, (nvbi->offset + src_off) >> 32);
|
|
PUSH_DATA (push, (nvbi->offset + src_off));
|
|
PUSH_DATA (push, (nvbo->offset + dst_off) >> 32);
|
|
PUSH_DATA (push, (nvbo->offset + dst_off));
|
|
PUSH_DATA (push, src->pitch);
|
|
PUSH_DATA (push, dst->pitch);
|
|
PUSH_DATA (push, w / cpp);
|
|
PUSH_DATA (push, h);
|
|
PUSH_DATA (push, 0x03333120);
|
|
|
|
BEGIN_NVXX(push, SUBC_COPY(0x0338), 3);
|
|
PUSH_DATA (push, (query_bo->offset) >> 32);
|
|
PUSH_DATA (push, (query_bo->offset));
|
|
PUSH_DATA (push, ++query_counter);
|
|
|
|
BEGIN_NVXX(push, SUBC_COPY(0x0300), 1);
|
|
PUSH_DATA (push, exec);
|
|
|
|
ret = nouveau_pushbuf_kick(push, push->channel);
|
|
while (!ret && *query < query_counter) { usleep(1000); }
|
|
|
|
igt_assert_eq(ret, 0);
|
|
}
|
|
|
|
static void check1_macro(uint32_t *p, uint32_t w, uint32_t h)
|
|
{
|
|
uint32_t i, val, j;
|
|
|
|
for (i = 0; i < 256; ++i, p += 4) {
|
|
val = (i) | (i << 8) | (i << 16) | (i << 24);
|
|
igt_assert_f(p[0] == val && p[1] == val && p[2] == val && p[3] == val,
|
|
"Retile check failed in first tile!\n"
|
|
"%08x %08x %08x %08x instead of %08x\n",
|
|
p[0], p[1], p[2], p[3], val);
|
|
}
|
|
|
|
val = 0x3e3e3e3e;
|
|
for (i = 0; i < 256 * (w-1); ++i, p += 4) {
|
|
igt_assert_f(p[0] == val && p[1] == val && p[2] == val && p[3] == val,
|
|
"Retile check failed in second tile!\n"
|
|
"%08x %08x %08x %08x instead of %08x\n",
|
|
p[0], p[1], p[2], p[3], val);
|
|
}
|
|
|
|
for (j = 1; j < h; ++j) {
|
|
val = 0x7e7e7e7e;
|
|
for (i = 0; i < 256; ++i, p += 4) {
|
|
igt_assert_f(p[0] == val && p[1] == val && p[2] == val && p[3] == val,
|
|
"Retile check failed in third tile!\n"
|
|
"%08x %08x %08x %08x instead of %08x\n",
|
|
p[0], p[1], p[2], p[3], val);
|
|
}
|
|
|
|
val = 0xcececece;
|
|
for (i = 0; i < 256 * (w-1); ++i, p += 4) {
|
|
igt_assert_f(p[0] == val && p[1] == val && p[2] == val && p[3] == val,
|
|
"Retile check failed in fourth tile!\n"
|
|
"%08x %08x %08x %08x instead of %08x\n",
|
|
p[0], p[1], p[2], p[3], val);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* test 1, see if we can copy from linear to intel Y format safely */
|
|
static void test1_macro(void)
|
|
{
|
|
int prime_fd = -1;
|
|
struct nouveau_bo *nvbo = NULL, *nvbi = NULL;
|
|
rect dst, src;
|
|
uint8_t *ptr;
|
|
uint32_t w = 2 * 128, h = 2 * 32, x, y;
|
|
|
|
nv_bo_alloc(&nvbi, &src, w, h, 0, -1, NOUVEAU_BO_GART);
|
|
nv_bo_alloc(&nvbo, &dst, w, h, tile_intel_y, -1, NOUVEAU_BO_GART);
|
|
|
|
nouveau_bo_set_prime(nvbo, &prime_fd);
|
|
|
|
/* Set up something for our tile that should map into the first
|
|
* y-major tile, assuming my understanding of documentation is
|
|
* correct
|
|
*/
|
|
|
|
/* First tile should be read out in groups of 16 bytes that
|
|
* are all set to a linear increasing value..
|
|
*/
|
|
ptr = nvbi->map;
|
|
for (x = 0; x < 128; x += 16)
|
|
for (y = 0; y < 32; ++y)
|
|
fill16(&ptr[y * w + x], x * 2 + y);
|
|
|
|
/* second tile */
|
|
for (x = 128; x < w; x += 16)
|
|
for (y = 0; y < 32; ++y)
|
|
fill16(&ptr[y * w + x], 0x3e);
|
|
|
|
/* third tile */
|
|
for (x = 0; x < 128; x += 16)
|
|
for (y = 32; y < h; ++y)
|
|
fill16(&ptr[y * w + x], 0x7e);
|
|
|
|
/* last tile */
|
|
for (x = 128; x < w; x += 16)
|
|
for (y = 32; y < h; ++y)
|
|
fill16(&ptr[y * w + x], 0xce);
|
|
memset(nvbo->map, 0xfc, w * h);
|
|
|
|
if (pcopy)
|
|
perform_copy(nvbo, &dst, 0, 0, nvbi, &src, 0, 0, w, h);
|
|
else
|
|
swtile_y(nvbo->map, nvbi->map, w, h);
|
|
check1_macro(nvbo->map, w/128, h/32);
|
|
|
|
nouveau_bo_ref(NULL, &nvbo);
|
|
nouveau_bo_ref(NULL, &nvbi);
|
|
close(prime_fd);
|
|
}
|
|
|
|
static void dump_line(uint8_t *map)
|
|
{
|
|
uint32_t dx, dy;
|
|
igt_debug("Dumping sub-tile:\n");
|
|
for (dy = 0; dy < 32; ++dy) {
|
|
for (dx = 0; dx < 15; ++dx, ++map) {
|
|
igt_debug("%02x ", *map);
|
|
}
|
|
igt_debug("%02x\n", *(map++));
|
|
}
|
|
}
|
|
|
|
static void check1_micro(void *map, uint32_t pitch, uint32_t lines,
|
|
uint32_t dst_x, uint32_t dst_y, uint32_t w, uint32_t h)
|
|
{
|
|
uint32_t x, y;
|
|
|
|
/* check only the relevant subrectangle [0..w) [0...h) */
|
|
uint8_t *m = map;
|
|
for (y = 0; y < h; ++y, m += pitch) {
|
|
for (x = 0; x < w; ++x) {
|
|
uint8_t expected = ((y & 3) << 6) | (x & 0x3f);
|
|
|
|
if (expected != m[x])
|
|
dump_line(m);
|
|
|
|
igt_assert_f(expected == m[x],
|
|
"failed check at x=%u y=%u, expected %02x got %02x\n",
|
|
x, y, expected, m[x]);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* test 1, but check micro format, should be unaffected by bit9 swizzling */
|
|
static void test1_micro(void)
|
|
{
|
|
struct nouveau_bo *bo_intel = NULL, *bo_nvidia = NULL, *bo_linear = NULL;
|
|
rect intel, nvidia, linear;
|
|
uint32_t tiling = I915_TILING_Y;
|
|
|
|
uint32_t src_x = 0, src_y = 0;
|
|
uint32_t dst_x = 0, dst_y = 0;
|
|
uint32_t x, y, w = 256, h = 64;
|
|
|
|
drm_intel_bo *test_intel_bo;
|
|
int prime_fd;
|
|
|
|
test_intel_bo = drm_intel_bo_alloc(bufmgr, "test bo", w * h, 4096);
|
|
igt_assert(test_intel_bo);
|
|
drm_intel_bo_set_tiling(test_intel_bo, &tiling, w);
|
|
igt_assert(tiling == I915_TILING_Y);
|
|
igt_assert(drm_intel_gem_bo_map_gtt(test_intel_bo) == 0);
|
|
|
|
drm_intel_bo_gem_export_to_prime(test_intel_bo, &prime_fd);
|
|
igt_assert_lte(0, prime_fd);
|
|
noop_intel(test_intel_bo);
|
|
|
|
nv_bo_alloc(&bo_intel, &intel, w, h, tile_intel_y, prime_fd, 0);
|
|
nv_bo_alloc(&bo_nvidia, &nvidia, w, h, 0x10, -1, NOUVEAU_BO_VRAM);
|
|
nv_bo_alloc(&bo_linear, &linear, w, h, 0, -1, NOUVEAU_BO_GART);
|
|
|
|
for (y = 0; y < linear.h; ++y) {
|
|
uint8_t *map = bo_linear->map;
|
|
map += y * linear.pitch;
|
|
for (x = 0; x < linear.pitch; ++x) {
|
|
uint8_t pos = x & 0x3f;
|
|
/* low 4 bits: micro tile pos */
|
|
/* 2 bits: x pos in tile (wraps) */
|
|
/* 2 bits: y pos in tile (wraps) */
|
|
pos |= (y & 3) << 6;
|
|
map[x] = pos;
|
|
}
|
|
}
|
|
|
|
perform_copy(bo_nvidia, &nvidia, 0, 0, bo_linear, &linear, 0, 0, nvidia.pitch, nvidia.h);
|
|
|
|
/* Perform the actual sub rectangle copy */
|
|
if (pcopy)
|
|
perform_copy(bo_intel, &intel, dst_x, dst_y, bo_nvidia, &nvidia, src_x, src_y, w, h);
|
|
else
|
|
swtile_y(test_intel_bo->virtual, bo_linear->map, w, h);
|
|
|
|
noop_intel(test_intel_bo);
|
|
check1_micro(test_intel_bo->virtual, intel.pitch, intel.h, dst_x, dst_y, w, h);
|
|
|
|
nouveau_bo_ref(NULL, &bo_linear);
|
|
nouveau_bo_ref(NULL, &bo_nvidia);
|
|
nouveau_bo_ref(NULL, &bo_intel);
|
|
drm_intel_bo_unreference(test_intel_bo);
|
|
}
|
|
|
|
/* test 2, see if we can copy from linear to intel X format safely
|
|
* Seems nvidia lacks a method to do it, so just keep this test
|
|
* as a reference for potential future tests. Software tiling is
|
|
* used for now
|
|
*/
|
|
static void test2(void)
|
|
{
|
|
struct nouveau_bo *nvbo = NULL, *nvbi = NULL;
|
|
rect dst, src;
|
|
uint8_t *ptr;
|
|
uint32_t w = 1024, h = 16, x, y;
|
|
|
|
nv_bo_alloc(&nvbi, &src, w, h, 0, -1, NOUVEAU_BO_GART);
|
|
nv_bo_alloc(&nvbo, &dst, w, h, tile_intel_x, -1, NOUVEAU_BO_GART);
|
|
|
|
/* Set up something for our tile that should map into the first
|
|
* y-major tile, assuming my understanding of documentation is
|
|
* correct
|
|
*/
|
|
|
|
/* First tile should be read out in groups of 16 bytes that
|
|
* are all set to a linear increasing value..
|
|
*/
|
|
ptr = nvbi->map;
|
|
for (y = 0; y < 8; ++y)
|
|
for (x = 0; x < 512; x += 16)
|
|
fill16(&ptr[y * w + x], (y * 512 + x)/16);
|
|
|
|
for (y = 0; y < 8; ++y)
|
|
for (x = 512; x < w; x += 16)
|
|
fill16(&ptr[y * w + x], 0x3e);
|
|
|
|
for (y = 8; y < h; ++y)
|
|
for (x = 0; x < 512; x += 16)
|
|
fill16(&ptr[y * w + x], 0x7e);
|
|
|
|
for (y = 8; y < h; ++y)
|
|
for (x = 512; x < w; x += 16)
|
|
fill16(&ptr[y * w + x], 0xce);
|
|
memset(nvbo->map, 0xfc, w * h);
|
|
|
|
/* do this in software, there is no X major tiling in PCOPY (yet?) */
|
|
if (0 && pcopy)
|
|
perform_copy(nvbo, &dst, 0, 0, nvbi, &src, 0, 0, w, h);
|
|
else
|
|
swtile_x(nvbo->map, nvbi->map, w, h);
|
|
check1_macro(nvbo->map, w/512, h/8);
|
|
|
|
nouveau_bo_ref(NULL, &nvbo);
|
|
nouveau_bo_ref(NULL, &nvbi);
|
|
}
|
|
|
|
static void check3(const uint32_t *p, uint32_t pitch, uint32_t lines,
|
|
uint32_t sub_x, uint32_t sub_y,
|
|
uint32_t sub_w, uint32_t sub_h)
|
|
{
|
|
uint32_t x, y;
|
|
|
|
sub_w += sub_x;
|
|
sub_h += sub_y;
|
|
|
|
igt_assert_f(p[pitch * lines / 4 - 1] != 0x03030303,
|
|
"copy failed: Not all lines have been copied back!\n");
|
|
|
|
for (y = 0; y < lines; ++y) {
|
|
for (x = 0; x < pitch; x += 4, ++p) {
|
|
uint32_t expected;
|
|
if ((x < sub_x || x >= sub_w) ||
|
|
(y < sub_y || y >= sub_h))
|
|
expected = 0x80808080;
|
|
else
|
|
expected = 0x04040404;
|
|
igt_assert_f(*p == expected,
|
|
"%u,%u should be %08x, but is %08x\n",
|
|
x, y, expected, *p);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* copy from nvidia bo to intel bo and copy to a linear bo to check if tiling went succesful */
|
|
static void test3_base(int tile_src, int tile_dst)
|
|
{
|
|
struct nouveau_bo *bo_intel = NULL, *bo_nvidia = NULL, *bo_linear = NULL;
|
|
rect intel, nvidia, linear;
|
|
uint32_t cpp = 4;
|
|
|
|
uint32_t src_x = 1 * cpp, src_y = 1;
|
|
uint32_t dst_x = 2 * cpp, dst_y = 26;
|
|
uint32_t w = 298 * cpp, h = 298;
|
|
|
|
drm_intel_bo *test_intel_bo;
|
|
int prime_fd;
|
|
|
|
test_intel_bo = drm_intel_bo_alloc(bufmgr, "test bo", 2048 * cpp * 768, 4096);
|
|
igt_assert(test_intel_bo);
|
|
|
|
drm_intel_bo_gem_export_to_prime(test_intel_bo, &prime_fd);
|
|
igt_assert_lte(0, prime_fd);
|
|
|
|
nv_bo_alloc(&bo_intel, &intel, 2048 * cpp, 768, tile_dst, prime_fd, 0);
|
|
nv_bo_alloc(&bo_nvidia, &nvidia, 300 * cpp, 300, tile_src, -1, NOUVEAU_BO_VRAM);
|
|
nv_bo_alloc(&bo_linear, &linear, 2048 * cpp, 768, 0, -1, NOUVEAU_BO_GART);
|
|
|
|
noop_intel(test_intel_bo);
|
|
memset(bo_linear->map, 0x80, bo_linear->size);
|
|
perform_copy(bo_intel, &intel, 0, 0, bo_linear, &linear, 0, 0, linear.pitch, linear.h);
|
|
noop_intel(test_intel_bo);
|
|
|
|
memset(bo_linear->map, 0x04, bo_linear->size);
|
|
perform_copy(bo_nvidia, &nvidia, 0, 0, bo_linear, &linear, 0, 0, nvidia.pitch, nvidia.h);
|
|
|
|
/* Perform the actual sub rectangle copy */
|
|
noop_intel(test_intel_bo);
|
|
perform_copy(bo_intel, &intel, dst_x, dst_y, bo_nvidia, &nvidia, src_x, src_y, w, h);
|
|
noop_intel(test_intel_bo);
|
|
|
|
memset(bo_linear->map, 0x3, bo_linear->size);
|
|
noop_intel(test_intel_bo);
|
|
perform_copy(bo_linear, &linear, 0, 0, bo_intel, &intel, 0, 0, intel.pitch, intel.h);
|
|
noop_intel(test_intel_bo);
|
|
|
|
check3(bo_linear->map, linear.pitch, linear.h, dst_x, dst_y, w, h);
|
|
|
|
nouveau_bo_ref(NULL, &bo_linear);
|
|
nouveau_bo_ref(NULL, &bo_nvidia);
|
|
nouveau_bo_ref(NULL, &bo_intel);
|
|
drm_intel_bo_unreference(test_intel_bo);
|
|
}
|
|
|
|
static void test3_1(void)
|
|
{
|
|
/* nvidia tiling to intel */
|
|
test3_base(0x40, tile_intel_y);
|
|
}
|
|
|
|
static void test3_2(void)
|
|
{
|
|
/* intel tiling to nvidia */
|
|
test3_base(tile_intel_y, 0x40);
|
|
}
|
|
|
|
static void test3_3(void)
|
|
{
|
|
/* intel tiling to linear */
|
|
test3_base(tile_intel_y, 0);
|
|
}
|
|
|
|
static void test3_4(void)
|
|
{
|
|
/* linear tiling to intel */
|
|
test3_base(0, tile_intel_y);
|
|
}
|
|
|
|
static void test3_5(void)
|
|
{
|
|
/* linear to linear */
|
|
test3_base(0, 0);
|
|
}
|
|
|
|
/* Acquire when == SEQUENCE */
|
|
#define SEMA_ACQUIRE_EQUAL 1
|
|
|
|
/* Release, and write a 16 byte query structure to sema:
|
|
* { (uint32)seq, (uint32)0, (uint64)timestamp } */
|
|
#define SEMA_WRITE_LONG 2
|
|
|
|
/* Acquire when >= SEQUENCE */
|
|
#define SEMA_ACQUIRE_GEQUAL 4
|
|
|
|
/* Test only new style semaphores, old ones are AWFUL */
|
|
static void test_semaphore(void)
|
|
{
|
|
drm_intel_bo *test_intel_bo = NULL;
|
|
struct nouveau_bo *sema_bo = NULL;
|
|
int prime_fd;
|
|
uint32_t *sema;
|
|
struct nouveau_pushbuf *push = npush;
|
|
|
|
igt_skip_on(ndev->chipset < 0x84);
|
|
|
|
/* Should probably be kept in sysmem */
|
|
test_intel_bo = drm_intel_bo_alloc(bufmgr, "semaphore bo", 4096, 4096);
|
|
igt_assert(test_intel_bo);
|
|
|
|
drm_intel_bo_gem_export_to_prime(test_intel_bo, &prime_fd);
|
|
igt_assert_lte(0, prime_fd);
|
|
igt_assert(nouveau_bo_prime_handle_ref(ndev, prime_fd, &sema_bo) == 0);
|
|
close(prime_fd);
|
|
|
|
igt_assert(drm_intel_gem_bo_map_gtt(test_intel_bo) == 0);
|
|
sema = test_intel_bo->virtual;
|
|
sema++;
|
|
*sema = 0;
|
|
|
|
igt_assert(nouveau_pushbuf_space(push, 64, 0, 0) == 0);
|
|
igt_assert(nouveau_pushbuf_refn(push, &(struct nouveau_pushbuf_refn)
|
|
{ sema_bo, NOUVEAU_BO_GART|NOUVEAU_BO_RDWR }, 1) == 0);
|
|
|
|
if (ndev->chipset < 0xc0) {
|
|
struct nv04_fifo *nv04_fifo = nchannel->data;
|
|
/* kernel binds it's own dma object here and overwrites old one,
|
|
* so just rebind vram every time we submit
|
|
*/
|
|
BEGIN_NV04(npush, SUBC_COPY(0x0060), 1);
|
|
PUSH_DATA(npush, nv04_fifo->vram);
|
|
}
|
|
BEGIN_NVXX(push, SUBC_COPY(0x0010), 4);
|
|
PUSH_DATA(push, sema_bo->offset >> 32);
|
|
PUSH_DATA(push, sema_bo->offset + 4);
|
|
PUSH_DATA(push, 2); // SEQUENCE
|
|
PUSH_DATA(push, SEMA_WRITE_LONG); // TRIGGER
|
|
|
|
BEGIN_NVXX(push, SUBC_COPY(0x0018), 2);
|
|
PUSH_DATA(push, 3);
|
|
PUSH_DATA(push, SEMA_ACQUIRE_EQUAL);
|
|
BEGIN_NVXX(push, SUBC_COPY(0x0018), 2);
|
|
PUSH_DATA(push, 4);
|
|
PUSH_DATA(push, SEMA_WRITE_LONG);
|
|
|
|
BEGIN_NVXX(push, SUBC_COPY(0x0018), 2);
|
|
PUSH_DATA(push, 5);
|
|
PUSH_DATA(push, SEMA_ACQUIRE_GEQUAL);
|
|
BEGIN_NVXX(push, SUBC_COPY(0x0018), 2);
|
|
PUSH_DATA(push, 6);
|
|
PUSH_DATA(push, SEMA_WRITE_LONG);
|
|
|
|
BEGIN_NVXX(push, SUBC_COPY(0x0018), 2);
|
|
PUSH_DATA(push, 7);
|
|
PUSH_DATA(push, SEMA_ACQUIRE_GEQUAL);
|
|
BEGIN_NVXX(push, SUBC_COPY(0x0018), 2);
|
|
PUSH_DATA(push, 9);
|
|
PUSH_DATA(push, SEMA_WRITE_LONG);
|
|
nouveau_pushbuf_kick(push, push->channel);
|
|
|
|
usleep(1000);
|
|
igt_assert(*sema == 2);
|
|
|
|
*sema = 3;
|
|
usleep(1000);
|
|
igt_assert(*sema == 4);
|
|
|
|
*sema = 5;
|
|
usleep(1000);
|
|
igt_assert(*sema == 6);
|
|
|
|
*sema = 8;
|
|
usleep(1000);
|
|
igt_assert(*sema == 9);
|
|
|
|
nouveau_bo_ref(NULL, &sema_bo);
|
|
drm_intel_bo_unreference(test_intel_bo);
|
|
}
|
|
|
|
igt_main
|
|
{
|
|
igt_fixture {
|
|
find_and_open_devices();
|
|
|
|
igt_require(nouveau_fd != -1);
|
|
igt_require(intel_fd != -1);
|
|
|
|
/* set up intel bufmgr */
|
|
bufmgr = drm_intel_bufmgr_gem_init(intel_fd, 4096);
|
|
igt_assert(bufmgr);
|
|
/* Do not enable reuse, we share (almost) all buffers. */
|
|
//drm_intel_bufmgr_gem_enable_reuse(bufmgr);
|
|
|
|
/* set up nouveau bufmgr */
|
|
init_nouveau();
|
|
|
|
/* set up an intel batch buffer */
|
|
devid = intel_get_drm_devid(intel_fd);
|
|
batch = intel_batchbuffer_alloc(bufmgr, devid);
|
|
igt_assert(batch);
|
|
}
|
|
|
|
#define xtest(x, args...) \
|
|
igt_subtest( #x ) \
|
|
(x)(args);
|
|
|
|
xtest(test1_macro);
|
|
xtest(test1_micro);
|
|
//xtest(test1_swizzle);
|
|
xtest(test2);
|
|
xtest(test3_1);
|
|
xtest(test3_2);
|
|
xtest(test3_3);
|
|
xtest(test3_4);
|
|
xtest(test3_5);
|
|
xtest(test_semaphore);
|
|
|
|
igt_fixture {
|
|
nouveau_bo_ref(NULL, &query_bo);
|
|
nouveau_object_del(&pcopy);
|
|
nouveau_bufctx_del(&nbufctx);
|
|
nouveau_pushbuf_del(&npush);
|
|
nouveau_object_del(&nchannel);
|
|
|
|
intel_batchbuffer_free(batch);
|
|
|
|
nouveau_client_del(&nclient);
|
|
nouveau_device_del(&ndev);
|
|
drm_intel_bufmgr_destroy(bufmgr);
|
|
|
|
close(intel_fd);
|
|
close(nouveau_fd);
|
|
}
|
|
}
|