613 lines
15 KiB
C
613 lines
15 KiB
C
// SPDX-License-Identifier: GPL-2.0 AND MIT
|
|
/*
|
|
* Copyright © 2022 Intel Corporation
|
|
*/
|
|
|
|
#include <kunit/test.h>
|
|
#include <kunit/visibility.h>
|
|
|
|
#include <linux/iosys-map.h>
|
|
#include <linux/math64.h>
|
|
#include <linux/prandom.h>
|
|
#include <linux/swap.h>
|
|
|
|
#include <uapi/linux/sysinfo.h>
|
|
|
|
#include "tests/xe_kunit_helpers.h"
|
|
#include "tests/xe_pci_test.h"
|
|
#include "tests/xe_test.h"
|
|
|
|
#include "xe_bo_evict.h"
|
|
#include "xe_pci.h"
|
|
#include "xe_pm.h"
|
|
|
|
static int ccs_test_migrate(struct xe_tile *tile, struct xe_bo *bo,
|
|
bool clear, u64 get_val, u64 assign_val,
|
|
struct kunit *test)
|
|
{
|
|
struct dma_fence *fence;
|
|
struct ttm_tt *ttm;
|
|
struct page *page;
|
|
pgoff_t ccs_page;
|
|
long timeout;
|
|
u64 *cpu_map;
|
|
int ret;
|
|
u32 offset;
|
|
|
|
/* Move bo to VRAM if not already there. */
|
|
ret = xe_bo_validate(bo, NULL, false);
|
|
if (ret) {
|
|
KUNIT_FAIL(test, "Failed to validate bo.\n");
|
|
return ret;
|
|
}
|
|
|
|
/* Optionally clear bo *and* CCS data in VRAM. */
|
|
if (clear) {
|
|
fence = xe_migrate_clear(tile->migrate, bo, bo->ttm.resource,
|
|
XE_MIGRATE_CLEAR_FLAG_FULL);
|
|
if (IS_ERR(fence)) {
|
|
KUNIT_FAIL(test, "Failed to submit bo clear.\n");
|
|
return PTR_ERR(fence);
|
|
}
|
|
dma_fence_put(fence);
|
|
}
|
|
|
|
/* Evict to system. CCS data should be copied. */
|
|
ret = xe_bo_evict(bo, true);
|
|
if (ret) {
|
|
KUNIT_FAIL(test, "Failed to evict bo.\n");
|
|
return ret;
|
|
}
|
|
|
|
/* Sync all migration blits */
|
|
timeout = dma_resv_wait_timeout(bo->ttm.base.resv,
|
|
DMA_RESV_USAGE_KERNEL,
|
|
true,
|
|
5 * HZ);
|
|
if (timeout <= 0) {
|
|
KUNIT_FAIL(test, "Failed to sync bo eviction.\n");
|
|
return -ETIME;
|
|
}
|
|
|
|
/*
|
|
* Bo with CCS data is now in system memory. Verify backing store
|
|
* and data integrity. Then assign for the next testing round while
|
|
* we still have a CPU map.
|
|
*/
|
|
ttm = bo->ttm.ttm;
|
|
if (!ttm || !ttm_tt_is_populated(ttm)) {
|
|
KUNIT_FAIL(test, "Bo was not in expected placement.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
ccs_page = xe_bo_ccs_pages_start(bo) >> PAGE_SHIFT;
|
|
if (ccs_page >= ttm->num_pages) {
|
|
KUNIT_FAIL(test, "No TTM CCS pages present.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
page = ttm->pages[ccs_page];
|
|
cpu_map = kmap_local_page(page);
|
|
|
|
/* Check first CCS value */
|
|
if (cpu_map[0] != get_val) {
|
|
KUNIT_FAIL(test,
|
|
"Expected CCS readout 0x%016llx, got 0x%016llx.\n",
|
|
(unsigned long long)get_val,
|
|
(unsigned long long)cpu_map[0]);
|
|
ret = -EINVAL;
|
|
}
|
|
|
|
/* Check last CCS value, or at least last value in page. */
|
|
offset = xe_device_ccs_bytes(tile_to_xe(tile), bo->size);
|
|
offset = min_t(u32, offset, PAGE_SIZE) / sizeof(u64) - 1;
|
|
if (cpu_map[offset] != get_val) {
|
|
KUNIT_FAIL(test,
|
|
"Expected CCS readout 0x%016llx, got 0x%016llx.\n",
|
|
(unsigned long long)get_val,
|
|
(unsigned long long)cpu_map[offset]);
|
|
ret = -EINVAL;
|
|
}
|
|
|
|
cpu_map[0] = assign_val;
|
|
cpu_map[offset] = assign_val;
|
|
kunmap_local(cpu_map);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void ccs_test_run_tile(struct xe_device *xe, struct xe_tile *tile,
|
|
struct kunit *test)
|
|
{
|
|
struct xe_bo *bo;
|
|
|
|
int ret;
|
|
|
|
/* TODO: Sanity check */
|
|
unsigned int bo_flags = XE_BO_FLAG_VRAM_IF_DGFX(tile);
|
|
|
|
if (IS_DGFX(xe))
|
|
kunit_info(test, "Testing vram id %u\n", tile->id);
|
|
else
|
|
kunit_info(test, "Testing system memory\n");
|
|
|
|
bo = xe_bo_create_user(xe, NULL, NULL, SZ_1M, DRM_XE_GEM_CPU_CACHING_WC,
|
|
bo_flags);
|
|
if (IS_ERR(bo)) {
|
|
KUNIT_FAIL(test, "Failed to create bo.\n");
|
|
return;
|
|
}
|
|
|
|
xe_bo_lock(bo, false);
|
|
|
|
kunit_info(test, "Verifying that CCS data is cleared on creation.\n");
|
|
ret = ccs_test_migrate(tile, bo, false, 0ULL, 0xdeadbeefdeadbeefULL,
|
|
test);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
kunit_info(test, "Verifying that CCS data survives migration.\n");
|
|
ret = ccs_test_migrate(tile, bo, false, 0xdeadbeefdeadbeefULL,
|
|
0xdeadbeefdeadbeefULL, test);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
kunit_info(test, "Verifying that CCS data can be properly cleared.\n");
|
|
ret = ccs_test_migrate(tile, bo, true, 0ULL, 0ULL, test);
|
|
|
|
out_unlock:
|
|
xe_bo_unlock(bo);
|
|
xe_bo_put(bo);
|
|
}
|
|
|
|
static int ccs_test_run_device(struct xe_device *xe)
|
|
{
|
|
struct kunit *test = kunit_get_current_test();
|
|
struct xe_tile *tile;
|
|
int id;
|
|
|
|
if (!xe_device_has_flat_ccs(xe)) {
|
|
kunit_skip(test, "non-flat-ccs device\n");
|
|
return 0;
|
|
}
|
|
|
|
/* For xe2+ dgfx, we don't handle ccs metadata */
|
|
if (GRAPHICS_VER(xe) >= 20 && IS_DGFX(xe)) {
|
|
kunit_skip(test, "xe2+ dgfx device\n");
|
|
return 0;
|
|
}
|
|
|
|
xe_pm_runtime_get(xe);
|
|
|
|
for_each_tile(tile, xe, id) {
|
|
/* For igfx run only for primary tile */
|
|
if (!IS_DGFX(xe) && id > 0)
|
|
continue;
|
|
ccs_test_run_tile(xe, tile, test);
|
|
}
|
|
|
|
xe_pm_runtime_put(xe);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void xe_ccs_migrate_kunit(struct kunit *test)
|
|
{
|
|
struct xe_device *xe = test->priv;
|
|
|
|
ccs_test_run_device(xe);
|
|
}
|
|
|
|
static int evict_test_run_tile(struct xe_device *xe, struct xe_tile *tile, struct kunit *test)
|
|
{
|
|
struct xe_bo *bo, *external;
|
|
unsigned int bo_flags = XE_BO_FLAG_VRAM_IF_DGFX(tile);
|
|
struct xe_vm *vm = xe_migrate_get_vm(xe_device_get_root_tile(xe)->migrate);
|
|
struct xe_gt *__gt;
|
|
int err, i, id;
|
|
|
|
kunit_info(test, "Testing device %s vram id %u\n",
|
|
dev_name(xe->drm.dev), tile->id);
|
|
|
|
for (i = 0; i < 2; ++i) {
|
|
xe_vm_lock(vm, false);
|
|
bo = xe_bo_create_user(xe, NULL, vm, 0x10000,
|
|
DRM_XE_GEM_CPU_CACHING_WC,
|
|
bo_flags);
|
|
xe_vm_unlock(vm);
|
|
if (IS_ERR(bo)) {
|
|
KUNIT_FAIL(test, "bo create err=%pe\n", bo);
|
|
break;
|
|
}
|
|
|
|
external = xe_bo_create_user(xe, NULL, NULL, 0x10000,
|
|
DRM_XE_GEM_CPU_CACHING_WC,
|
|
bo_flags);
|
|
if (IS_ERR(external)) {
|
|
KUNIT_FAIL(test, "external bo create err=%pe\n", external);
|
|
goto cleanup_bo;
|
|
}
|
|
|
|
xe_bo_lock(external, false);
|
|
err = xe_bo_pin_external(external);
|
|
xe_bo_unlock(external);
|
|
if (err) {
|
|
KUNIT_FAIL(test, "external bo pin err=%pe\n",
|
|
ERR_PTR(err));
|
|
goto cleanup_external;
|
|
}
|
|
|
|
err = xe_bo_evict_all(xe);
|
|
if (err) {
|
|
KUNIT_FAIL(test, "evict err=%pe\n", ERR_PTR(err));
|
|
goto cleanup_all;
|
|
}
|
|
|
|
for_each_gt(__gt, xe, id)
|
|
xe_gt_sanitize(__gt);
|
|
err = xe_bo_restore_kernel(xe);
|
|
/*
|
|
* Snapshotting the CTB and copying back a potentially old
|
|
* version seems risky, depending on what might have been
|
|
* inflight. Also it seems snapshotting the ADS object and
|
|
* copying back results in serious breakage. Normally when
|
|
* calling xe_bo_restore_kernel() we always fully restart the
|
|
* GT, which re-intializes such things. We could potentially
|
|
* skip saving and restoring such objects in xe_bo_evict_all()
|
|
* however seems quite fragile not to also restart the GT. Try
|
|
* to do that here by triggering a GT reset.
|
|
*/
|
|
for_each_gt(__gt, xe, id)
|
|
xe_gt_reset(__gt);
|
|
|
|
if (err) {
|
|
KUNIT_FAIL(test, "restore kernel err=%pe\n",
|
|
ERR_PTR(err));
|
|
goto cleanup_all;
|
|
}
|
|
|
|
err = xe_bo_restore_user(xe);
|
|
if (err) {
|
|
KUNIT_FAIL(test, "restore user err=%pe\n", ERR_PTR(err));
|
|
goto cleanup_all;
|
|
}
|
|
|
|
if (!xe_bo_is_vram(external)) {
|
|
KUNIT_FAIL(test, "external bo is not vram\n");
|
|
err = -EPROTO;
|
|
goto cleanup_all;
|
|
}
|
|
|
|
if (xe_bo_is_vram(bo)) {
|
|
KUNIT_FAIL(test, "bo is vram\n");
|
|
err = -EPROTO;
|
|
goto cleanup_all;
|
|
}
|
|
|
|
if (i) {
|
|
down_read(&vm->lock);
|
|
xe_vm_lock(vm, false);
|
|
err = xe_bo_validate(bo, bo->vm, false);
|
|
xe_vm_unlock(vm);
|
|
up_read(&vm->lock);
|
|
if (err) {
|
|
KUNIT_FAIL(test, "bo valid err=%pe\n",
|
|
ERR_PTR(err));
|
|
goto cleanup_all;
|
|
}
|
|
xe_bo_lock(external, false);
|
|
err = xe_bo_validate(external, NULL, false);
|
|
xe_bo_unlock(external);
|
|
if (err) {
|
|
KUNIT_FAIL(test, "external bo valid err=%pe\n",
|
|
ERR_PTR(err));
|
|
goto cleanup_all;
|
|
}
|
|
}
|
|
|
|
xe_bo_lock(external, false);
|
|
xe_bo_unpin_external(external);
|
|
xe_bo_unlock(external);
|
|
|
|
xe_bo_put(external);
|
|
|
|
xe_bo_lock(bo, false);
|
|
__xe_bo_unset_bulk_move(bo);
|
|
xe_bo_unlock(bo);
|
|
xe_bo_put(bo);
|
|
continue;
|
|
|
|
cleanup_all:
|
|
xe_bo_lock(external, false);
|
|
xe_bo_unpin_external(external);
|
|
xe_bo_unlock(external);
|
|
cleanup_external:
|
|
xe_bo_put(external);
|
|
cleanup_bo:
|
|
xe_bo_lock(bo, false);
|
|
__xe_bo_unset_bulk_move(bo);
|
|
xe_bo_unlock(bo);
|
|
xe_bo_put(bo);
|
|
break;
|
|
}
|
|
|
|
xe_vm_put(vm);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int evict_test_run_device(struct xe_device *xe)
|
|
{
|
|
struct kunit *test = kunit_get_current_test();
|
|
struct xe_tile *tile;
|
|
int id;
|
|
|
|
if (!IS_DGFX(xe)) {
|
|
kunit_skip(test, "non-discrete device\n");
|
|
return 0;
|
|
}
|
|
|
|
xe_pm_runtime_get(xe);
|
|
|
|
for_each_tile(tile, xe, id)
|
|
evict_test_run_tile(xe, tile, test);
|
|
|
|
xe_pm_runtime_put(xe);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void xe_bo_evict_kunit(struct kunit *test)
|
|
{
|
|
struct xe_device *xe = test->priv;
|
|
|
|
evict_test_run_device(xe);
|
|
}
|
|
|
|
struct xe_bo_link {
|
|
struct list_head link;
|
|
struct xe_bo *bo;
|
|
u32 val;
|
|
};
|
|
|
|
#define XE_BO_SHRINK_SIZE ((unsigned long)SZ_64M)
|
|
|
|
static int shrink_test_fill_random(struct xe_bo *bo, struct rnd_state *state,
|
|
struct xe_bo_link *link)
|
|
{
|
|
struct iosys_map map;
|
|
int ret = ttm_bo_vmap(&bo->ttm, &map);
|
|
size_t __maybe_unused i;
|
|
|
|
if (ret)
|
|
return ret;
|
|
|
|
for (i = 0; i < bo->ttm.base.size; i += sizeof(u32)) {
|
|
u32 val = prandom_u32_state(state);
|
|
|
|
iosys_map_wr(&map, i, u32, val);
|
|
if (i == 0)
|
|
link->val = val;
|
|
}
|
|
|
|
ttm_bo_vunmap(&bo->ttm, &map);
|
|
return 0;
|
|
}
|
|
|
|
static bool shrink_test_verify(struct kunit *test, struct xe_bo *bo,
|
|
unsigned int bo_nr, struct rnd_state *state,
|
|
struct xe_bo_link *link)
|
|
{
|
|
struct iosys_map map;
|
|
int ret = ttm_bo_vmap(&bo->ttm, &map);
|
|
size_t i;
|
|
bool failed = false;
|
|
|
|
if (ret) {
|
|
KUNIT_FAIL(test, "Error mapping bo %u for content check.\n", bo_nr);
|
|
return true;
|
|
}
|
|
|
|
for (i = 0; i < bo->ttm.base.size; i += sizeof(u32)) {
|
|
u32 val = prandom_u32_state(state);
|
|
|
|
if (iosys_map_rd(&map, i, u32) != val) {
|
|
KUNIT_FAIL(test, "Content not preserved, bo %u offset 0x%016llx",
|
|
bo_nr, (unsigned long long)i);
|
|
kunit_info(test, "Failed value is 0x%08x, recorded 0x%08x\n",
|
|
(unsigned int)iosys_map_rd(&map, i, u32), val);
|
|
if (i == 0 && val != link->val)
|
|
kunit_info(test, "Looks like PRNG is out of sync.\n");
|
|
failed = true;
|
|
break;
|
|
}
|
|
}
|
|
|
|
ttm_bo_vunmap(&bo->ttm, &map);
|
|
|
|
return failed;
|
|
}
|
|
|
|
/*
|
|
* Try to create system bos corresponding to twice the amount
|
|
* of available system memory to test shrinker functionality.
|
|
* If no swap space is available to accommodate the
|
|
* memory overcommit, mark bos purgeable.
|
|
*/
|
|
static int shrink_test_run_device(struct xe_device *xe)
|
|
{
|
|
struct kunit *test = kunit_get_current_test();
|
|
LIST_HEAD(bos);
|
|
struct xe_bo_link *link, *next;
|
|
struct sysinfo si;
|
|
u64 ram, ram_and_swap, purgeable = 0, alloced, to_alloc, limit;
|
|
unsigned int interrupted = 0, successful = 0, count = 0;
|
|
struct rnd_state prng;
|
|
u64 rand_seed;
|
|
bool failed = false;
|
|
|
|
rand_seed = get_random_u64();
|
|
prandom_seed_state(&prng, rand_seed);
|
|
kunit_info(test, "Random seed is 0x%016llx.\n",
|
|
(unsigned long long)rand_seed);
|
|
|
|
/* Skip if execution time is expected to be too long. */
|
|
|
|
limit = SZ_32G;
|
|
/* IGFX with flat CCS needs to copy when swapping / shrinking */
|
|
if (!IS_DGFX(xe) && xe_device_has_flat_ccs(xe))
|
|
limit = SZ_16G;
|
|
|
|
si_meminfo(&si);
|
|
ram = (size_t)si.freeram * si.mem_unit;
|
|
if (ram > limit) {
|
|
kunit_skip(test, "Too long expected execution time.\n");
|
|
return 0;
|
|
}
|
|
to_alloc = ram * 2;
|
|
|
|
ram_and_swap = ram + get_nr_swap_pages() * PAGE_SIZE;
|
|
if (to_alloc > ram_and_swap)
|
|
purgeable = to_alloc - ram_and_swap;
|
|
purgeable += div64_u64(purgeable, 5);
|
|
|
|
kunit_info(test, "Free ram is %lu bytes. Will allocate twice of that.\n",
|
|
(unsigned long)ram);
|
|
for (alloced = 0; alloced < to_alloc; alloced += XE_BO_SHRINK_SIZE) {
|
|
struct xe_bo *bo;
|
|
unsigned int mem_type;
|
|
struct xe_ttm_tt *xe_tt;
|
|
|
|
link = kzalloc(sizeof(*link), GFP_KERNEL);
|
|
if (!link) {
|
|
KUNIT_FAIL(test, "Unexpected link allocation failure\n");
|
|
failed = true;
|
|
break;
|
|
}
|
|
|
|
INIT_LIST_HEAD(&link->link);
|
|
|
|
/* We can create bos using WC caching here. But it is slower. */
|
|
bo = xe_bo_create_user(xe, NULL, NULL, XE_BO_SHRINK_SIZE,
|
|
DRM_XE_GEM_CPU_CACHING_WB,
|
|
XE_BO_FLAG_SYSTEM);
|
|
if (IS_ERR(bo)) {
|
|
if (bo != ERR_PTR(-ENOMEM) && bo != ERR_PTR(-ENOSPC) &&
|
|
bo != ERR_PTR(-EINTR) && bo != ERR_PTR(-ERESTARTSYS))
|
|
KUNIT_FAIL(test, "Error creating bo: %pe\n", bo);
|
|
kfree(link);
|
|
failed = true;
|
|
break;
|
|
}
|
|
xe_bo_lock(bo, false);
|
|
xe_tt = container_of(bo->ttm.ttm, typeof(*xe_tt), ttm);
|
|
|
|
/*
|
|
* Allocate purgeable bos first, because if we do it the
|
|
* other way around, they may not be subject to swapping...
|
|
*/
|
|
if (alloced < purgeable) {
|
|
xe_tt->purgeable = true;
|
|
bo->ttm.priority = 0;
|
|
} else {
|
|
int ret = shrink_test_fill_random(bo, &prng, link);
|
|
|
|
if (ret) {
|
|
xe_bo_unlock(bo);
|
|
xe_bo_put(bo);
|
|
KUNIT_FAIL(test, "Error filling bo with random data: %pe\n",
|
|
ERR_PTR(ret));
|
|
kfree(link);
|
|
failed = true;
|
|
break;
|
|
}
|
|
}
|
|
|
|
mem_type = bo->ttm.resource->mem_type;
|
|
xe_bo_unlock(bo);
|
|
link->bo = bo;
|
|
list_add_tail(&link->link, &bos);
|
|
|
|
if (mem_type != XE_PL_TT) {
|
|
KUNIT_FAIL(test, "Bo in incorrect memory type: %u\n",
|
|
bo->ttm.resource->mem_type);
|
|
failed = true;
|
|
}
|
|
cond_resched();
|
|
if (signal_pending(current))
|
|
break;
|
|
}
|
|
|
|
/*
|
|
* Read back and destroy bos. Reset the pseudo-random seed to get an
|
|
* identical pseudo-random number sequence for readback.
|
|
*/
|
|
prandom_seed_state(&prng, rand_seed);
|
|
list_for_each_entry_safe(link, next, &bos, link) {
|
|
static struct ttm_operation_ctx ctx = {.interruptible = true};
|
|
struct xe_bo *bo = link->bo;
|
|
struct xe_ttm_tt *xe_tt;
|
|
int ret;
|
|
|
|
count++;
|
|
if (!signal_pending(current) && !failed) {
|
|
bool purgeable, intr = false;
|
|
|
|
xe_bo_lock(bo, NULL);
|
|
|
|
/* xe_tt->purgeable is cleared on validate. */
|
|
xe_tt = container_of(bo->ttm.ttm, typeof(*xe_tt), ttm);
|
|
purgeable = xe_tt->purgeable;
|
|
do {
|
|
ret = ttm_bo_validate(&bo->ttm, &tt_placement, &ctx);
|
|
if (ret == -EINTR)
|
|
intr = true;
|
|
} while (ret == -EINTR && !signal_pending(current));
|
|
|
|
if (!ret && !purgeable)
|
|
failed = shrink_test_verify(test, bo, count, &prng, link);
|
|
|
|
xe_bo_unlock(bo);
|
|
if (ret) {
|
|
KUNIT_FAIL(test, "Validation failed: %pe\n",
|
|
ERR_PTR(ret));
|
|
failed = true;
|
|
} else if (intr) {
|
|
interrupted++;
|
|
} else {
|
|
successful++;
|
|
}
|
|
}
|
|
xe_bo_put(link->bo);
|
|
list_del(&link->link);
|
|
kfree(link);
|
|
}
|
|
kunit_info(test, "Readbacks interrupted: %u successful: %u\n",
|
|
interrupted, successful);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void xe_bo_shrink_kunit(struct kunit *test)
|
|
{
|
|
struct xe_device *xe = test->priv;
|
|
|
|
shrink_test_run_device(xe);
|
|
}
|
|
|
|
static struct kunit_case xe_bo_tests[] = {
|
|
KUNIT_CASE_PARAM(xe_ccs_migrate_kunit, xe_pci_live_device_gen_param),
|
|
KUNIT_CASE_PARAM(xe_bo_evict_kunit, xe_pci_live_device_gen_param),
|
|
KUNIT_CASE_PARAM_ATTR(xe_bo_shrink_kunit, xe_pci_live_device_gen_param,
|
|
{.speed = KUNIT_SPEED_SLOW}),
|
|
{}
|
|
};
|
|
|
|
VISIBLE_IF_KUNIT
|
|
struct kunit_suite xe_bo_test_suite = {
|
|
.name = "xe_bo",
|
|
.test_cases = xe_bo_tests,
|
|
.init = xe_kunit_helper_xe_device_live_test_init,
|
|
};
|
|
EXPORT_SYMBOL_IF_KUNIT(xe_bo_test_suite);
|