203 lines
5.2 KiB
C
203 lines
5.2 KiB
C
// SPDX-License-Identifier: MIT
|
|
/*
|
|
* Copyright © 2023 Intel Corporation
|
|
*/
|
|
|
|
#include <drm/drm_managed.h>
|
|
|
|
#include "regs/xe_gt_regs.h"
|
|
#include "xe_assert.h"
|
|
#include "xe_gt.h"
|
|
#include "xe_gt_ccs_mode.h"
|
|
#include "xe_gt_printk.h"
|
|
#include "xe_gt_sysfs.h"
|
|
#include "xe_mmio.h"
|
|
#include "xe_sriov.h"
|
|
|
|
static void __xe_gt_apply_ccs_mode(struct xe_gt *gt, u32 num_engines)
|
|
{
|
|
u32 mode = CCS_MODE_CSLICE_0_3_MASK; /* disable all by default */
|
|
int num_slices = hweight32(CCS_MASK(gt));
|
|
struct xe_device *xe = gt_to_xe(gt);
|
|
int width, cslice = 0;
|
|
u32 config = 0;
|
|
|
|
xe_assert(xe, xe_gt_ccs_mode_enabled(gt));
|
|
|
|
xe_assert(xe, num_engines && num_engines <= num_slices);
|
|
xe_assert(xe, !(num_slices % num_engines));
|
|
|
|
/*
|
|
* Loop over all available slices and assign each a user engine.
|
|
* For example, if there are four compute slices available, the
|
|
* assignment of compute slices to compute engines would be,
|
|
*
|
|
* With 1 engine (ccs0):
|
|
* slice 0, 1, 2, 3: ccs0
|
|
*
|
|
* With 2 engines (ccs0, ccs1):
|
|
* slice 0, 2: ccs0
|
|
* slice 1, 3: ccs1
|
|
*
|
|
* With 4 engines (ccs0, ccs1, ccs2, ccs3):
|
|
* slice 0: ccs0
|
|
* slice 1: ccs1
|
|
* slice 2: ccs2
|
|
* slice 3: ccs3
|
|
*/
|
|
for (width = num_slices / num_engines; width; width--) {
|
|
struct xe_hw_engine *hwe;
|
|
enum xe_hw_engine_id id;
|
|
|
|
for_each_hw_engine(hwe, gt, id) {
|
|
if (hwe->class != XE_ENGINE_CLASS_COMPUTE)
|
|
continue;
|
|
|
|
if (hwe->logical_instance >= num_engines)
|
|
break;
|
|
|
|
config |= BIT(hwe->instance) << XE_HW_ENGINE_CCS0;
|
|
|
|
/* If a slice is fused off, leave disabled */
|
|
while ((CCS_MASK(gt) & BIT(cslice)) == 0)
|
|
cslice++;
|
|
|
|
mode &= ~CCS_MODE_CSLICE(cslice, CCS_MODE_CSLICE_MASK);
|
|
mode |= CCS_MODE_CSLICE(cslice, hwe->instance);
|
|
cslice++;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Mask bits need to be set for the register. Though only Xe2+
|
|
* platforms require setting of mask bits, it won't harm for older
|
|
* platforms as these bits are unused there.
|
|
*/
|
|
mode |= CCS_MODE_CSLICE_0_3_MASK << 16;
|
|
xe_mmio_write32(>->mmio, CCS_MODE, mode);
|
|
|
|
xe_gt_dbg(gt, "CCS_MODE=%x config:%08x, num_engines:%d, num_slices:%d\n",
|
|
mode, config, num_engines, num_slices);
|
|
}
|
|
|
|
void xe_gt_apply_ccs_mode(struct xe_gt *gt)
|
|
{
|
|
if (!gt->ccs_mode || IS_SRIOV_VF(gt_to_xe(gt)))
|
|
return;
|
|
|
|
__xe_gt_apply_ccs_mode(gt, gt->ccs_mode);
|
|
}
|
|
|
|
static ssize_t
|
|
num_cslices_show(struct device *kdev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct xe_gt *gt = kobj_to_gt(&kdev->kobj);
|
|
|
|
return sysfs_emit(buf, "%u\n", hweight32(CCS_MASK(gt)));
|
|
}
|
|
|
|
static DEVICE_ATTR_RO(num_cslices);
|
|
|
|
static ssize_t
|
|
ccs_mode_show(struct device *kdev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct xe_gt *gt = kobj_to_gt(&kdev->kobj);
|
|
|
|
return sysfs_emit(buf, "%u\n", gt->ccs_mode);
|
|
}
|
|
|
|
static ssize_t
|
|
ccs_mode_store(struct device *kdev, struct device_attribute *attr,
|
|
const char *buff, size_t count)
|
|
{
|
|
struct xe_gt *gt = kobj_to_gt(&kdev->kobj);
|
|
struct xe_device *xe = gt_to_xe(gt);
|
|
u32 num_engines, num_slices;
|
|
int ret;
|
|
|
|
if (IS_SRIOV(xe)) {
|
|
xe_gt_dbg(gt, "Can't change compute mode when running as %s\n",
|
|
xe_sriov_mode_to_string(xe_device_sriov_mode(xe)));
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
ret = kstrtou32(buff, 0, &num_engines);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/*
|
|
* Ensure number of engines specified is valid and there is an
|
|
* exact multiple of engines for slices.
|
|
*/
|
|
num_slices = hweight32(CCS_MASK(gt));
|
|
if (!num_engines || num_engines > num_slices || num_slices % num_engines) {
|
|
xe_gt_dbg(gt, "Invalid compute config, %d engines %d slices\n",
|
|
num_engines, num_slices);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* CCS mode can only be updated when there are no drm clients */
|
|
mutex_lock(&xe->drm.filelist_mutex);
|
|
if (!list_empty(&xe->drm.filelist)) {
|
|
mutex_unlock(&xe->drm.filelist_mutex);
|
|
xe_gt_dbg(gt, "Rejecting compute mode change as there are active drm clients\n");
|
|
return -EBUSY;
|
|
}
|
|
|
|
if (gt->ccs_mode != num_engines) {
|
|
xe_gt_info(gt, "Setting compute mode to %d\n", num_engines);
|
|
gt->ccs_mode = num_engines;
|
|
xe_gt_record_user_engines(gt);
|
|
xe_gt_reset(gt);
|
|
}
|
|
|
|
mutex_unlock(&xe->drm.filelist_mutex);
|
|
|
|
return count;
|
|
}
|
|
|
|
static DEVICE_ATTR_RW(ccs_mode);
|
|
|
|
static const struct attribute *gt_ccs_mode_attrs[] = {
|
|
&dev_attr_ccs_mode.attr,
|
|
&dev_attr_num_cslices.attr,
|
|
NULL,
|
|
};
|
|
|
|
static void xe_gt_ccs_mode_sysfs_fini(void *arg)
|
|
{
|
|
struct xe_gt *gt = arg;
|
|
|
|
sysfs_remove_files(gt->sysfs, gt_ccs_mode_attrs);
|
|
}
|
|
|
|
/**
|
|
* xe_gt_ccs_mode_sysfs_init - Initialize CCS mode sysfs interfaces
|
|
* @gt: GT structure
|
|
*
|
|
* Through a per-gt 'ccs_mode' sysfs interface, the user can enable a fixed
|
|
* number of compute hardware engines to which the available compute slices
|
|
* are to be allocated. This user configuration change triggers a gt reset
|
|
* and it is expected that there are no open drm clients while doing so.
|
|
* The number of available compute slices is exposed to user through a per-gt
|
|
* 'num_cslices' sysfs interface.
|
|
*
|
|
* Returns: Returns error value for failure and 0 for success.
|
|
*/
|
|
int xe_gt_ccs_mode_sysfs_init(struct xe_gt *gt)
|
|
{
|
|
struct xe_device *xe = gt_to_xe(gt);
|
|
int err;
|
|
|
|
if (!xe_gt_ccs_mode_enabled(gt))
|
|
return 0;
|
|
|
|
err = sysfs_create_files(gt->sysfs, gt_ccs_mode_attrs);
|
|
if (err)
|
|
return err;
|
|
|
|
return devm_add_action_or_reset(xe->drm.dev, xe_gt_ccs_mode_sysfs_fini, gt);
|
|
}
|