dmaengine: idxd: add interrupt handle request and release support
DSA spec states that when Request Interrupt Handle and Release Interrupt Handle command bits are set in the CMDCAP register, these device commands must be supported by the driver. The interrupt handle is programmed in a descriptor. When Request Interrupt Handle is not supported, the interrupt handle is the index of the desired entry in the MSI-X table. When the command is supported, driver must use the command to obtain a handle to be programmed in the submitted descriptor. A requested handle may be revoked. After the handle is revoked, any use of the handle will result in Invalid Interrupt Handle error. Signed-off-by: Dave Jiang <dave.jiang@intel.com> Link: https://lore.kernel.org/r/161894439422.3202472.17579543737810265471.stgit@djiang5-desk3.ch.intel.com Signed-off-by: Vinod Koul <vkoul@kernel.org>
This commit is contained in:
parent
8c66bbdc4f
commit
eb15e7154f
@ -598,6 +598,77 @@ void idxd_device_drain_pasid(struct idxd_device *idxd, int pasid)
|
||||
dev_dbg(dev, "pasid %d drained\n", pasid);
|
||||
}
|
||||
|
||||
int idxd_device_request_int_handle(struct idxd_device *idxd, int idx, int *handle,
|
||||
enum idxd_interrupt_type irq_type)
|
||||
{
|
||||
struct device *dev = &idxd->pdev->dev;
|
||||
u32 operand, status;
|
||||
|
||||
if (!(idxd->hw.cmd_cap & BIT(IDXD_CMD_REQUEST_INT_HANDLE)))
|
||||
return -EOPNOTSUPP;
|
||||
|
||||
dev_dbg(dev, "get int handle, idx %d\n", idx);
|
||||
|
||||
operand = idx & GENMASK(15, 0);
|
||||
if (irq_type == IDXD_IRQ_IMS)
|
||||
operand |= CMD_INT_HANDLE_IMS;
|
||||
|
||||
dev_dbg(dev, "cmd: %u operand: %#x\n", IDXD_CMD_REQUEST_INT_HANDLE, operand);
|
||||
|
||||
idxd_cmd_exec(idxd, IDXD_CMD_REQUEST_INT_HANDLE, operand, &status);
|
||||
|
||||
if ((status & IDXD_CMDSTS_ERR_MASK) != IDXD_CMDSTS_SUCCESS) {
|
||||
dev_dbg(dev, "request int handle failed: %#x\n", status);
|
||||
return -ENXIO;
|
||||
}
|
||||
|
||||
*handle = (status >> IDXD_CMDSTS_RES_SHIFT) & GENMASK(15, 0);
|
||||
|
||||
dev_dbg(dev, "int handle acquired: %u\n", *handle);
|
||||
return 0;
|
||||
}
|
||||
|
||||
int idxd_device_release_int_handle(struct idxd_device *idxd, int handle,
|
||||
enum idxd_interrupt_type irq_type)
|
||||
{
|
||||
struct device *dev = &idxd->pdev->dev;
|
||||
u32 operand, status;
|
||||
union idxd_command_reg cmd;
|
||||
unsigned long flags;
|
||||
|
||||
if (!(idxd->hw.cmd_cap & BIT(IDXD_CMD_RELEASE_INT_HANDLE)))
|
||||
return -EOPNOTSUPP;
|
||||
|
||||
dev_dbg(dev, "release int handle, handle %d\n", handle);
|
||||
|
||||
memset(&cmd, 0, sizeof(cmd));
|
||||
operand = handle & GENMASK(15, 0);
|
||||
|
||||
if (irq_type == IDXD_IRQ_IMS)
|
||||
operand |= CMD_INT_HANDLE_IMS;
|
||||
|
||||
cmd.cmd = IDXD_CMD_RELEASE_INT_HANDLE;
|
||||
cmd.operand = operand;
|
||||
|
||||
dev_dbg(dev, "cmd: %u operand: %#x\n", IDXD_CMD_RELEASE_INT_HANDLE, operand);
|
||||
|
||||
spin_lock_irqsave(&idxd->dev_lock, flags);
|
||||
iowrite32(cmd.bits, idxd->reg_base + IDXD_CMD_OFFSET);
|
||||
|
||||
while (ioread32(idxd->reg_base + IDXD_CMDSTS_OFFSET) & IDXD_CMDSTS_ACTIVE)
|
||||
cpu_relax();
|
||||
status = ioread32(idxd->reg_base + IDXD_CMDSTS_OFFSET);
|
||||
spin_unlock_irqrestore(&idxd->dev_lock, flags);
|
||||
|
||||
if ((status & IDXD_CMDSTS_ERR_MASK) != IDXD_CMDSTS_SUCCESS) {
|
||||
dev_dbg(dev, "release int handle failed: %#x\n", status);
|
||||
return -ENXIO;
|
||||
}
|
||||
|
||||
dev_dbg(dev, "int handle released.\n");
|
||||
return 0;
|
||||
}
|
||||
|
||||
/* Device configuration bits */
|
||||
void idxd_msix_perm_setup(struct idxd_device *idxd)
|
||||
{
|
||||
|
@ -160,6 +160,7 @@ struct idxd_hw {
|
||||
union group_cap_reg group_cap;
|
||||
union engine_cap_reg engine_cap;
|
||||
struct opcap opcap;
|
||||
u32 cmd_cap;
|
||||
};
|
||||
|
||||
enum idxd_device_state {
|
||||
@ -237,6 +238,8 @@ struct idxd_device {
|
||||
struct idxd_dma_dev *idxd_dma;
|
||||
struct workqueue_struct *wq;
|
||||
struct work_struct work;
|
||||
|
||||
int *int_handles;
|
||||
};
|
||||
|
||||
/* IDXD software descriptor */
|
||||
@ -256,6 +259,7 @@ struct idxd_desc {
|
||||
struct list_head list;
|
||||
int id;
|
||||
int cpu;
|
||||
unsigned int vector;
|
||||
struct idxd_wq *wq;
|
||||
};
|
||||
|
||||
@ -330,6 +334,11 @@ enum idxd_portal_prot {
|
||||
IDXD_PORTAL_LIMITED,
|
||||
};
|
||||
|
||||
enum idxd_interrupt_type {
|
||||
IDXD_IRQ_MSIX = 0,
|
||||
IDXD_IRQ_IMS,
|
||||
};
|
||||
|
||||
static inline int idxd_get_wq_portal_offset(enum idxd_portal_prot prot)
|
||||
{
|
||||
return prot * 0x1000;
|
||||
@ -385,6 +394,10 @@ int idxd_device_config(struct idxd_device *idxd);
|
||||
void idxd_device_wqs_clear_state(struct idxd_device *idxd);
|
||||
void idxd_device_drain_pasid(struct idxd_device *idxd, int pasid);
|
||||
int idxd_device_load_config(struct idxd_device *idxd);
|
||||
int idxd_device_request_int_handle(struct idxd_device *idxd, int idx, int *handle,
|
||||
enum idxd_interrupt_type irq_type);
|
||||
int idxd_device_release_int_handle(struct idxd_device *idxd, int handle,
|
||||
enum idxd_interrupt_type irq_type);
|
||||
|
||||
/* work queue control */
|
||||
int idxd_wq_alloc_resources(struct idxd_wq *wq);
|
||||
|
@ -125,7 +125,25 @@ static int idxd_setup_interrupts(struct idxd_device *idxd)
|
||||
dev_err(dev, "Failed to allocate irq %d.\n", irq_entry->vector);
|
||||
goto err_wq_irqs;
|
||||
}
|
||||
|
||||
dev_dbg(dev, "Allocated idxd-msix %d for vector %d\n", i, irq_entry->vector);
|
||||
if (idxd->hw.cmd_cap & BIT(IDXD_CMD_REQUEST_INT_HANDLE)) {
|
||||
/*
|
||||
* The MSIX vector enumeration starts at 1 with vector 0 being the
|
||||
* misc interrupt that handles non I/O completion events. The
|
||||
* interrupt handles are for IMS enumeration on guest. The misc
|
||||
* interrupt vector does not require a handle and therefore we start
|
||||
* the int_handles at index 0. Since 'i' starts at 1, the first
|
||||
* int_handles index will be 0.
|
||||
*/
|
||||
rc = idxd_device_request_int_handle(idxd, i, &idxd->int_handles[i - 1],
|
||||
IDXD_IRQ_MSIX);
|
||||
if (rc < 0) {
|
||||
free_irq(irq_entry->vector, irq_entry);
|
||||
goto err_wq_irqs;
|
||||
}
|
||||
dev_dbg(dev, "int handle requested: %u\n", idxd->int_handles[i - 1]);
|
||||
}
|
||||
}
|
||||
|
||||
idxd_unmask_error_interrupts(idxd);
|
||||
@ -136,6 +154,9 @@ static int idxd_setup_interrupts(struct idxd_device *idxd)
|
||||
while (--i >= 0) {
|
||||
irq_entry = &idxd->irq_entries[i];
|
||||
free_irq(irq_entry->vector, irq_entry);
|
||||
if (i != 0)
|
||||
idxd_device_release_int_handle(idxd,
|
||||
idxd->int_handles[i], IDXD_IRQ_MSIX);
|
||||
}
|
||||
err_misc_irq:
|
||||
/* Disable error interrupt generation */
|
||||
@ -288,9 +309,15 @@ static int idxd_setup_internals(struct idxd_device *idxd)
|
||||
|
||||
init_waitqueue_head(&idxd->cmd_waitq);
|
||||
|
||||
if (idxd->hw.cmd_cap & BIT(IDXD_CMD_REQUEST_INT_HANDLE)) {
|
||||
idxd->int_handles = devm_kcalloc(dev, idxd->max_wqs, sizeof(int), GFP_KERNEL);
|
||||
if (!idxd->int_handles)
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
rc = idxd_setup_wqs(idxd);
|
||||
if (rc < 0)
|
||||
return rc;
|
||||
goto err_wqs;
|
||||
|
||||
rc = idxd_setup_engines(idxd);
|
||||
if (rc < 0)
|
||||
@ -317,6 +344,8 @@ static int idxd_setup_internals(struct idxd_device *idxd)
|
||||
err_engine:
|
||||
for (i = 0; i < idxd->max_wqs; i++)
|
||||
put_device(&idxd->wqs[i]->conf_dev);
|
||||
err_wqs:
|
||||
kfree(idxd->int_handles);
|
||||
return rc;
|
||||
}
|
||||
|
||||
@ -345,6 +374,12 @@ static void idxd_read_caps(struct idxd_device *idxd)
|
||||
/* reading generic capabilities */
|
||||
idxd->hw.gen_cap.bits = ioread64(idxd->reg_base + IDXD_GENCAP_OFFSET);
|
||||
dev_dbg(dev, "gen_cap: %#llx\n", idxd->hw.gen_cap.bits);
|
||||
|
||||
if (idxd->hw.gen_cap.cmd_cap) {
|
||||
idxd->hw.cmd_cap = ioread32(idxd->reg_base + IDXD_CMDCAP_OFFSET);
|
||||
dev_dbg(dev, "cmd_cap: %#x\n", idxd->hw.cmd_cap);
|
||||
}
|
||||
|
||||
idxd->max_xfer_bytes = 1ULL << idxd->hw.gen_cap.max_xfer_shift;
|
||||
dev_dbg(dev, "max xfer size: %llu bytes\n", idxd->max_xfer_bytes);
|
||||
idxd->max_batch_size = 1U << idxd->hw.gen_cap.max_batch_shift;
|
||||
@ -604,6 +639,24 @@ static void idxd_flush_work_list(struct idxd_irq_entry *ie)
|
||||
}
|
||||
}
|
||||
|
||||
static void idxd_release_int_handles(struct idxd_device *idxd)
|
||||
{
|
||||
struct device *dev = &idxd->pdev->dev;
|
||||
int i, rc;
|
||||
|
||||
for (i = 0; i < idxd->num_wq_irqs; i++) {
|
||||
if (idxd->hw.cmd_cap & BIT(IDXD_CMD_RELEASE_INT_HANDLE)) {
|
||||
rc = idxd_device_release_int_handle(idxd, idxd->int_handles[i],
|
||||
IDXD_IRQ_MSIX);
|
||||
if (rc < 0)
|
||||
dev_warn(dev, "irq handle %d release failed\n",
|
||||
idxd->int_handles[i]);
|
||||
else
|
||||
dev_dbg(dev, "int handle requested: %u\n", idxd->int_handles[i]);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
static void idxd_shutdown(struct pci_dev *pdev)
|
||||
{
|
||||
struct idxd_device *idxd = pci_get_drvdata(pdev);
|
||||
@ -630,6 +683,7 @@ static void idxd_shutdown(struct pci_dev *pdev)
|
||||
}
|
||||
|
||||
idxd_msix_perm_clear(idxd);
|
||||
idxd_release_int_handles(idxd);
|
||||
pci_free_irq_vectors(pdev);
|
||||
pci_iounmap(pdev, idxd->reg_base);
|
||||
pci_disable_device(pdev);
|
||||
|
@ -24,8 +24,8 @@ union gen_cap_reg {
|
||||
u64 overlap_copy:1;
|
||||
u64 cache_control_mem:1;
|
||||
u64 cache_control_cache:1;
|
||||
u64 cmd_cap:1;
|
||||
u64 rsvd:3;
|
||||
u64 int_handle_req:1;
|
||||
u64 dest_readback:1;
|
||||
u64 drain_readback:1;
|
||||
u64 rsvd2:6;
|
||||
@ -180,8 +180,11 @@ enum idxd_cmd {
|
||||
IDXD_CMD_DRAIN_PASID,
|
||||
IDXD_CMD_ABORT_PASID,
|
||||
IDXD_CMD_REQUEST_INT_HANDLE,
|
||||
IDXD_CMD_RELEASE_INT_HANDLE,
|
||||
};
|
||||
|
||||
#define CMD_INT_HANDLE_IMS 0x10000
|
||||
|
||||
#define IDXD_CMDSTS_OFFSET 0xa8
|
||||
union cmdsts_reg {
|
||||
struct {
|
||||
@ -193,6 +196,8 @@ union cmdsts_reg {
|
||||
u32 bits;
|
||||
} __packed;
|
||||
#define IDXD_CMDSTS_ACTIVE 0x80000000
|
||||
#define IDXD_CMDSTS_ERR_MASK 0xff
|
||||
#define IDXD_CMDSTS_RES_SHIFT 8
|
||||
|
||||
enum idxd_cmdsts_err {
|
||||
IDXD_CMDSTS_SUCCESS = 0,
|
||||
@ -228,6 +233,8 @@ enum idxd_cmdsts_err {
|
||||
IDXD_CMDSTS_ERR_NO_HANDLE,
|
||||
};
|
||||
|
||||
#define IDXD_CMDCAP_OFFSET 0xb0
|
||||
|
||||
#define IDXD_SWERR_OFFSET 0xc0
|
||||
#define IDXD_SWERR_VALID 0x00000001
|
||||
#define IDXD_SWERR_OVERFLOW 0x00000002
|
||||
|
@ -22,11 +22,23 @@ static struct idxd_desc *__get_desc(struct idxd_wq *wq, int idx, int cpu)
|
||||
desc->hw->pasid = idxd->pasid;
|
||||
|
||||
/*
|
||||
* Descriptor completion vectors are 1-8 for MSIX. We will round
|
||||
* robin through the 8 vectors.
|
||||
* Descriptor completion vectors are 1...N for MSIX. We will round
|
||||
* robin through the N vectors.
|
||||
*/
|
||||
wq->vec_ptr = (wq->vec_ptr % idxd->num_wq_irqs) + 1;
|
||||
desc->hw->int_handle = wq->vec_ptr;
|
||||
if (!idxd->int_handles) {
|
||||
desc->hw->int_handle = wq->vec_ptr;
|
||||
} else {
|
||||
desc->vector = wq->vec_ptr;
|
||||
/*
|
||||
* int_handles are only for descriptor completion. However for device
|
||||
* MSIX enumeration, vec 0 is used for misc interrupts. Therefore even
|
||||
* though we are rotating through 1...N for descriptor interrupts, we
|
||||
* need to acqurie the int_handles from 0..N-1.
|
||||
*/
|
||||
desc->hw->int_handle = idxd->int_handles[desc->vector - 1];
|
||||
}
|
||||
|
||||
return desc;
|
||||
}
|
||||
|
||||
@ -79,7 +91,6 @@ void idxd_free_desc(struct idxd_wq *wq, struct idxd_desc *desc)
|
||||
int idxd_submit_desc(struct idxd_wq *wq, struct idxd_desc *desc)
|
||||
{
|
||||
struct idxd_device *idxd = wq->idxd;
|
||||
int vec = desc->hw->int_handle;
|
||||
void __iomem *portal;
|
||||
int rc;
|
||||
|
||||
@ -117,9 +128,19 @@ int idxd_submit_desc(struct idxd_wq *wq, struct idxd_desc *desc)
|
||||
* Pending the descriptor to the lockless list for the irq_entry
|
||||
* that we designated the descriptor to.
|
||||
*/
|
||||
if (desc->hw->flags & IDXD_OP_FLAG_RCI)
|
||||
llist_add(&desc->llnode,
|
||||
&idxd->irq_entries[vec].pending_llist);
|
||||
if (desc->hw->flags & IDXD_OP_FLAG_RCI) {
|
||||
int vec;
|
||||
|
||||
/*
|
||||
* If the driver is on host kernel, it would be the value
|
||||
* assigned to interrupt handle, which is index for MSIX
|
||||
* vector. If it's guest then can't use the int_handle since
|
||||
* that is the index to IMS for the entire device. The guest
|
||||
* device local index will be used.
|
||||
*/
|
||||
vec = !idxd->int_handles ? desc->hw->int_handle : desc->vector;
|
||||
llist_add(&desc->llnode, &idxd->irq_entries[vec].pending_llist);
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
@ -1600,6 +1600,7 @@ static void idxd_conf_device_release(struct device *dev)
|
||||
kfree(idxd->wqs);
|
||||
kfree(idxd->engines);
|
||||
kfree(idxd->irq_entries);
|
||||
kfree(idxd->int_handles);
|
||||
ida_free(&idxd_ida, idxd->id);
|
||||
kfree(idxd);
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user