powerpc/xive: Remove P9 DD1 flag XIVE_IRQ_FLAG_EOI_FW
This flag was used to support the P9 DD1 and we have stopped supporting this CPU when DD2 came out. See skiboot commit: https://github.com/open-power/skiboot/commit/0b0d15e3c170 Also, remove eoi handler which is now unused. Signed-off-by: Cédric Le Goater <clg@kaod.org> Reviewed-by: Greg Kurz <groug@kaod.org> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au> Link: https://lore.kernel.org/r/20201210171450.1933725-11-clg@kaod.org
This commit is contained in:
committed by
Michael Ellerman
parent
b5277d18c6
commit
cf58b74666
@ -1093,7 +1093,7 @@ enum {
|
|||||||
OPAL_XIVE_IRQ_LSI = 0x00000004,
|
OPAL_XIVE_IRQ_LSI = 0x00000004,
|
||||||
OPAL_XIVE_IRQ_SHIFT_BUG = 0x00000008, /* P9 DD1.0 workaround */
|
OPAL_XIVE_IRQ_SHIFT_BUG = 0x00000008, /* P9 DD1.0 workaround */
|
||||||
OPAL_XIVE_IRQ_MASK_VIA_FW = 0x00000010, /* P9 DD1.0 workaround */
|
OPAL_XIVE_IRQ_MASK_VIA_FW = 0x00000010, /* P9 DD1.0 workaround */
|
||||||
OPAL_XIVE_IRQ_EOI_VIA_FW = 0x00000020,
|
OPAL_XIVE_IRQ_EOI_VIA_FW = 0x00000020, /* P9 DD1.0 workaround */
|
||||||
};
|
};
|
||||||
|
|
||||||
/* Flags for OPAL_XIVE_GET/SET_QUEUE_INFO */
|
/* Flags for OPAL_XIVE_GET/SET_QUEUE_INFO */
|
||||||
|
@ -62,7 +62,7 @@ struct xive_irq_data {
|
|||||||
#define XIVE_IRQ_FLAG_LSI 0x02
|
#define XIVE_IRQ_FLAG_LSI 0x02
|
||||||
/* #define XIVE_IRQ_FLAG_SHIFT_BUG 0x04 */ /* P9 DD1.0 workaround */
|
/* #define XIVE_IRQ_FLAG_SHIFT_BUG 0x04 */ /* P9 DD1.0 workaround */
|
||||||
/* #define XIVE_IRQ_FLAG_MASK_FW 0x08 */ /* P9 DD1.0 workaround */
|
/* #define XIVE_IRQ_FLAG_MASK_FW 0x08 */ /* P9 DD1.0 workaround */
|
||||||
#define XIVE_IRQ_FLAG_EOI_FW 0x10
|
/* #define XIVE_IRQ_FLAG_EOI_FW 0x10 */ /* P9 DD1.0 workaround */
|
||||||
#define XIVE_IRQ_FLAG_H_INT_ESB 0x20
|
#define XIVE_IRQ_FLAG_H_INT_ESB 0x20
|
||||||
|
|
||||||
/* Special flag set by KVM for excalation interrupts */
|
/* Special flag set by KVM for excalation interrupts */
|
||||||
|
@ -74,8 +74,6 @@ static void GLUE(X_PFX,source_eoi)(u32 hw_irq, struct xive_irq_data *xd)
|
|||||||
/* If the XIVE supports the new "store EOI facility, use it */
|
/* If the XIVE supports the new "store EOI facility, use it */
|
||||||
if (xd->flags & XIVE_IRQ_FLAG_STORE_EOI)
|
if (xd->flags & XIVE_IRQ_FLAG_STORE_EOI)
|
||||||
__x_writeq(0, __x_eoi_page(xd) + XIVE_ESB_STORE_EOI);
|
__x_writeq(0, __x_eoi_page(xd) + XIVE_ESB_STORE_EOI);
|
||||||
else if (hw_irq && xd->flags & XIVE_IRQ_FLAG_EOI_FW)
|
|
||||||
opal_int_eoi(hw_irq);
|
|
||||||
else if (xd->flags & XIVE_IRQ_FLAG_LSI) {
|
else if (xd->flags & XIVE_IRQ_FLAG_LSI) {
|
||||||
/*
|
/*
|
||||||
* For LSIs the HW EOI cycle is used rather than PQ bits,
|
* For LSIs the HW EOI cycle is used rather than PQ bits,
|
||||||
|
@ -354,18 +354,7 @@ static void xive_do_source_eoi(u32 hw_irq, struct xive_irq_data *xd)
|
|||||||
/* If the XIVE supports the new "store EOI facility, use it */
|
/* If the XIVE supports the new "store EOI facility, use it */
|
||||||
if (xd->flags & XIVE_IRQ_FLAG_STORE_EOI)
|
if (xd->flags & XIVE_IRQ_FLAG_STORE_EOI)
|
||||||
xive_esb_write(xd, XIVE_ESB_STORE_EOI, 0);
|
xive_esb_write(xd, XIVE_ESB_STORE_EOI, 0);
|
||||||
else if (hw_irq && xd->flags & XIVE_IRQ_FLAG_EOI_FW) {
|
else {
|
||||||
/*
|
|
||||||
* The FW told us to call it. This happens for some
|
|
||||||
* interrupt sources that need additional HW whacking
|
|
||||||
* beyond the ESB manipulation. For example LPC interrupts
|
|
||||||
* on P9 DD1.0 needed a latch to be clared in the LPC bridge
|
|
||||||
* itself. The Firmware will take care of it.
|
|
||||||
*/
|
|
||||||
if (WARN_ON_ONCE(!xive_ops->eoi))
|
|
||||||
return;
|
|
||||||
xive_ops->eoi(hw_irq);
|
|
||||||
} else {
|
|
||||||
u8 eoi_val;
|
u8 eoi_val;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -1267,7 +1256,6 @@ static const struct {
|
|||||||
} xive_irq_flags[] = {
|
} xive_irq_flags[] = {
|
||||||
{ XIVE_IRQ_FLAG_STORE_EOI, "STORE_EOI" },
|
{ XIVE_IRQ_FLAG_STORE_EOI, "STORE_EOI" },
|
||||||
{ XIVE_IRQ_FLAG_LSI, "LSI" },
|
{ XIVE_IRQ_FLAG_LSI, "LSI" },
|
||||||
{ XIVE_IRQ_FLAG_EOI_FW, "EOI_FW" },
|
|
||||||
{ XIVE_IRQ_FLAG_H_INT_ESB, "H_INT_ESB" },
|
{ XIVE_IRQ_FLAG_H_INT_ESB, "H_INT_ESB" },
|
||||||
{ XIVE_IRQ_FLAG_NO_EOI, "NO_EOI" },
|
{ XIVE_IRQ_FLAG_NO_EOI, "NO_EOI" },
|
||||||
};
|
};
|
||||||
|
@ -64,8 +64,6 @@ int xive_native_populate_irq_data(u32 hw_irq, struct xive_irq_data *data)
|
|||||||
data->flags |= XIVE_IRQ_FLAG_STORE_EOI;
|
data->flags |= XIVE_IRQ_FLAG_STORE_EOI;
|
||||||
if (opal_flags & OPAL_XIVE_IRQ_LSI)
|
if (opal_flags & OPAL_XIVE_IRQ_LSI)
|
||||||
data->flags |= XIVE_IRQ_FLAG_LSI;
|
data->flags |= XIVE_IRQ_FLAG_LSI;
|
||||||
if (opal_flags & OPAL_XIVE_IRQ_EOI_VIA_FW)
|
|
||||||
data->flags |= XIVE_IRQ_FLAG_EOI_FW;
|
|
||||||
data->eoi_page = be64_to_cpu(eoi_page);
|
data->eoi_page = be64_to_cpu(eoi_page);
|
||||||
data->trig_page = be64_to_cpu(trig_page);
|
data->trig_page = be64_to_cpu(trig_page);
|
||||||
data->esb_shift = be32_to_cpu(esb_shift);
|
data->esb_shift = be32_to_cpu(esb_shift);
|
||||||
@ -380,15 +378,6 @@ static void xive_native_update_pending(struct xive_cpu *xc)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void xive_native_eoi(u32 hw_irq)
|
|
||||||
{
|
|
||||||
/*
|
|
||||||
* Not normally used except if specific interrupts need
|
|
||||||
* a workaround on EOI.
|
|
||||||
*/
|
|
||||||
opal_int_eoi(hw_irq);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void xive_native_setup_cpu(unsigned int cpu, struct xive_cpu *xc)
|
static void xive_native_setup_cpu(unsigned int cpu, struct xive_cpu *xc)
|
||||||
{
|
{
|
||||||
s64 rc;
|
s64 rc;
|
||||||
@ -471,7 +460,6 @@ static const struct xive_ops xive_native_ops = {
|
|||||||
.match = xive_native_match,
|
.match = xive_native_match,
|
||||||
.shutdown = xive_native_shutdown,
|
.shutdown = xive_native_shutdown,
|
||||||
.update_pending = xive_native_update_pending,
|
.update_pending = xive_native_update_pending,
|
||||||
.eoi = xive_native_eoi,
|
|
||||||
.setup_cpu = xive_native_setup_cpu,
|
.setup_cpu = xive_native_setup_cpu,
|
||||||
.teardown_cpu = xive_native_teardown_cpu,
|
.teardown_cpu = xive_native_teardown_cpu,
|
||||||
.sync_source = xive_native_sync_source,
|
.sync_source = xive_native_sync_source,
|
||||||
|
@ -628,11 +628,6 @@ static void xive_spapr_update_pending(struct xive_cpu *xc)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void xive_spapr_eoi(u32 hw_irq)
|
|
||||||
{
|
|
||||||
/* Not used */;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void xive_spapr_setup_cpu(unsigned int cpu, struct xive_cpu *xc)
|
static void xive_spapr_setup_cpu(unsigned int cpu, struct xive_cpu *xc)
|
||||||
{
|
{
|
||||||
/* Only some debug on the TIMA settings */
|
/* Only some debug on the TIMA settings */
|
||||||
@ -677,7 +672,6 @@ static const struct xive_ops xive_spapr_ops = {
|
|||||||
.match = xive_spapr_match,
|
.match = xive_spapr_match,
|
||||||
.shutdown = xive_spapr_shutdown,
|
.shutdown = xive_spapr_shutdown,
|
||||||
.update_pending = xive_spapr_update_pending,
|
.update_pending = xive_spapr_update_pending,
|
||||||
.eoi = xive_spapr_eoi,
|
|
||||||
.setup_cpu = xive_spapr_setup_cpu,
|
.setup_cpu = xive_spapr_setup_cpu,
|
||||||
.teardown_cpu = xive_spapr_teardown_cpu,
|
.teardown_cpu = xive_spapr_teardown_cpu,
|
||||||
.sync_source = xive_spapr_sync_source,
|
.sync_source = xive_spapr_sync_source,
|
||||||
|
@ -52,7 +52,6 @@ struct xive_ops {
|
|||||||
void (*shutdown)(void);
|
void (*shutdown)(void);
|
||||||
|
|
||||||
void (*update_pending)(struct xive_cpu *xc);
|
void (*update_pending)(struct xive_cpu *xc);
|
||||||
void (*eoi)(u32 hw_irq);
|
|
||||||
void (*sync_source)(u32 hw_irq);
|
void (*sync_source)(u32 hw_irq);
|
||||||
u64 (*esb_rw)(u32 hw_irq, u32 offset, u64 data, bool write);
|
u64 (*esb_rw)(u32 hw_irq, u32 offset, u64 data, bool write);
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
|
Reference in New Issue
Block a user