19d31c5f11
A series of hcalls have been added to the PAPR which allow a regular guest partition to create and manage guest partitions of its own. KVM already had an interface that allowed this on powernv platforms. This existing interface will now be called "nestedv1". The newly added PAPR interface will be called "nestedv2". PHYP will support the nestedv2 interface. At this time the host side of the nestedv2 interface has not been implemented on powernv but there is no technical reason why it could not be added. The nestedv1 interface is still supported. Add support to KVM to utilize these hcalls to enable running nested guests as a pseries guest on PHYP. Overview of the new hcall usage: - L1 and L0 negotiate capabilities with H_GUEST_{G,S}ET_CAPABILITIES() - L1 requests the L0 create a L2 with H_GUEST_CREATE() and receives a handle to use in future hcalls - L1 requests the L0 create a L2 vCPU with H_GUEST_CREATE_VCPU() - L1 sets up the L2 using H_GUEST_SET and the H_GUEST_VCPU_RUN input buffer - L1 requests the L0 runs the L2 vCPU using H_GUEST_VCPU_RUN() - L2 returns to L1 with an exit reason and L1 reads the H_GUEST_VCPU_RUN output buffer populated by the L0 - L1 handles the exit using H_GET_STATE if necessary - L1 reruns L2 vCPU with H_GUEST_VCPU_RUN - L1 frees the L2 in the L0 with H_GUEST_DELETE() Support for the new API is determined by trying H_GUEST_GET_CAPABILITIES. On a successful return, use the nestedv2 interface. Use the vcpu register state setters for tracking modified guest state elements and copy the thread wide values into the H_GUEST_VCPU_RUN input buffer immediately before running a L2. The guest wide elements can not be added to the input buffer so send them with a separate H_GUEST_SET call if necessary. Make the vcpu register getter load the corresponding value from the real host with H_GUEST_GET. To avoid unnecessarily calling H_GUEST_GET, track which values have already been loaded between H_GUEST_VCPU_RUN calls. If an element is present in the H_GUEST_VCPU_RUN output buffer it also does not need to be loaded again. Tested-by: Sachin Sant <sachinp@linux.ibm.com> Signed-off-by: Vaibhav Jain <vaibhav@linux.ibm.com> Signed-off-by: Gautam Menghani <gautam@linux.ibm.com> Signed-off-by: Kautuk Consul <kconsul@linux.vnet.ibm.com> Signed-off-by: Amit Machhiwal <amachhiw@linux.vnet.ibm.com> Signed-off-by: Jordan Niethe <jniethe5@gmail.com> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au> Link: https://msgid.link/20230914030600.16993-11-jniethe5@gmail.com
676 lines
15 KiB
C
676 lines
15 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef _ASM_POWERPC_PLPAR_WRAPPERS_H
|
|
#define _ASM_POWERPC_PLPAR_WRAPPERS_H
|
|
|
|
#ifdef CONFIG_PPC_PSERIES
|
|
|
|
#include <linux/string.h>
|
|
#include <linux/irqflags.h>
|
|
#include <linux/delay.h>
|
|
|
|
#include <asm/hvcall.h>
|
|
#include <asm/paca.h>
|
|
#include <asm/lppaca.h>
|
|
#include <asm/page.h>
|
|
|
|
static inline long poll_pending(void)
|
|
{
|
|
return plpar_hcall_norets(H_POLL_PENDING);
|
|
}
|
|
|
|
static inline u8 get_cede_latency_hint(void)
|
|
{
|
|
return get_lppaca()->cede_latency_hint;
|
|
}
|
|
|
|
static inline void set_cede_latency_hint(u8 latency_hint)
|
|
{
|
|
get_lppaca()->cede_latency_hint = latency_hint;
|
|
}
|
|
|
|
static inline long cede_processor(void)
|
|
{
|
|
/*
|
|
* We cannot call tracepoints inside RCU idle regions which
|
|
* means we must not trace H_CEDE.
|
|
*/
|
|
return plpar_hcall_norets_notrace(H_CEDE);
|
|
}
|
|
|
|
static inline long extended_cede_processor(unsigned long latency_hint)
|
|
{
|
|
long rc;
|
|
u8 old_latency_hint = get_cede_latency_hint();
|
|
|
|
set_cede_latency_hint(latency_hint);
|
|
|
|
rc = cede_processor();
|
|
|
|
/* Ensure that H_CEDE returns with IRQs on */
|
|
if (WARN_ON(IS_ENABLED(CONFIG_PPC_IRQ_SOFT_MASK_DEBUG) && !(mfmsr() & MSR_EE)))
|
|
__hard_irq_enable();
|
|
|
|
set_cede_latency_hint(old_latency_hint);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long vpa_call(unsigned long flags, unsigned long cpu,
|
|
unsigned long vpa)
|
|
{
|
|
flags = flags << H_VPA_FUNC_SHIFT;
|
|
|
|
return plpar_hcall_norets(H_REGISTER_VPA, flags, cpu, vpa);
|
|
}
|
|
|
|
static inline long unregister_vpa(unsigned long cpu)
|
|
{
|
|
return vpa_call(H_VPA_DEREG_VPA, cpu, 0);
|
|
}
|
|
|
|
static inline long register_vpa(unsigned long cpu, unsigned long vpa)
|
|
{
|
|
return vpa_call(H_VPA_REG_VPA, cpu, vpa);
|
|
}
|
|
|
|
static inline long unregister_slb_shadow(unsigned long cpu)
|
|
{
|
|
return vpa_call(H_VPA_DEREG_SLB, cpu, 0);
|
|
}
|
|
|
|
static inline long register_slb_shadow(unsigned long cpu, unsigned long vpa)
|
|
{
|
|
return vpa_call(H_VPA_REG_SLB, cpu, vpa);
|
|
}
|
|
|
|
static inline long unregister_dtl(unsigned long cpu)
|
|
{
|
|
return vpa_call(H_VPA_DEREG_DTL, cpu, 0);
|
|
}
|
|
|
|
static inline long register_dtl(unsigned long cpu, unsigned long vpa)
|
|
{
|
|
return vpa_call(H_VPA_REG_DTL, cpu, vpa);
|
|
}
|
|
|
|
extern void vpa_init(int cpu);
|
|
|
|
static inline long plpar_pte_enter(unsigned long flags,
|
|
unsigned long hpte_group, unsigned long hpte_v,
|
|
unsigned long hpte_r, unsigned long *slot)
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
rc = plpar_hcall(H_ENTER, retbuf, flags, hpte_group, hpte_v, hpte_r);
|
|
|
|
*slot = retbuf[0];
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_pte_remove(unsigned long flags, unsigned long ptex,
|
|
unsigned long avpn, unsigned long *old_pteh_ret,
|
|
unsigned long *old_ptel_ret)
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
rc = plpar_hcall(H_REMOVE, retbuf, flags, ptex, avpn);
|
|
|
|
*old_pteh_ret = retbuf[0];
|
|
*old_ptel_ret = retbuf[1];
|
|
|
|
return rc;
|
|
}
|
|
|
|
/* plpar_pte_remove_raw can be called in real mode. It calls plpar_hcall_raw */
|
|
static inline long plpar_pte_remove_raw(unsigned long flags, unsigned long ptex,
|
|
unsigned long avpn, unsigned long *old_pteh_ret,
|
|
unsigned long *old_ptel_ret)
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
rc = plpar_hcall_raw(H_REMOVE, retbuf, flags, ptex, avpn);
|
|
|
|
*old_pteh_ret = retbuf[0];
|
|
*old_ptel_ret = retbuf[1];
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_pte_read(unsigned long flags, unsigned long ptex,
|
|
unsigned long *old_pteh_ret, unsigned long *old_ptel_ret)
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
rc = plpar_hcall(H_READ, retbuf, flags, ptex);
|
|
|
|
*old_pteh_ret = retbuf[0];
|
|
*old_ptel_ret = retbuf[1];
|
|
|
|
return rc;
|
|
}
|
|
|
|
/* plpar_pte_read_raw can be called in real mode. It calls plpar_hcall_raw */
|
|
static inline long plpar_pte_read_raw(unsigned long flags, unsigned long ptex,
|
|
unsigned long *old_pteh_ret, unsigned long *old_ptel_ret)
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
rc = plpar_hcall_raw(H_READ, retbuf, flags, ptex);
|
|
|
|
*old_pteh_ret = retbuf[0];
|
|
*old_ptel_ret = retbuf[1];
|
|
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* ptes must be 8*sizeof(unsigned long)
|
|
*/
|
|
static inline long plpar_pte_read_4(unsigned long flags, unsigned long ptex,
|
|
unsigned long *ptes)
|
|
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL9_BUFSIZE];
|
|
|
|
rc = plpar_hcall9(H_READ, retbuf, flags | H_READ_4, ptex);
|
|
|
|
memcpy(ptes, retbuf, 8*sizeof(unsigned long));
|
|
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* plpar_pte_read_4_raw can be called in real mode.
|
|
* ptes must be 8*sizeof(unsigned long)
|
|
*/
|
|
static inline long plpar_pte_read_4_raw(unsigned long flags, unsigned long ptex,
|
|
unsigned long *ptes)
|
|
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL9_BUFSIZE];
|
|
|
|
rc = plpar_hcall9_raw(H_READ, retbuf, flags | H_READ_4, ptex);
|
|
|
|
memcpy(ptes, retbuf, 8*sizeof(unsigned long));
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_pte_protect(unsigned long flags, unsigned long ptex,
|
|
unsigned long avpn)
|
|
{
|
|
return plpar_hcall_norets(H_PROTECT, flags, ptex, avpn);
|
|
}
|
|
|
|
static inline long plpar_resize_hpt_prepare(unsigned long flags,
|
|
unsigned long shift)
|
|
{
|
|
return plpar_hcall_norets(H_RESIZE_HPT_PREPARE, flags, shift);
|
|
}
|
|
|
|
static inline long plpar_resize_hpt_commit(unsigned long flags,
|
|
unsigned long shift)
|
|
{
|
|
return plpar_hcall_norets(H_RESIZE_HPT_COMMIT, flags, shift);
|
|
}
|
|
|
|
static inline long plpar_tce_get(unsigned long liobn, unsigned long ioba,
|
|
unsigned long *tce_ret)
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
rc = plpar_hcall(H_GET_TCE, retbuf, liobn, ioba);
|
|
|
|
*tce_ret = retbuf[0];
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_tce_put(unsigned long liobn, unsigned long ioba,
|
|
unsigned long tceval)
|
|
{
|
|
return plpar_hcall_norets(H_PUT_TCE, liobn, ioba, tceval);
|
|
}
|
|
|
|
static inline long plpar_tce_put_indirect(unsigned long liobn,
|
|
unsigned long ioba, unsigned long page, unsigned long count)
|
|
{
|
|
return plpar_hcall_norets(H_PUT_TCE_INDIRECT, liobn, ioba, page, count);
|
|
}
|
|
|
|
static inline long plpar_tce_stuff(unsigned long liobn, unsigned long ioba,
|
|
unsigned long tceval, unsigned long count)
|
|
{
|
|
return plpar_hcall_norets(H_STUFF_TCE, liobn, ioba, tceval, count);
|
|
}
|
|
|
|
/* Set various resource mode parameters */
|
|
static inline long plpar_set_mode(unsigned long mflags, unsigned long resource,
|
|
unsigned long value1, unsigned long value2)
|
|
{
|
|
return plpar_hcall_norets(H_SET_MODE, mflags, resource, value1, value2);
|
|
}
|
|
|
|
/*
|
|
* Enable relocation on exceptions on this partition
|
|
*
|
|
* Note: this call has a partition wide scope and can take a while to complete.
|
|
* If it returns H_LONG_BUSY_* it should be retried periodically until it
|
|
* returns H_SUCCESS.
|
|
*/
|
|
static inline long enable_reloc_on_exceptions(void)
|
|
{
|
|
/* mflags = 3: Exceptions at 0xC000000000004000 */
|
|
return plpar_set_mode(3, H_SET_MODE_RESOURCE_ADDR_TRANS_MODE, 0, 0);
|
|
}
|
|
|
|
/*
|
|
* Disable relocation on exceptions on this partition
|
|
*
|
|
* Note: this call has a partition wide scope and can take a while to complete.
|
|
* If it returns H_LONG_BUSY_* it should be retried periodically until it
|
|
* returns H_SUCCESS.
|
|
*/
|
|
static inline long disable_reloc_on_exceptions(void) {
|
|
return plpar_set_mode(0, H_SET_MODE_RESOURCE_ADDR_TRANS_MODE, 0, 0);
|
|
}
|
|
|
|
/*
|
|
* Take exceptions in big endian mode on this partition
|
|
*
|
|
* Note: this call has a partition wide scope and can take a while to complete.
|
|
* If it returns H_LONG_BUSY_* it should be retried periodically until it
|
|
* returns H_SUCCESS.
|
|
*/
|
|
static inline long enable_big_endian_exceptions(void)
|
|
{
|
|
/* mflags = 0: big endian exceptions */
|
|
return plpar_set_mode(0, H_SET_MODE_RESOURCE_LE, 0, 0);
|
|
}
|
|
|
|
/*
|
|
* Take exceptions in little endian mode on this partition
|
|
*
|
|
* Note: this call has a partition wide scope and can take a while to complete.
|
|
* If it returns H_LONG_BUSY_* it should be retried periodically until it
|
|
* returns H_SUCCESS.
|
|
*/
|
|
static inline long enable_little_endian_exceptions(void)
|
|
{
|
|
/* mflags = 1: little endian exceptions */
|
|
return plpar_set_mode(1, H_SET_MODE_RESOURCE_LE, 0, 0);
|
|
}
|
|
|
|
static inline long plpar_set_ciabr(unsigned long ciabr)
|
|
{
|
|
return plpar_set_mode(0, H_SET_MODE_RESOURCE_SET_CIABR, ciabr, 0);
|
|
}
|
|
|
|
static inline long plpar_set_watchpoint0(unsigned long dawr0, unsigned long dawrx0)
|
|
{
|
|
return plpar_set_mode(0, H_SET_MODE_RESOURCE_SET_DAWR0, dawr0, dawrx0);
|
|
}
|
|
|
|
static inline long plpar_set_watchpoint1(unsigned long dawr1, unsigned long dawrx1)
|
|
{
|
|
return plpar_set_mode(0, H_SET_MODE_RESOURCE_SET_DAWR1, dawr1, dawrx1);
|
|
}
|
|
|
|
static inline long plpar_signal_sys_reset(long cpu)
|
|
{
|
|
return plpar_hcall_norets(H_SIGNAL_SYS_RESET, cpu);
|
|
}
|
|
|
|
static inline long plpar_get_cpu_characteristics(struct h_cpu_char_result *p)
|
|
{
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
long rc;
|
|
|
|
rc = plpar_hcall(H_GET_CPU_CHARACTERISTICS, retbuf);
|
|
if (rc == H_SUCCESS) {
|
|
p->character = retbuf[0];
|
|
p->behaviour = retbuf[1];
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_guest_create(unsigned long flags, unsigned long *guest_id)
|
|
{
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
unsigned long token;
|
|
long rc;
|
|
|
|
token = -1UL;
|
|
do {
|
|
rc = plpar_hcall(H_GUEST_CREATE, retbuf, flags, token);
|
|
if (rc == H_SUCCESS)
|
|
*guest_id = retbuf[0];
|
|
|
|
if (rc == H_BUSY) {
|
|
token = retbuf[0];
|
|
cond_resched();
|
|
}
|
|
|
|
if (H_IS_LONG_BUSY(rc)) {
|
|
token = retbuf[0];
|
|
msleep(get_longbusy_msecs(rc));
|
|
rc = H_BUSY;
|
|
}
|
|
|
|
} while (rc == H_BUSY);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_guest_create_vcpu(unsigned long flags,
|
|
unsigned long guest_id,
|
|
unsigned long vcpu_id)
|
|
{
|
|
long rc;
|
|
|
|
do {
|
|
rc = plpar_hcall_norets(H_GUEST_CREATE_VCPU, 0, guest_id, vcpu_id);
|
|
|
|
if (rc == H_BUSY)
|
|
cond_resched();
|
|
|
|
if (H_IS_LONG_BUSY(rc)) {
|
|
msleep(get_longbusy_msecs(rc));
|
|
rc = H_BUSY;
|
|
}
|
|
|
|
} while (rc == H_BUSY);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_guest_set_state(unsigned long flags,
|
|
unsigned long guest_id,
|
|
unsigned long vcpu_id,
|
|
unsigned long data_buffer,
|
|
unsigned long data_size,
|
|
unsigned long *failed_index)
|
|
{
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
long rc;
|
|
|
|
while (true) {
|
|
rc = plpar_hcall(H_GUEST_SET_STATE, retbuf, flags, guest_id,
|
|
vcpu_id, data_buffer, data_size);
|
|
|
|
if (rc == H_BUSY) {
|
|
cpu_relax();
|
|
continue;
|
|
}
|
|
|
|
if (H_IS_LONG_BUSY(rc)) {
|
|
mdelay(get_longbusy_msecs(rc));
|
|
continue;
|
|
}
|
|
|
|
if (rc == H_INVALID_ELEMENT_ID)
|
|
*failed_index = retbuf[0];
|
|
else if (rc == H_INVALID_ELEMENT_SIZE)
|
|
*failed_index = retbuf[0];
|
|
else if (rc == H_INVALID_ELEMENT_VALUE)
|
|
*failed_index = retbuf[0];
|
|
|
|
break;
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_guest_get_state(unsigned long flags,
|
|
unsigned long guest_id,
|
|
unsigned long vcpu_id,
|
|
unsigned long data_buffer,
|
|
unsigned long data_size,
|
|
unsigned long *failed_index)
|
|
{
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
long rc;
|
|
|
|
while (true) {
|
|
rc = plpar_hcall(H_GUEST_GET_STATE, retbuf, flags, guest_id,
|
|
vcpu_id, data_buffer, data_size);
|
|
|
|
if (rc == H_BUSY) {
|
|
cpu_relax();
|
|
continue;
|
|
}
|
|
|
|
if (H_IS_LONG_BUSY(rc)) {
|
|
mdelay(get_longbusy_msecs(rc));
|
|
continue;
|
|
}
|
|
|
|
if (rc == H_INVALID_ELEMENT_ID)
|
|
*failed_index = retbuf[0];
|
|
else if (rc == H_INVALID_ELEMENT_SIZE)
|
|
*failed_index = retbuf[0];
|
|
else if (rc == H_INVALID_ELEMENT_VALUE)
|
|
*failed_index = retbuf[0];
|
|
|
|
break;
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_guest_run_vcpu(unsigned long flags, unsigned long guest_id,
|
|
unsigned long vcpu_id, int *trap,
|
|
unsigned long *failed_index)
|
|
{
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
long rc;
|
|
|
|
rc = plpar_hcall(H_GUEST_RUN_VCPU, retbuf, flags, guest_id, vcpu_id);
|
|
if (rc == H_SUCCESS)
|
|
*trap = retbuf[0];
|
|
else if (rc == H_INVALID_ELEMENT_ID)
|
|
*failed_index = retbuf[0];
|
|
else if (rc == H_INVALID_ELEMENT_SIZE)
|
|
*failed_index = retbuf[0];
|
|
else if (rc == H_INVALID_ELEMENT_VALUE)
|
|
*failed_index = retbuf[0];
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_guest_delete(unsigned long flags, u64 guest_id)
|
|
{
|
|
long rc;
|
|
|
|
do {
|
|
rc = plpar_hcall_norets(H_GUEST_DELETE, flags, guest_id);
|
|
if (rc == H_BUSY)
|
|
cond_resched();
|
|
|
|
if (H_IS_LONG_BUSY(rc)) {
|
|
msleep(get_longbusy_msecs(rc));
|
|
rc = H_BUSY;
|
|
}
|
|
|
|
} while (rc == H_BUSY);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_guest_set_capabilities(unsigned long flags,
|
|
unsigned long capabilities)
|
|
{
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
long rc;
|
|
|
|
do {
|
|
rc = plpar_hcall(H_GUEST_SET_CAPABILITIES, retbuf, flags, capabilities);
|
|
if (rc == H_BUSY)
|
|
cond_resched();
|
|
|
|
if (H_IS_LONG_BUSY(rc)) {
|
|
msleep(get_longbusy_msecs(rc));
|
|
rc = H_BUSY;
|
|
}
|
|
} while (rc == H_BUSY);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_guest_get_capabilities(unsigned long flags,
|
|
unsigned long *capabilities)
|
|
{
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
long rc;
|
|
|
|
do {
|
|
rc = plpar_hcall(H_GUEST_GET_CAPABILITIES, retbuf, flags);
|
|
if (rc == H_BUSY)
|
|
cond_resched();
|
|
|
|
if (H_IS_LONG_BUSY(rc)) {
|
|
msleep(get_longbusy_msecs(rc));
|
|
rc = H_BUSY;
|
|
}
|
|
} while (rc == H_BUSY);
|
|
|
|
if (rc == H_SUCCESS)
|
|
*capabilities = retbuf[0];
|
|
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* Wrapper to H_RPT_INVALIDATE hcall that handles return values appropriately
|
|
*
|
|
* - Returns H_SUCCESS on success
|
|
* - For H_BUSY return value, we retry the hcall.
|
|
* - For any other hcall failures, attempt a full flush once before
|
|
* resorting to BUG().
|
|
*
|
|
* Note: This hcall is expected to fail only very rarely. The correct
|
|
* error recovery of killing the process/guest will be eventually
|
|
* needed.
|
|
*/
|
|
static inline long pseries_rpt_invalidate(u64 pid, u64 target, u64 type,
|
|
u64 page_sizes, u64 start, u64 end)
|
|
{
|
|
long rc;
|
|
unsigned long all;
|
|
|
|
while (true) {
|
|
rc = plpar_hcall_norets(H_RPT_INVALIDATE, pid, target, type,
|
|
page_sizes, start, end);
|
|
if (rc == H_BUSY) {
|
|
cpu_relax();
|
|
continue;
|
|
} else if (rc == H_SUCCESS)
|
|
return rc;
|
|
|
|
/* Flush request failed, try with a full flush once */
|
|
if (type & H_RPTI_TYPE_NESTED)
|
|
all = H_RPTI_TYPE_NESTED | H_RPTI_TYPE_NESTED_ALL;
|
|
else
|
|
all = H_RPTI_TYPE_ALL;
|
|
retry:
|
|
rc = plpar_hcall_norets(H_RPT_INVALIDATE, pid, target,
|
|
all, page_sizes, 0, -1UL);
|
|
if (rc == H_BUSY) {
|
|
cpu_relax();
|
|
goto retry;
|
|
} else if (rc == H_SUCCESS)
|
|
return rc;
|
|
|
|
BUG();
|
|
}
|
|
}
|
|
|
|
#else /* !CONFIG_PPC_PSERIES */
|
|
|
|
static inline long plpar_set_ciabr(unsigned long ciabr)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline long plpar_pte_read_4(unsigned long flags, unsigned long ptex,
|
|
unsigned long *ptes)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline long pseries_rpt_invalidate(u64 pid, u64 target, u64 type,
|
|
u64 page_sizes, u64 start, u64 end)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline long plpar_guest_create_vcpu(unsigned long flags,
|
|
unsigned long guest_id,
|
|
unsigned long vcpu_id)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline long plpar_guest_get_state(unsigned long flags,
|
|
unsigned long guest_id,
|
|
unsigned long vcpu_id,
|
|
unsigned long data_buffer,
|
|
unsigned long data_size,
|
|
unsigned long *failed_index)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline long plpar_guest_set_state(unsigned long flags,
|
|
unsigned long guest_id,
|
|
unsigned long vcpu_id,
|
|
unsigned long data_buffer,
|
|
unsigned long data_size,
|
|
unsigned long *failed_index)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline long plpar_guest_run_vcpu(unsigned long flags, unsigned long guest_id,
|
|
unsigned long vcpu_id, int *trap,
|
|
unsigned long *failed_index)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline long plpar_guest_create(unsigned long flags, unsigned long *guest_id)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline long plpar_guest_delete(unsigned long flags, u64 guest_id)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline long plpar_guest_get_capabilities(unsigned long flags,
|
|
unsigned long *capabilities)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline long plpar_guest_set_capabilities(unsigned long flags,
|
|
unsigned long capabilities)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
#endif /* CONFIG_PPC_PSERIES */
|
|
|
|
#endif /* _ASM_POWERPC_PLPAR_WRAPPERS_H */
|