3b5d56b931
When a host stops or suspends a VM it will set a flag to show this. The watchdog will use these functions to determine if a softlockup is real, or the result of a suspended VM. Signed-off-by: Eric B Munson <emunson@mgebm.net> asm-generic changes Acked-by: Arnd Bergmann <arnd@arndb.de> Signed-off-by: Marcelo Tosatti <mtosatti@redhat.com> Signed-off-by: Avi Kivity <avi@redhat.com>
217 lines
4.7 KiB
C
217 lines
4.7 KiB
C
/*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License, version 2, as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
|
|
*
|
|
* Copyright IBM Corp. 2008
|
|
*
|
|
* Authors: Hollis Blanchard <hollisb@us.ibm.com>
|
|
*/
|
|
|
|
#ifndef __POWERPC_KVM_PARA_H__
|
|
#define __POWERPC_KVM_PARA_H__
|
|
|
|
#include <linux/types.h>
|
|
|
|
/*
|
|
* Additions to this struct must only occur at the end, and should be
|
|
* accompanied by a KVM_MAGIC_FEAT flag to advertise that they are present
|
|
* (albeit not necessarily relevant to the current target hardware platform).
|
|
*
|
|
* Struct fields are always 32 or 64 bit aligned, depending on them being 32
|
|
* or 64 bit wide respectively.
|
|
*
|
|
* See Documentation/virtual/kvm/ppc-pv.txt
|
|
*/
|
|
struct kvm_vcpu_arch_shared {
|
|
__u64 scratch1;
|
|
__u64 scratch2;
|
|
__u64 scratch3;
|
|
__u64 critical; /* Guest may not get interrupts if == r1 */
|
|
__u64 sprg0;
|
|
__u64 sprg1;
|
|
__u64 sprg2;
|
|
__u64 sprg3;
|
|
__u64 srr0;
|
|
__u64 srr1;
|
|
__u64 dar; /* dear on BookE */
|
|
__u64 msr;
|
|
__u32 dsisr;
|
|
__u32 int_pending; /* Tells the guest if we have an interrupt */
|
|
__u32 sr[16];
|
|
__u32 mas0;
|
|
__u32 mas1;
|
|
__u64 mas7_3;
|
|
__u64 mas2;
|
|
__u32 mas4;
|
|
__u32 mas6;
|
|
__u32 esr;
|
|
__u32 pir;
|
|
|
|
/*
|
|
* SPRG4-7 are user-readable, so we can only keep these consistent
|
|
* between the shared area and the real registers when there's an
|
|
* intervening exit to KVM. This also applies to SPRG3 on some
|
|
* chips.
|
|
*
|
|
* This suffices for access by guest userspace, since in PR-mode
|
|
* KVM, an exit must occur when changing the guest's MSR[PR].
|
|
* If the guest kernel writes to SPRG3-7 via the shared area, it
|
|
* must also use the shared area for reading while in kernel space.
|
|
*/
|
|
__u64 sprg4;
|
|
__u64 sprg5;
|
|
__u64 sprg6;
|
|
__u64 sprg7;
|
|
};
|
|
|
|
#define KVM_SC_MAGIC_R0 0x4b564d21 /* "KVM!" */
|
|
#define HC_VENDOR_KVM (42 << 16)
|
|
#define HC_EV_SUCCESS 0
|
|
#define HC_EV_UNIMPLEMENTED 12
|
|
|
|
#define KVM_FEATURE_MAGIC_PAGE 1
|
|
|
|
#define KVM_MAGIC_FEAT_SR (1 << 0)
|
|
|
|
/* MASn, ESR, PIR, and high SPRGs */
|
|
#define KVM_MAGIC_FEAT_MAS0_TO_SPRG7 (1 << 1)
|
|
|
|
#ifdef __KERNEL__
|
|
|
|
#ifdef CONFIG_KVM_GUEST
|
|
|
|
#include <linux/of.h>
|
|
|
|
static inline int kvm_para_available(void)
|
|
{
|
|
struct device_node *hyper_node;
|
|
|
|
hyper_node = of_find_node_by_path("/hypervisor");
|
|
if (!hyper_node)
|
|
return 0;
|
|
|
|
if (!of_device_is_compatible(hyper_node, "linux,kvm"))
|
|
return 0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
extern unsigned long kvm_hypercall(unsigned long *in,
|
|
unsigned long *out,
|
|
unsigned long nr);
|
|
|
|
#else
|
|
|
|
static inline int kvm_para_available(void)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static unsigned long kvm_hypercall(unsigned long *in,
|
|
unsigned long *out,
|
|
unsigned long nr)
|
|
{
|
|
return HC_EV_UNIMPLEMENTED;
|
|
}
|
|
|
|
#endif
|
|
|
|
static inline long kvm_hypercall0_1(unsigned int nr, unsigned long *r2)
|
|
{
|
|
unsigned long in[8];
|
|
unsigned long out[8];
|
|
unsigned long r;
|
|
|
|
r = kvm_hypercall(in, out, nr | HC_VENDOR_KVM);
|
|
*r2 = out[0];
|
|
|
|
return r;
|
|
}
|
|
|
|
static inline long kvm_hypercall0(unsigned int nr)
|
|
{
|
|
unsigned long in[8];
|
|
unsigned long out[8];
|
|
|
|
return kvm_hypercall(in, out, nr | HC_VENDOR_KVM);
|
|
}
|
|
|
|
static inline long kvm_hypercall1(unsigned int nr, unsigned long p1)
|
|
{
|
|
unsigned long in[8];
|
|
unsigned long out[8];
|
|
|
|
in[0] = p1;
|
|
return kvm_hypercall(in, out, nr | HC_VENDOR_KVM);
|
|
}
|
|
|
|
static inline long kvm_hypercall2(unsigned int nr, unsigned long p1,
|
|
unsigned long p2)
|
|
{
|
|
unsigned long in[8];
|
|
unsigned long out[8];
|
|
|
|
in[0] = p1;
|
|
in[1] = p2;
|
|
return kvm_hypercall(in, out, nr | HC_VENDOR_KVM);
|
|
}
|
|
|
|
static inline long kvm_hypercall3(unsigned int nr, unsigned long p1,
|
|
unsigned long p2, unsigned long p3)
|
|
{
|
|
unsigned long in[8];
|
|
unsigned long out[8];
|
|
|
|
in[0] = p1;
|
|
in[1] = p2;
|
|
in[2] = p3;
|
|
return kvm_hypercall(in, out, nr | HC_VENDOR_KVM);
|
|
}
|
|
|
|
static inline long kvm_hypercall4(unsigned int nr, unsigned long p1,
|
|
unsigned long p2, unsigned long p3,
|
|
unsigned long p4)
|
|
{
|
|
unsigned long in[8];
|
|
unsigned long out[8];
|
|
|
|
in[0] = p1;
|
|
in[1] = p2;
|
|
in[2] = p3;
|
|
in[3] = p4;
|
|
return kvm_hypercall(in, out, nr | HC_VENDOR_KVM);
|
|
}
|
|
|
|
|
|
static inline unsigned int kvm_arch_para_features(void)
|
|
{
|
|
unsigned long r;
|
|
|
|
if (!kvm_para_available())
|
|
return 0;
|
|
|
|
if(kvm_hypercall0_1(KVM_HC_FEATURES, &r))
|
|
return 0;
|
|
|
|
return r;
|
|
}
|
|
|
|
static inline bool kvm_check_and_clear_guest_paused(void)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
#endif /* __KERNEL__ */
|
|
|
|
#endif /* __POWERPC_KVM_PARA_H__ */
|