kvm_para.h 4.7 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22
/*
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License, version 2, as
 * published by the Free Software Foundation.
 *
 * This program is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program; if not, write to the Free Software
 * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301, USA.
 *
 * Copyright IBM Corp. 2008
 *
 * Authors: Hollis Blanchard <hollisb@us.ibm.com>
 */

#ifndef __POWERPC_KVM_PARA_H__
#define __POWERPC_KVM_PARA_H__

A
Alexander Graf 已提交
23 24
#include <linux/types.h>

25 26 27 28 29 30 31 32 33 34
/*
 * Additions to this struct must only occur at the end, and should be
 * accompanied by a KVM_MAGIC_FEAT flag to advertise that they are present
 * (albeit not necessarily relevant to the current target hardware platform).
 *
 * Struct fields are always 32 or 64 bit aligned, depending on them being 32
 * or 64 bit wide respectively.
 *
 * See Documentation/virtual/kvm/ppc-pv.txt
 */
A
Alexander Graf 已提交
35
struct kvm_vcpu_arch_shared {
36 37 38
	__u64 scratch1;
	__u64 scratch2;
	__u64 scratch3;
39
	__u64 critical;		/* Guest may not get interrupts if == r1 */
40 41 42 43
	__u64 sprg0;
	__u64 sprg1;
	__u64 sprg2;
	__u64 sprg3;
44 45
	__u64 srr0;
	__u64 srr1;
46
	__u64 dar;		/* dear on BookE */
47
	__u64 msr;
48
	__u32 dsisr;
49
	__u32 int_pending;	/* Tells the guest if we have an interrupt */
50
	__u32 sr[16];
51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74
	__u32 mas0;
	__u32 mas1;
	__u64 mas7_3;
	__u64 mas2;
	__u32 mas4;
	__u32 mas6;
	__u32 esr;
	__u32 pir;

	/*
	 * SPRG4-7 are user-readable, so we can only keep these consistent
	 * between the shared area and the real registers when there's an
	 * intervening exit to KVM.  This also applies to SPRG3 on some
	 * chips.
	 *
	 * This suffices for access by guest userspace, since in PR-mode
	 * KVM, an exit must occur when changing the guest's MSR[PR].
	 * If the guest kernel writes to SPRG3-7 via the shared area, it
	 * must also use the shared area for reading while in kernel space.
	 */
	__u64 sprg4;
	__u64 sprg5;
	__u64 sprg6;
	__u64 sprg7;
A
Alexander Graf 已提交
75 76
};

77
#define KVM_SC_MAGIC_R0		0x4b564d21 /* "KVM!" */
78 79 80 81

#define KVM_HCALL_TOKEN(num)     _EV_HCALL_TOKEN(EV_KVM_VENDOR_ID, num)

#include <asm/epapr_hcalls.h>
82

83 84
#define KVM_FEATURE_MAGIC_PAGE	1

85 86 87 88
#define KVM_MAGIC_FEAT_SR		(1 << 0)

/* MASn, ESR, PIR, and high SPRGs */
#define KVM_MAGIC_FEAT_MAS0_TO_SPRG7	(1 << 1)
89

90 91
#ifdef __KERNEL__

92 93
#ifdef CONFIG_KVM_GUEST

94 95
#include <linux/of.h>

96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115
static inline int kvm_para_available(void)
{
	struct device_node *hyper_node;

	hyper_node = of_find_node_by_path("/hypervisor");
	if (!hyper_node)
		return 0;

	if (!of_device_is_compatible(hyper_node, "linux,kvm"))
		return 0;

	return 1;
}

extern unsigned long kvm_hypercall(unsigned long *in,
				   unsigned long *out,
				   unsigned long nr);

#else

116 117 118 119 120
static inline int kvm_para_available(void)
{
	return 0;
}

121 122 123 124
static unsigned long kvm_hypercall(unsigned long *in,
				   unsigned long *out,
				   unsigned long nr)
{
125
	return EV_UNIMPLEMENTED;
126 127 128 129 130 131 132 133 134 135
}

#endif

static inline long kvm_hypercall0_1(unsigned int nr, unsigned long *r2)
{
	unsigned long in[8];
	unsigned long out[8];
	unsigned long r;

136
	r = kvm_hypercall(in, out, KVM_HCALL_TOKEN(nr));
137 138 139 140 141 142 143 144 145 146
	*r2 = out[0];

	return r;
}

static inline long kvm_hypercall0(unsigned int nr)
{
	unsigned long in[8];
	unsigned long out[8];

147
	return kvm_hypercall(in, out, KVM_HCALL_TOKEN(nr));
148 149 150 151 152 153 154 155
}

static inline long kvm_hypercall1(unsigned int nr, unsigned long p1)
{
	unsigned long in[8];
	unsigned long out[8];

	in[0] = p1;
156
	return kvm_hypercall(in, out, KVM_HCALL_TOKEN(nr));
157 158 159 160 161 162 163 164 165 166
}

static inline long kvm_hypercall2(unsigned int nr, unsigned long p1,
				  unsigned long p2)
{
	unsigned long in[8];
	unsigned long out[8];

	in[0] = p1;
	in[1] = p2;
167
	return kvm_hypercall(in, out, KVM_HCALL_TOKEN(nr));
168 169 170 171 172 173 174 175 176 177 178
}

static inline long kvm_hypercall3(unsigned int nr, unsigned long p1,
				  unsigned long p2, unsigned long p3)
{
	unsigned long in[8];
	unsigned long out[8];

	in[0] = p1;
	in[1] = p2;
	in[2] = p3;
179
	return kvm_hypercall(in, out, KVM_HCALL_TOKEN(nr));
180 181 182 183 184 185 186 187 188 189 190 191 192
}

static inline long kvm_hypercall4(unsigned int nr, unsigned long p1,
				  unsigned long p2, unsigned long p3,
				  unsigned long p4)
{
	unsigned long in[8];
	unsigned long out[8];

	in[0] = p1;
	in[1] = p2;
	in[2] = p3;
	in[3] = p4;
193
	return kvm_hypercall(in, out, KVM_HCALL_TOKEN(nr));
194 195 196
}


197 198
static inline unsigned int kvm_arch_para_features(void)
{
199 200 201 202 203 204 205 206 207
	unsigned long r;

	if (!kvm_para_available())
		return 0;

	if(kvm_hypercall0_1(KVM_HC_FEATURES, &r))
		return 0;

	return r;
208 209
}

210 211 212 213 214
static inline bool kvm_check_and_clear_guest_paused(void)
{
	return false;
}

215 216 217
#endif /* __KERNEL__ */

#endif /* __POWERPC_KVM_PARA_H__ */