kvm_host.h 6.4 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14
/* SPDX-License-Identifier: GPL-2.0-only */
/*
 * Copyright (C) 2019 Western Digital Corporation or its affiliates.
 *
 * Authors:
 *     Anup Patel <anup.patel@wdc.com>
 */

#ifndef __RISCV_KVM_HOST_H__
#define __RISCV_KVM_HOST_H__

#include <linux/types.h>
#include <linux/kvm.h>
#include <linux/kvm_types.h>
15
#include <asm/csr.h>
16
#include <asm/kvm_vcpu_fp.h>
A
Atish Patra 已提交
17
#include <asm/kvm_vcpu_timer.h>
18

19 20
#define KVM_MAX_VCPUS			\
	((HGATP_VMID_MASK >> HGATP_VMID_SHIFT) + 1)
21 22 23 24 25 26 27 28

#define KVM_HALT_POLL_NS_DEFAULT	500000

#define KVM_VCPU_MAX_FEATURES		0

#define KVM_REQ_SLEEP \
	KVM_ARCH_REQ_FLAGS(0, KVM_REQUEST_WAIT | KVM_REQUEST_NO_WAKEUP)
#define KVM_REQ_VCPU_RESET		KVM_ARCH_REQ(1)
A
Anup Patel 已提交
29
#define KVM_REQ_UPDATE_HGATP		KVM_ARCH_REQ(2)
30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46

struct kvm_vm_stat {
	struct kvm_vm_stat_generic generic;
};

struct kvm_vcpu_stat {
	struct kvm_vcpu_stat_generic generic;
	u64 ecall_exit_stat;
	u64 wfi_exit_stat;
	u64 mmio_exit_user;
	u64 mmio_exit_kernel;
	u64 exits;
};

struct kvm_arch_memory_slot {
};

A
Anup Patel 已提交
47 48 49 50 51 52 53 54 55
struct kvm_vmid {
	/*
	 * Writes to vmid_version and vmid happen with vmid_lock held
	 * whereas reads happen without any lock held.
	 */
	unsigned long vmid_version;
	unsigned long vmid;
};

56
struct kvm_arch {
A
Anup Patel 已提交
57 58 59
	/* stage2 vmid */
	struct kvm_vmid vmid;

60 61 62
	/* stage2 page table */
	pgd_t *pgd;
	phys_addr_t pgd_phys;
A
Atish Patra 已提交
63 64 65

	/* Guest Timer */
	struct kvm_guest_timer timer;
66 67
};

68 69 70 71 72 73 74 75
struct kvm_mmio_decode {
	unsigned long insn;
	int insn_len;
	int len;
	int shift;
	int return_handled;
};

A
Atish Patra 已提交
76 77 78 79
struct kvm_sbi_context {
	int return_handled;
};

80 81 82 83 84 85 86 87
struct kvm_cpu_trap {
	unsigned long sepc;
	unsigned long scause;
	unsigned long stval;
	unsigned long htval;
	unsigned long htinst;
};

88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123
struct kvm_cpu_context {
	unsigned long zero;
	unsigned long ra;
	unsigned long sp;
	unsigned long gp;
	unsigned long tp;
	unsigned long t0;
	unsigned long t1;
	unsigned long t2;
	unsigned long s0;
	unsigned long s1;
	unsigned long a0;
	unsigned long a1;
	unsigned long a2;
	unsigned long a3;
	unsigned long a4;
	unsigned long a5;
	unsigned long a6;
	unsigned long a7;
	unsigned long s2;
	unsigned long s3;
	unsigned long s4;
	unsigned long s5;
	unsigned long s6;
	unsigned long s7;
	unsigned long s8;
	unsigned long s9;
	unsigned long s10;
	unsigned long s11;
	unsigned long t3;
	unsigned long t4;
	unsigned long t5;
	unsigned long t6;
	unsigned long sepc;
	unsigned long sstatus;
	unsigned long hstatus;
A
Atish Patra 已提交
124
	union __riscv_fp_state fp;
125 126 127 128 129 130 131 132 133 134 135 136 137 138 139
};

struct kvm_vcpu_csr {
	unsigned long vsstatus;
	unsigned long vsie;
	unsigned long vstvec;
	unsigned long vsscratch;
	unsigned long vsepc;
	unsigned long vscause;
	unsigned long vstval;
	unsigned long hvip;
	unsigned long vsatp;
	unsigned long scounteren;
};

140
struct kvm_vcpu_arch {
141 142 143 144 145 146
	/* VCPU ran at least once */
	bool ran_atleast_once;

	/* ISA feature bits (similar to MISA) */
	unsigned long isa;

147 148 149 150 151 152 153 154
	/* SSCRATCH, STVEC, and SCOUNTEREN of Host */
	unsigned long host_sscratch;
	unsigned long host_stvec;
	unsigned long host_scounteren;

	/* CPU context of Host */
	struct kvm_cpu_context host_context;

155 156 157 158 159 160 161 162 163 164 165 166
	/* CPU context of Guest VCPU */
	struct kvm_cpu_context guest_context;

	/* CPU CSR context of Guest VCPU */
	struct kvm_vcpu_csr guest_csr;

	/* CPU context upon Guest VCPU reset */
	struct kvm_cpu_context guest_reset_context;

	/* CPU CSR context upon Guest VCPU reset */
	struct kvm_vcpu_csr guest_reset_csr;

167 168 169 170 171 172 173 174 175 176 177 178
	/*
	 * VCPU interrupts
	 *
	 * We have a lockless approach for tracking pending VCPU interrupts
	 * implemented using atomic bitops. The irqs_pending bitmap represent
	 * pending interrupts whereas irqs_pending_mask represent bits changed
	 * in irqs_pending. Our approach is modeled around multiple producer
	 * and single consumer problem where the consumer is the VCPU itself.
	 */
	unsigned long irqs_pending;
	unsigned long irqs_pending_mask;

A
Atish Patra 已提交
179 180 181
	/* VCPU Timer */
	struct kvm_vcpu_timer timer;

182 183 184
	/* MMIO instruction details */
	struct kvm_mmio_decode mmio_decode;

A
Atish Patra 已提交
185 186 187
	/* SBI context */
	struct kvm_sbi_context sbi_context;

188
	/* Cache pages needed to program page tables with spinlock held */
189
	struct kvm_mmu_memory_cache mmu_page_cache;
190

191 192 193
	/* VCPU power-off state */
	bool power_off;

194 195 196 197 198 199 200 201 202 203 204
	/* Don't run the VCPU (blocked) */
	bool pause;

	/* SRCU lock index for in-kernel run loop */
	int srcu_idx;
};

static inline void kvm_arch_hardware_unsetup(void) {}
static inline void kvm_arch_sync_events(struct kvm *kvm) {}
static inline void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) {}

A
Anup Patel 已提交
205 206
#define KVM_ARCH_WANT_MMU_NOTIFIER

207 208
void __kvm_riscv_hfence_gvma_vmid_gpa(unsigned long gpa_divby_4,
				      unsigned long vmid);
A
Anup Patel 已提交
209
void __kvm_riscv_hfence_gvma_vmid(unsigned long vmid);
210
void __kvm_riscv_hfence_gvma_gpa(unsigned long gpa_divby_4);
A
Anup Patel 已提交
211 212
void __kvm_riscv_hfence_gvma_all(void);

213 214 215
int kvm_riscv_stage2_map(struct kvm_vcpu *vcpu,
			 struct kvm_memory_slot *memslot,
			 gpa_t gpa, unsigned long hva, bool is_write);
216 217 218
int kvm_riscv_stage2_alloc_pgd(struct kvm *kvm);
void kvm_riscv_stage2_free_pgd(struct kvm *kvm);
void kvm_riscv_stage2_update_hgatp(struct kvm_vcpu *vcpu);
219 220
void kvm_riscv_stage2_mode_detect(void);
unsigned long kvm_riscv_stage2_mode(void);
221
int kvm_riscv_stage2_gpa_bits(void);
A
Anup Patel 已提交
222 223 224 225 226 227

void kvm_riscv_stage2_vmid_detect(void);
unsigned long kvm_riscv_stage2_vmid_bits(void);
int kvm_riscv_stage2_vmid_init(struct kvm *kvm);
bool kvm_riscv_stage2_vmid_ver_changed(struct kvm_vmid *vmid);
void kvm_riscv_stage2_vmid_update(struct kvm_vcpu *vcpu);
228

229 230
void __kvm_riscv_unpriv_trap(void);

231
void kvm_riscv_vcpu_wfi(struct kvm_vcpu *vcpu);
232 233 234 235 236 237
unsigned long kvm_riscv_vcpu_unpriv_read(struct kvm_vcpu *vcpu,
					 bool read_insn,
					 unsigned long guest_addr,
					 struct kvm_cpu_trap *trap);
void kvm_riscv_vcpu_trap_redirect(struct kvm_vcpu *vcpu,
				  struct kvm_cpu_trap *trap);
238 239 240 241
int kvm_riscv_vcpu_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run);
int kvm_riscv_vcpu_exit(struct kvm_vcpu *vcpu, struct kvm_run *run,
			struct kvm_cpu_trap *trap);

242
void __kvm_riscv_switch_to(struct kvm_vcpu_arch *vcpu_arch);
243

244 245 246 247 248 249 250 251
int kvm_riscv_vcpu_set_interrupt(struct kvm_vcpu *vcpu, unsigned int irq);
int kvm_riscv_vcpu_unset_interrupt(struct kvm_vcpu *vcpu, unsigned int irq);
void kvm_riscv_vcpu_flush_interrupts(struct kvm_vcpu *vcpu);
void kvm_riscv_vcpu_sync_interrupts(struct kvm_vcpu *vcpu);
bool kvm_riscv_vcpu_has_interrupts(struct kvm_vcpu *vcpu, unsigned long mask);
void kvm_riscv_vcpu_power_off(struct kvm_vcpu *vcpu);
void kvm_riscv_vcpu_power_on(struct kvm_vcpu *vcpu);

A
Atish Patra 已提交
252 253 254
int kvm_riscv_vcpu_sbi_return(struct kvm_vcpu *vcpu, struct kvm_run *run);
int kvm_riscv_vcpu_sbi_ecall(struct kvm_vcpu *vcpu, struct kvm_run *run);

255
#endif /* __RISCV_KVM_HOST_H__ */