forked from mirrors/linux
		
	MIPS: Remove KVM_TE support
After removal of the guest part of KVM TE (trap and emulate), also remove the host part. Signed-off-by: Thomas Bogendoerfer <tsbogend@alpha.franken.de>
This commit is contained in:
		
							parent
							
								
									a1515ec720
								
							
						
					
					
						commit
						45c7e8af4a
					
				
					 16 changed files with 23 additions and 4262 deletions
				
			
		| 
						 | 
					@ -39,7 +39,6 @@ CONFIG_MIPS32_O32=y
 | 
				
			||||||
CONFIG_MIPS32_N32=y
 | 
					CONFIG_MIPS32_N32=y
 | 
				
			||||||
CONFIG_VIRTUALIZATION=y
 | 
					CONFIG_VIRTUALIZATION=y
 | 
				
			||||||
CONFIG_KVM=m
 | 
					CONFIG_KVM=m
 | 
				
			||||||
CONFIG_KVM_MIPS_VZ=y
 | 
					 | 
				
			||||||
CONFIG_MODULES=y
 | 
					CONFIG_MODULES=y
 | 
				
			||||||
CONFIG_MODULE_FORCE_LOAD=y
 | 
					CONFIG_MODULE_FORCE_LOAD=y
 | 
				
			||||||
CONFIG_MODULE_UNLOAD=y
 | 
					CONFIG_MODULE_UNLOAD=y
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -88,44 +88,10 @@
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define KVM_HALT_POLL_NS_DEFAULT 500000
 | 
					#define KVM_HALT_POLL_NS_DEFAULT 500000
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
extern unsigned long GUESTID_MASK;
 | 
					extern unsigned long GUESTID_MASK;
 | 
				
			||||||
extern unsigned long GUESTID_FIRST_VERSION;
 | 
					extern unsigned long GUESTID_FIRST_VERSION;
 | 
				
			||||||
extern unsigned long GUESTID_VERSION_MASK;
 | 
					extern unsigned long GUESTID_VERSION_MASK;
 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					 | 
				
			||||||
/*
 | 
					 | 
				
			||||||
 * Special address that contains the comm page, used for reducing # of traps
 | 
					 | 
				
			||||||
 * This needs to be within 32Kb of 0x0 (so the zero register can be used), but
 | 
					 | 
				
			||||||
 * preferably not at 0x0 so that most kernel NULL pointer dereferences can be
 | 
					 | 
				
			||||||
 * caught.
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
#define KVM_GUEST_COMMPAGE_ADDR		((PAGE_SIZE > 0x8000) ?	0 : \
 | 
					 | 
				
			||||||
					 (0x8000 - PAGE_SIZE))
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#define KVM_GUEST_KERNEL_MODE(vcpu)	((kvm_read_c0_guest_status(vcpu->arch.cop0) & (ST0_EXL | ST0_ERL)) || \
 | 
					 | 
				
			||||||
					((kvm_read_c0_guest_status(vcpu->arch.cop0) & KSU_USER) == 0))
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#define KVM_GUEST_KUSEG			0x00000000UL
 | 
					 | 
				
			||||||
#define KVM_GUEST_KSEG0			0x40000000UL
 | 
					 | 
				
			||||||
#define KVM_GUEST_KSEG1			0x40000000UL
 | 
					 | 
				
			||||||
#define KVM_GUEST_KSEG23		0x60000000UL
 | 
					 | 
				
			||||||
#define KVM_GUEST_KSEGX(a)		((_ACAST32_(a)) & 0xe0000000)
 | 
					 | 
				
			||||||
#define KVM_GUEST_CPHYSADDR(a)		((_ACAST32_(a)) & 0x1fffffff)
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#define KVM_GUEST_CKSEG0ADDR(a)		(KVM_GUEST_CPHYSADDR(a) | KVM_GUEST_KSEG0)
 | 
					 | 
				
			||||||
#define KVM_GUEST_CKSEG1ADDR(a)		(KVM_GUEST_CPHYSADDR(a) | KVM_GUEST_KSEG1)
 | 
					 | 
				
			||||||
#define KVM_GUEST_CKSEG23ADDR(a)	(KVM_GUEST_CPHYSADDR(a) | KVM_GUEST_KSEG23)
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/*
 | 
					 | 
				
			||||||
 * Map an address to a certain kernel segment
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
#define KVM_GUEST_KSEG0ADDR(a)		(KVM_GUEST_CPHYSADDR(a) | KVM_GUEST_KSEG0)
 | 
					 | 
				
			||||||
#define KVM_GUEST_KSEG1ADDR(a)		(KVM_GUEST_CPHYSADDR(a) | KVM_GUEST_KSEG1)
 | 
					 | 
				
			||||||
#define KVM_GUEST_KSEG23ADDR(a)		(KVM_GUEST_CPHYSADDR(a) | KVM_GUEST_KSEG23)
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#define KVM_INVALID_PAGE		0xdeadbeef
 | 
					 | 
				
			||||||
#define KVM_INVALID_ADDR		0xdeadbeef
 | 
					#define KVM_INVALID_ADDR		0xdeadbeef
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/*
 | 
					/*
 | 
				
			||||||
| 
						 | 
					@ -165,7 +131,6 @@ struct kvm_vcpu_stat {
 | 
				
			||||||
	u64 fpe_exits;
 | 
						u64 fpe_exits;
 | 
				
			||||||
	u64 msa_disabled_exits;
 | 
						u64 msa_disabled_exits;
 | 
				
			||||||
	u64 flush_dcache_exits;
 | 
						u64 flush_dcache_exits;
 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
	u64 vz_gpsi_exits;
 | 
						u64 vz_gpsi_exits;
 | 
				
			||||||
	u64 vz_gsfc_exits;
 | 
						u64 vz_gsfc_exits;
 | 
				
			||||||
	u64 vz_hc_exits;
 | 
						u64 vz_hc_exits;
 | 
				
			||||||
| 
						 | 
					@ -176,7 +141,6 @@ struct kvm_vcpu_stat {
 | 
				
			||||||
	u64 vz_resvd_exits;
 | 
						u64 vz_resvd_exits;
 | 
				
			||||||
#ifdef CONFIG_CPU_LOONGSON64
 | 
					#ifdef CONFIG_CPU_LOONGSON64
 | 
				
			||||||
	u64 vz_cpucfg_exits;
 | 
						u64 vz_cpucfg_exits;
 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
	u64 halt_successful_poll;
 | 
						u64 halt_successful_poll;
 | 
				
			||||||
	u64 halt_attempted_poll;
 | 
						u64 halt_attempted_poll;
 | 
				
			||||||
| 
						 | 
					@ -303,14 +267,6 @@ enum emulation_result {
 | 
				
			||||||
	EMULATE_HYPERCALL,	/* HYPCALL instruction */
 | 
						EMULATE_HYPERCALL,	/* HYPCALL instruction */
 | 
				
			||||||
};
 | 
					};
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define mips3_paddr_to_tlbpfn(x) \
 | 
					 | 
				
			||||||
	(((unsigned long)(x) >> MIPS3_PG_SHIFT) & MIPS3_PG_FRAME)
 | 
					 | 
				
			||||||
#define mips3_tlbpfn_to_paddr(x) \
 | 
					 | 
				
			||||||
	((unsigned long)((x) & MIPS3_PG_FRAME) << MIPS3_PG_SHIFT)
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#define MIPS3_PG_SHIFT		6
 | 
					 | 
				
			||||||
#define MIPS3_PG_FRAME		0x3fffffc0
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#if defined(CONFIG_64BIT)
 | 
					#if defined(CONFIG_64BIT)
 | 
				
			||||||
#define VPN2_MASK		GENMASK(cpu_vmbits - 1, 13)
 | 
					#define VPN2_MASK		GENMASK(cpu_vmbits - 1, 13)
 | 
				
			||||||
#else
 | 
					#else
 | 
				
			||||||
| 
						 | 
					@ -337,7 +293,6 @@ struct kvm_mips_tlb {
 | 
				
			||||||
#define KVM_MIPS_AUX_FPU	0x1
 | 
					#define KVM_MIPS_AUX_FPU	0x1
 | 
				
			||||||
#define KVM_MIPS_AUX_MSA	0x2
 | 
					#define KVM_MIPS_AUX_MSA	0x2
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define KVM_MIPS_GUEST_TLB_SIZE	64
 | 
					 | 
				
			||||||
struct kvm_vcpu_arch {
 | 
					struct kvm_vcpu_arch {
 | 
				
			||||||
	void *guest_ebase;
 | 
						void *guest_ebase;
 | 
				
			||||||
	int (*vcpu_run)(struct kvm_vcpu *vcpu);
 | 
						int (*vcpu_run)(struct kvm_vcpu *vcpu);
 | 
				
			||||||
| 
						 | 
					@ -370,9 +325,6 @@ struct kvm_vcpu_arch {
 | 
				
			||||||
	/* COP0 State */
 | 
						/* COP0 State */
 | 
				
			||||||
	struct mips_coproc *cop0;
 | 
						struct mips_coproc *cop0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* Host KSEG0 address of the EI/DI offset */
 | 
					 | 
				
			||||||
	void *kseg0_commpage;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Resume PC after MMIO completion */
 | 
						/* Resume PC after MMIO completion */
 | 
				
			||||||
	unsigned long io_pc;
 | 
						unsigned long io_pc;
 | 
				
			||||||
	/* GPR used as IO source/target */
 | 
						/* GPR used as IO source/target */
 | 
				
			||||||
| 
						 | 
					@ -398,19 +350,9 @@ struct kvm_vcpu_arch {
 | 
				
			||||||
	/* Bitmask of pending exceptions to be cleared */
 | 
						/* Bitmask of pending exceptions to be cleared */
 | 
				
			||||||
	unsigned long pending_exceptions_clr;
 | 
						unsigned long pending_exceptions_clr;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* S/W Based TLB for guest */
 | 
					 | 
				
			||||||
	struct kvm_mips_tlb guest_tlb[KVM_MIPS_GUEST_TLB_SIZE];
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Guest kernel/user [partial] mm */
 | 
					 | 
				
			||||||
	struct mm_struct guest_kernel_mm, guest_user_mm;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Guest ASID of last user mode execution */
 | 
					 | 
				
			||||||
	unsigned int last_user_gasid;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Cache some mmu pages needed inside spinlock regions */
 | 
						/* Cache some mmu pages needed inside spinlock regions */
 | 
				
			||||||
	struct kvm_mmu_memory_cache mmu_page_cache;
 | 
						struct kvm_mmu_memory_cache mmu_page_cache;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
	/* vcpu's vzguestid is different on each host cpu in an smp system */
 | 
						/* vcpu's vzguestid is different on each host cpu in an smp system */
 | 
				
			||||||
	u32 vzguestid[NR_CPUS];
 | 
						u32 vzguestid[NR_CPUS];
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -421,7 +363,6 @@ struct kvm_vcpu_arch {
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* emulated guest MAAR registers */
 | 
						/* emulated guest MAAR registers */
 | 
				
			||||||
	unsigned long maar[6];
 | 
						unsigned long maar[6];
 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* Last CPU the VCPU state was loaded on */
 | 
						/* Last CPU the VCPU state was loaded on */
 | 
				
			||||||
	int last_sched_cpu;
 | 
						int last_sched_cpu;
 | 
				
			||||||
| 
						 | 
					@ -651,20 +592,6 @@ static inline void kvm_change_##name1(struct mips_coproc *cop0,		\
 | 
				
			||||||
	__BUILD_KVM_ATOMIC_SAVED(name, type, _reg, sel)			\
 | 
						__BUILD_KVM_ATOMIC_SAVED(name, type, _reg, sel)			\
 | 
				
			||||||
	__BUILD_KVM_SET_WRAP(c0_guest_##name, sw_gc0_##name, type)
 | 
						__BUILD_KVM_SET_WRAP(c0_guest_##name, sw_gc0_##name, type)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifndef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/*
 | 
					 | 
				
			||||||
 * T&E (trap & emulate software based virtualisation)
 | 
					 | 
				
			||||||
 * We generate the common accessors operating exclusively on the saved context
 | 
					 | 
				
			||||||
 * in RAM.
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#define __BUILD_KVM_RW_HW	__BUILD_KVM_RW_SW
 | 
					 | 
				
			||||||
#define __BUILD_KVM_SET_HW	__BUILD_KVM_SET_SW
 | 
					 | 
				
			||||||
#define __BUILD_KVM_ATOMIC_HW	__BUILD_KVM_ATOMIC_SW
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#else
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/*
 | 
					/*
 | 
				
			||||||
 * VZ (hardware assisted virtualisation)
 | 
					 * VZ (hardware assisted virtualisation)
 | 
				
			||||||
 * These macros use the active guest state in VZ mode (hardware registers),
 | 
					 * These macros use the active guest state in VZ mode (hardware registers),
 | 
				
			||||||
| 
						 | 
					@ -697,8 +624,6 @@ static inline void kvm_change_##name1(struct mips_coproc *cop0,		\
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
#define __BUILD_KVM_ATOMIC_HW	__BUILD_KVM_SET_HW
 | 
					#define __BUILD_KVM_ATOMIC_HW	__BUILD_KVM_SET_HW
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/*
 | 
					/*
 | 
				
			||||||
 * Define accessors for CP0 registers that are accessible to the guest. These
 | 
					 * Define accessors for CP0 registers that are accessible to the guest. These
 | 
				
			||||||
 * are primarily used by common emulation code, which may need to access the
 | 
					 * are primarily used by common emulation code, which may need to access the
 | 
				
			||||||
| 
						 | 
					@ -874,42 +799,9 @@ void kvm_drop_fpu(struct kvm_vcpu *vcpu);
 | 
				
			||||||
void kvm_lose_fpu(struct kvm_vcpu *vcpu);
 | 
					void kvm_lose_fpu(struct kvm_vcpu *vcpu);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/* TLB handling */
 | 
					/* TLB handling */
 | 
				
			||||||
u32 kvm_get_kernel_asid(struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
u32 kvm_get_user_asid(struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
u32 kvm_get_commpage_asid (struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
int kvm_mips_handle_vz_root_tlb_fault(unsigned long badvaddr,
 | 
					int kvm_mips_handle_vz_root_tlb_fault(unsigned long badvaddr,
 | 
				
			||||||
				      struct kvm_vcpu *vcpu, bool write_fault);
 | 
									      struct kvm_vcpu *vcpu, bool write_fault);
 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
extern int kvm_mips_handle_kseg0_tlb_fault(unsigned long badbaddr,
 | 
					 | 
				
			||||||
					   struct kvm_vcpu *vcpu,
 | 
					 | 
				
			||||||
					   bool write_fault);
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
extern int kvm_mips_handle_commpage_tlb_fault(unsigned long badvaddr,
 | 
					 | 
				
			||||||
					      struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern int kvm_mips_handle_mapped_seg_tlb_fault(struct kvm_vcpu *vcpu,
 | 
					 | 
				
			||||||
						struct kvm_mips_tlb *tlb,
 | 
					 | 
				
			||||||
						unsigned long gva,
 | 
					 | 
				
			||||||
						bool write_fault);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_handle_tlbmiss(u32 cause,
 | 
					 | 
				
			||||||
						     u32 *opc,
 | 
					 | 
				
			||||||
						     struct kvm_vcpu *vcpu,
 | 
					 | 
				
			||||||
						     bool write_fault);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern void kvm_mips_dump_host_tlbs(void);
 | 
					 | 
				
			||||||
extern void kvm_mips_dump_guest_tlbs(struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
extern int kvm_mips_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long entryhi,
 | 
					 | 
				
			||||||
				 bool user, bool kernel);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern int kvm_mips_guest_tlb_lookup(struct kvm_vcpu *vcpu,
 | 
					 | 
				
			||||||
				     unsigned long entryhi);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
int kvm_vz_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long entryhi);
 | 
					int kvm_vz_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long entryhi);
 | 
				
			||||||
int kvm_vz_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long gva,
 | 
					int kvm_vz_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long gva,
 | 
				
			||||||
			    unsigned long *gpa);
 | 
								    unsigned long *gpa);
 | 
				
			||||||
| 
						 | 
					@ -923,48 +815,13 @@ void kvm_vz_load_guesttlb(const struct kvm_mips_tlb *buf, unsigned int index,
 | 
				
			||||||
void kvm_loongson_clear_guest_vtlb(void);
 | 
					void kvm_loongson_clear_guest_vtlb(void);
 | 
				
			||||||
void kvm_loongson_clear_guest_ftlb(void);
 | 
					void kvm_loongson_clear_guest_ftlb(void);
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
void kvm_mips_suspend_mm(int cpu);
 | 
					 | 
				
			||||||
void kvm_mips_resume_mm(int cpu);
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
/* MMU handling */
 | 
					/* MMU handling */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/**
 | 
					 | 
				
			||||||
 * enum kvm_mips_flush - Types of MMU flushes.
 | 
					 | 
				
			||||||
 * @KMF_USER:	Flush guest user virtual memory mappings.
 | 
					 | 
				
			||||||
 *		Guest USeg only.
 | 
					 | 
				
			||||||
 * @KMF_KERN:	Flush guest kernel virtual memory mappings.
 | 
					 | 
				
			||||||
 *		Guest USeg and KSeg2/3.
 | 
					 | 
				
			||||||
 * @KMF_GPA:	Flush guest physical memory mappings.
 | 
					 | 
				
			||||||
 *		Also includes KSeg0 if KMF_KERN is set.
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
enum kvm_mips_flush {
 | 
					 | 
				
			||||||
	KMF_USER	= 0x0,
 | 
					 | 
				
			||||||
	KMF_KERN	= 0x1,
 | 
					 | 
				
			||||||
	KMF_GPA		= 0x2,
 | 
					 | 
				
			||||||
};
 | 
					 | 
				
			||||||
void kvm_mips_flush_gva_pt(pgd_t *pgd, enum kvm_mips_flush flags);
 | 
					 | 
				
			||||||
bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn);
 | 
					bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn);
 | 
				
			||||||
int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn);
 | 
					int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn);
 | 
				
			||||||
pgd_t *kvm_pgd_alloc(void);
 | 
					pgd_t *kvm_pgd_alloc(void);
 | 
				
			||||||
void kvm_mmu_free_memory_caches(struct kvm_vcpu *vcpu);
 | 
					void kvm_mmu_free_memory_caches(struct kvm_vcpu *vcpu);
 | 
				
			||||||
void kvm_trap_emul_invalidate_gva(struct kvm_vcpu *vcpu, unsigned long addr,
 | 
					 | 
				
			||||||
				  bool user);
 | 
					 | 
				
			||||||
void kvm_trap_emul_gva_lockless_begin(struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
void kvm_trap_emul_gva_lockless_end(struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
enum kvm_mips_fault_result {
 | 
					 | 
				
			||||||
	KVM_MIPS_MAPPED = 0,
 | 
					 | 
				
			||||||
	KVM_MIPS_GVA,
 | 
					 | 
				
			||||||
	KVM_MIPS_GPA,
 | 
					 | 
				
			||||||
	KVM_MIPS_TLB,
 | 
					 | 
				
			||||||
	KVM_MIPS_TLBINV,
 | 
					 | 
				
			||||||
	KVM_MIPS_TLBMOD,
 | 
					 | 
				
			||||||
};
 | 
					 | 
				
			||||||
enum kvm_mips_fault_result kvm_trap_emul_gva_fault(struct kvm_vcpu *vcpu,
 | 
					 | 
				
			||||||
						   unsigned long gva,
 | 
					 | 
				
			||||||
						   bool write);
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define KVM_ARCH_WANT_MMU_NOTIFIER
 | 
					#define KVM_ARCH_WANT_MMU_NOTIFIER
 | 
				
			||||||
int kvm_unmap_hva_range(struct kvm *kvm,
 | 
					int kvm_unmap_hva_range(struct kvm *kvm,
 | 
				
			||||||
| 
						 | 
					@ -974,7 +831,6 @@ int kvm_age_hva(struct kvm *kvm, unsigned long start, unsigned long end);
 | 
				
			||||||
int kvm_test_age_hva(struct kvm *kvm, unsigned long hva);
 | 
					int kvm_test_age_hva(struct kvm *kvm, unsigned long hva);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/* Emulation */
 | 
					/* Emulation */
 | 
				
			||||||
int kvm_get_inst(u32 *opc, struct kvm_vcpu *vcpu, u32 *out);
 | 
					 | 
				
			||||||
enum emulation_result update_pc(struct kvm_vcpu *vcpu, u32 cause);
 | 
					enum emulation_result update_pc(struct kvm_vcpu *vcpu, u32 cause);
 | 
				
			||||||
int kvm_get_badinstr(u32 *opc, struct kvm_vcpu *vcpu, u32 *out);
 | 
					int kvm_get_badinstr(u32 *opc, struct kvm_vcpu *vcpu, u32 *out);
 | 
				
			||||||
int kvm_get_badinstrp(u32 *opc, struct kvm_vcpu *vcpu, u32 *out);
 | 
					int kvm_get_badinstrp(u32 *opc, struct kvm_vcpu *vcpu, u32 *out);
 | 
				
			||||||
| 
						 | 
					@ -1006,68 +862,6 @@ static inline bool kvm_is_ifetch_fault(struct kvm_vcpu_arch *vcpu)
 | 
				
			||||||
	return false;
 | 
						return false;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_inst(u32 cause,
 | 
					 | 
				
			||||||
						   u32 *opc,
 | 
					 | 
				
			||||||
						   struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
long kvm_mips_guest_exception_base(struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_syscall(u32 cause,
 | 
					 | 
				
			||||||
						      u32 *opc,
 | 
					 | 
				
			||||||
						      struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_tlbmiss_ld(u32 cause,
 | 
					 | 
				
			||||||
							 u32 *opc,
 | 
					 | 
				
			||||||
							 struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_tlbinv_ld(u32 cause,
 | 
					 | 
				
			||||||
							u32 *opc,
 | 
					 | 
				
			||||||
							struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_tlbmiss_st(u32 cause,
 | 
					 | 
				
			||||||
							 u32 *opc,
 | 
					 | 
				
			||||||
							 struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_tlbinv_st(u32 cause,
 | 
					 | 
				
			||||||
							u32 *opc,
 | 
					 | 
				
			||||||
							struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_tlbmod(u32 cause,
 | 
					 | 
				
			||||||
						     u32 *opc,
 | 
					 | 
				
			||||||
						     struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_fpu_exc(u32 cause,
 | 
					 | 
				
			||||||
						      u32 *opc,
 | 
					 | 
				
			||||||
						      struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_handle_ri(u32 cause,
 | 
					 | 
				
			||||||
						u32 *opc,
 | 
					 | 
				
			||||||
						struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_ri_exc(u32 cause,
 | 
					 | 
				
			||||||
						     u32 *opc,
 | 
					 | 
				
			||||||
						     struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_bp_exc(u32 cause,
 | 
					 | 
				
			||||||
						     u32 *opc,
 | 
					 | 
				
			||||||
						     struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_trap_exc(u32 cause,
 | 
					 | 
				
			||||||
						       u32 *opc,
 | 
					 | 
				
			||||||
						       struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_msafpe_exc(u32 cause,
 | 
					 | 
				
			||||||
							 u32 *opc,
 | 
					 | 
				
			||||||
							 struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_fpe_exc(u32 cause,
 | 
					 | 
				
			||||||
						      u32 *opc,
 | 
					 | 
				
			||||||
						      struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_emulate_msadis_exc(u32 cause,
 | 
					 | 
				
			||||||
							 u32 *opc,
 | 
					 | 
				
			||||||
							 struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern enum emulation_result kvm_mips_complete_mmio_load(struct kvm_vcpu *vcpu);
 | 
					extern enum emulation_result kvm_mips_complete_mmio_load(struct kvm_vcpu *vcpu);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
u32 kvm_mips_read_count(struct kvm_vcpu *vcpu);
 | 
					u32 kvm_mips_read_count(struct kvm_vcpu *vcpu);
 | 
				
			||||||
| 
						 | 
					@ -1087,26 +881,9 @@ ktime_t kvm_mips_freeze_hrtimer(struct kvm_vcpu *vcpu, u32 *count);
 | 
				
			||||||
int kvm_mips_restore_hrtimer(struct kvm_vcpu *vcpu, ktime_t before,
 | 
					int kvm_mips_restore_hrtimer(struct kvm_vcpu *vcpu, ktime_t before,
 | 
				
			||||||
			     u32 count, int min_drift);
 | 
								     u32 count, int min_drift);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
void kvm_vz_acquire_htimer(struct kvm_vcpu *vcpu);
 | 
					void kvm_vz_acquire_htimer(struct kvm_vcpu *vcpu);
 | 
				
			||||||
void kvm_vz_lose_htimer(struct kvm_vcpu *vcpu);
 | 
					void kvm_vz_lose_htimer(struct kvm_vcpu *vcpu);
 | 
				
			||||||
#else
 | 
					 | 
				
			||||||
static inline void kvm_vz_acquire_htimer(struct kvm_vcpu *vcpu) {}
 | 
					 | 
				
			||||||
static inline void kvm_vz_lose_htimer(struct kvm_vcpu *vcpu) {}
 | 
					 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
enum emulation_result kvm_mips_check_privilege(u32 cause,
 | 
					 | 
				
			||||||
					       u32 *opc,
 | 
					 | 
				
			||||||
					       struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
enum emulation_result kvm_mips_emulate_cache(union mips_instruction inst,
 | 
					 | 
				
			||||||
					     u32 *opc,
 | 
					 | 
				
			||||||
					     u32 cause,
 | 
					 | 
				
			||||||
					     struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
enum emulation_result kvm_mips_emulate_CP0(union mips_instruction inst,
 | 
					 | 
				
			||||||
					   u32 *opc,
 | 
					 | 
				
			||||||
					   u32 cause,
 | 
					 | 
				
			||||||
					   struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
enum emulation_result kvm_mips_emulate_store(union mips_instruction inst,
 | 
					enum emulation_result kvm_mips_emulate_store(union mips_instruction inst,
 | 
				
			||||||
					     u32 cause,
 | 
										     u32 cause,
 | 
				
			||||||
					     struct kvm_vcpu *vcpu);
 | 
										     struct kvm_vcpu *vcpu);
 | 
				
			||||||
| 
						 | 
					@ -1117,27 +894,12 @@ enum emulation_result kvm_mips_emulate_load(union mips_instruction inst,
 | 
				
			||||||
/* COP0 */
 | 
					/* COP0 */
 | 
				
			||||||
enum emulation_result kvm_mips_emul_wait(struct kvm_vcpu *vcpu);
 | 
					enum emulation_result kvm_mips_emul_wait(struct kvm_vcpu *vcpu);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
unsigned int kvm_mips_config1_wrmask(struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
unsigned int kvm_mips_config3_wrmask(struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
unsigned int kvm_mips_config4_wrmask(struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
unsigned int kvm_mips_config5_wrmask(struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/* Hypercalls (hypcall.c) */
 | 
					/* Hypercalls (hypcall.c) */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
enum emulation_result kvm_mips_emul_hypcall(struct kvm_vcpu *vcpu,
 | 
					enum emulation_result kvm_mips_emul_hypcall(struct kvm_vcpu *vcpu,
 | 
				
			||||||
					    union mips_instruction inst);
 | 
										    union mips_instruction inst);
 | 
				
			||||||
int kvm_mips_handle_hypcall(struct kvm_vcpu *vcpu);
 | 
					int kvm_mips_handle_hypcall(struct kvm_vcpu *vcpu);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/* Dynamic binary translation */
 | 
					 | 
				
			||||||
extern int kvm_mips_trans_cache_index(union mips_instruction inst,
 | 
					 | 
				
			||||||
				      u32 *opc, struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
extern int kvm_mips_trans_cache_va(union mips_instruction inst, u32 *opc,
 | 
					 | 
				
			||||||
				   struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
extern int kvm_mips_trans_mfc0(union mips_instruction inst, u32 *opc,
 | 
					 | 
				
			||||||
			       struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
extern int kvm_mips_trans_mtc0(union mips_instruction inst, u32 *opc,
 | 
					 | 
				
			||||||
			       struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/* Misc */
 | 
					/* Misc */
 | 
				
			||||||
extern void kvm_mips_dump_stats(struct kvm_vcpu *vcpu);
 | 
					extern void kvm_mips_dump_stats(struct kvm_vcpu *vcpu);
 | 
				
			||||||
extern unsigned long kvm_mips_get_ramsize(struct kvm *kvm);
 | 
					extern unsigned long kvm_mips_get_ramsize(struct kvm *kvm);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -30,40 +30,6 @@ config KVM
 | 
				
			||||||
	help
 | 
						help
 | 
				
			||||||
	  Support for hosting Guest kernels.
 | 
						  Support for hosting Guest kernels.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
choice
 | 
					 | 
				
			||||||
	prompt "Virtualization mode"
 | 
					 | 
				
			||||||
	depends on KVM
 | 
					 | 
				
			||||||
	default KVM_MIPS_TE
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
config KVM_MIPS_TE
 | 
					 | 
				
			||||||
	bool "Trap & Emulate"
 | 
					 | 
				
			||||||
	depends on CPU_MIPS32_R2
 | 
					 | 
				
			||||||
	help
 | 
					 | 
				
			||||||
	  Use trap and emulate to virtualize 32-bit guests in user mode. This
 | 
					 | 
				
			||||||
	  does not require any special hardware Virtualization support beyond
 | 
					 | 
				
			||||||
	  standard MIPS32 r2 or later, but it does require the guest kernel
 | 
					 | 
				
			||||||
	  to be configured with CONFIG_KVM_GUEST=y so that it resides in the
 | 
					 | 
				
			||||||
	  user address segment.
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
config KVM_MIPS_VZ
 | 
					 | 
				
			||||||
	bool "MIPS Virtualization (VZ) ASE"
 | 
					 | 
				
			||||||
	help
 | 
					 | 
				
			||||||
	  Use the MIPS Virtualization (VZ) ASE to virtualize guests. This
 | 
					 | 
				
			||||||
	  supports running unmodified guest kernels (with CONFIG_KVM_GUEST=n),
 | 
					 | 
				
			||||||
	  but requires hardware support.
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
endchoice
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
config KVM_MIPS_DYN_TRANS
 | 
					 | 
				
			||||||
	bool "KVM/MIPS: Dynamic binary translation to reduce traps"
 | 
					 | 
				
			||||||
	depends on KVM_MIPS_TE
 | 
					 | 
				
			||||||
	default y
 | 
					 | 
				
			||||||
	help
 | 
					 | 
				
			||||||
	  When running in Trap & Emulate mode patch privileged
 | 
					 | 
				
			||||||
	  instructions to reduce the number of traps.
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	  If unsure, say Y.
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
config KVM_MIPS_DEBUG_COP0_COUNTERS
 | 
					config KVM_MIPS_DEBUG_COP0_COUNTERS
 | 
				
			||||||
	bool "Maintain counters for COP0 accesses"
 | 
						bool "Maintain counters for COP0 accesses"
 | 
				
			||||||
	depends on KVM
 | 
						depends on KVM
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -9,7 +9,7 @@ EXTRA_CFLAGS += -Ivirt/kvm -Iarch/mips/kvm
 | 
				
			||||||
common-objs-$(CONFIG_CPU_HAS_MSA) += msa.o
 | 
					common-objs-$(CONFIG_CPU_HAS_MSA) += msa.o
 | 
				
			||||||
 | 
					
 | 
				
			||||||
kvm-objs := $(common-objs-y) mips.o emulate.o entry.o \
 | 
					kvm-objs := $(common-objs-y) mips.o emulate.o entry.o \
 | 
				
			||||||
	    interrupt.o stats.o commpage.o \
 | 
						    interrupt.o stats.o \
 | 
				
			||||||
	    fpu.o
 | 
						    fpu.o
 | 
				
			||||||
kvm-objs += hypcall.o
 | 
					kvm-objs += hypcall.o
 | 
				
			||||||
kvm-objs += mmu.o
 | 
					kvm-objs += mmu.o
 | 
				
			||||||
| 
						 | 
					@ -17,11 +17,6 @@ ifdef CONFIG_CPU_LOONGSON64
 | 
				
			||||||
kvm-objs += loongson_ipi.o
 | 
					kvm-objs += loongson_ipi.o
 | 
				
			||||||
endif
 | 
					endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
kvm-objs		+= vz.o
 | 
					kvm-objs		+= vz.o
 | 
				
			||||||
else
 | 
					 | 
				
			||||||
kvm-objs		+= dyntrans.o
 | 
					 | 
				
			||||||
kvm-objs		+= trap_emul.o
 | 
					 | 
				
			||||||
endif
 | 
					 | 
				
			||||||
obj-$(CONFIG_KVM)	+= kvm.o
 | 
					obj-$(CONFIG_KVM)	+= kvm.o
 | 
				
			||||||
obj-y			+= callback.o tlb.o
 | 
					obj-y			+= callback.o tlb.o
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1,32 +0,0 @@
 | 
				
			||||||
/*
 | 
					 | 
				
			||||||
 * This file is subject to the terms and conditions of the GNU General Public
 | 
					 | 
				
			||||||
 * License.  See the file "COPYING" in the main directory of this archive
 | 
					 | 
				
			||||||
 * for more details.
 | 
					 | 
				
			||||||
 *
 | 
					 | 
				
			||||||
 * commpage, currently used for Virtual COP0 registers.
 | 
					 | 
				
			||||||
 * Mapped into the guest kernel @ KVM_GUEST_COMMPAGE_ADDR.
 | 
					 | 
				
			||||||
 *
 | 
					 | 
				
			||||||
 * Copyright (C) 2012  MIPS Technologies, Inc.  All rights reserved.
 | 
					 | 
				
			||||||
 * Authors: Sanjay Lal <sanjayl@kymasys.com>
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#include <linux/errno.h>
 | 
					 | 
				
			||||||
#include <linux/err.h>
 | 
					 | 
				
			||||||
#include <linux/vmalloc.h>
 | 
					 | 
				
			||||||
#include <linux/fs.h>
 | 
					 | 
				
			||||||
#include <linux/memblock.h>
 | 
					 | 
				
			||||||
#include <asm/page.h>
 | 
					 | 
				
			||||||
#include <asm/cacheflush.h>
 | 
					 | 
				
			||||||
#include <asm/mmu_context.h>
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#include <linux/kvm_host.h>
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#include "commpage.h"
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
void kvm_mips_commpage_init(struct kvm_vcpu *vcpu)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	struct kvm_mips_commpage *page = vcpu->arch.kseg0_commpage;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Specific init values for fields */
 | 
					 | 
				
			||||||
	vcpu->arch.cop0 = &page->cop0;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
| 
						 | 
					@ -1,24 +0,0 @@
 | 
				
			||||||
/*
 | 
					 | 
				
			||||||
 * This file is subject to the terms and conditions of the GNU General Public
 | 
					 | 
				
			||||||
 * License.  See the file "COPYING" in the main directory of this archive
 | 
					 | 
				
			||||||
 * for more details.
 | 
					 | 
				
			||||||
 *
 | 
					 | 
				
			||||||
 * KVM/MIPS: commpage: mapped into get kernel space
 | 
					 | 
				
			||||||
 *
 | 
					 | 
				
			||||||
 * Copyright (C) 2012  MIPS Technologies, Inc.  All rights reserved.
 | 
					 | 
				
			||||||
 * Authors: Sanjay Lal <sanjayl@kymasys.com>
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#ifndef __KVM_MIPS_COMMPAGE_H__
 | 
					 | 
				
			||||||
#define __KVM_MIPS_COMMPAGE_H__
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
struct kvm_mips_commpage {
 | 
					 | 
				
			||||||
	/* COP0 state is mapped into Guest kernel via commpage */
 | 
					 | 
				
			||||||
	struct mips_coproc cop0;
 | 
					 | 
				
			||||||
};
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#define KVM_MIPS_COMM_EIDI_OFFSET       0x0
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern void kvm_mips_commpage_init(struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#endif /* __KVM_MIPS_COMMPAGE_H__ */
 | 
					 | 
				
			||||||
| 
						 | 
					@ -1,143 +0,0 @@
 | 
				
			||||||
/*
 | 
					 | 
				
			||||||
 * This file is subject to the terms and conditions of the GNU General Public
 | 
					 | 
				
			||||||
 * License.  See the file "COPYING" in the main directory of this archive
 | 
					 | 
				
			||||||
 * for more details.
 | 
					 | 
				
			||||||
 *
 | 
					 | 
				
			||||||
 * KVM/MIPS: Binary Patching for privileged instructions, reduces traps.
 | 
					 | 
				
			||||||
 *
 | 
					 | 
				
			||||||
 * Copyright (C) 2012  MIPS Technologies, Inc.  All rights reserved.
 | 
					 | 
				
			||||||
 * Authors: Sanjay Lal <sanjayl@kymasys.com>
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#include <linux/errno.h>
 | 
					 | 
				
			||||||
#include <linux/err.h>
 | 
					 | 
				
			||||||
#include <linux/highmem.h>
 | 
					 | 
				
			||||||
#include <linux/kvm_host.h>
 | 
					 | 
				
			||||||
#include <linux/uaccess.h>
 | 
					 | 
				
			||||||
#include <linux/vmalloc.h>
 | 
					 | 
				
			||||||
#include <linux/fs.h>
 | 
					 | 
				
			||||||
#include <linux/memblock.h>
 | 
					 | 
				
			||||||
#include <asm/cacheflush.h>
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#include "commpage.h"
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/**
 | 
					 | 
				
			||||||
 * kvm_mips_trans_replace() - Replace trapping instruction in guest memory.
 | 
					 | 
				
			||||||
 * @vcpu:	Virtual CPU.
 | 
					 | 
				
			||||||
 * @opc:	PC of instruction to replace.
 | 
					 | 
				
			||||||
 * @replace:	Instruction to write
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
static int kvm_mips_trans_replace(struct kvm_vcpu *vcpu, u32 *opc,
 | 
					 | 
				
			||||||
				  union mips_instruction replace)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	unsigned long vaddr = (unsigned long)opc;
 | 
					 | 
				
			||||||
	int err;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
retry:
 | 
					 | 
				
			||||||
	/* The GVA page table is still active so use the Linux TLB handlers */
 | 
					 | 
				
			||||||
	kvm_trap_emul_gva_lockless_begin(vcpu);
 | 
					 | 
				
			||||||
	err = put_user(replace.word, opc);
 | 
					 | 
				
			||||||
	kvm_trap_emul_gva_lockless_end(vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (unlikely(err)) {
 | 
					 | 
				
			||||||
		/*
 | 
					 | 
				
			||||||
		 * We write protect clean pages in GVA page table so normal
 | 
					 | 
				
			||||||
		 * Linux TLB mod handler doesn't silently dirty the page.
 | 
					 | 
				
			||||||
		 * Its also possible we raced with a GVA invalidation.
 | 
					 | 
				
			||||||
		 * Try to force the page to become dirty.
 | 
					 | 
				
			||||||
		 */
 | 
					 | 
				
			||||||
		err = kvm_trap_emul_gva_fault(vcpu, vaddr, true);
 | 
					 | 
				
			||||||
		if (unlikely(err)) {
 | 
					 | 
				
			||||||
			kvm_info("%s: Address unwriteable: %p\n",
 | 
					 | 
				
			||||||
				 __func__, opc);
 | 
					 | 
				
			||||||
			return -EFAULT;
 | 
					 | 
				
			||||||
		}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		/*
 | 
					 | 
				
			||||||
		 * Try again. This will likely trigger a TLB refill, which will
 | 
					 | 
				
			||||||
		 * fetch the new dirty entry from the GVA page table, which
 | 
					 | 
				
			||||||
		 * should then succeed.
 | 
					 | 
				
			||||||
		 */
 | 
					 | 
				
			||||||
		goto retry;
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
	__local_flush_icache_user_range(vaddr, vaddr + 4);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return 0;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
int kvm_mips_trans_cache_index(union mips_instruction inst, u32 *opc,
 | 
					 | 
				
			||||||
			       struct kvm_vcpu *vcpu)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	union mips_instruction nop_inst = { 0 };
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Replace the CACHE instruction, with a NOP */
 | 
					 | 
				
			||||||
	return kvm_mips_trans_replace(vcpu, opc, nop_inst);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/*
 | 
					 | 
				
			||||||
 * Address based CACHE instructions are transformed into synci(s). A little
 | 
					 | 
				
			||||||
 * heavy for just D-cache invalidates, but avoids an expensive trap
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
int kvm_mips_trans_cache_va(union mips_instruction inst, u32 *opc,
 | 
					 | 
				
			||||||
			    struct kvm_vcpu *vcpu)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	union mips_instruction synci_inst = { 0 };
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	synci_inst.i_format.opcode = bcond_op;
 | 
					 | 
				
			||||||
	synci_inst.i_format.rs = inst.i_format.rs;
 | 
					 | 
				
			||||||
	synci_inst.i_format.rt = synci_op;
 | 
					 | 
				
			||||||
	if (cpu_has_mips_r6)
 | 
					 | 
				
			||||||
		synci_inst.i_format.simmediate = inst.spec3_format.simmediate;
 | 
					 | 
				
			||||||
	else
 | 
					 | 
				
			||||||
		synci_inst.i_format.simmediate = inst.i_format.simmediate;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return kvm_mips_trans_replace(vcpu, opc, synci_inst);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
int kvm_mips_trans_mfc0(union mips_instruction inst, u32 *opc,
 | 
					 | 
				
			||||||
			struct kvm_vcpu *vcpu)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	union mips_instruction mfc0_inst = { 0 };
 | 
					 | 
				
			||||||
	u32 rd, sel;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	rd = inst.c0r_format.rd;
 | 
					 | 
				
			||||||
	sel = inst.c0r_format.sel;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (rd == MIPS_CP0_ERRCTL && sel == 0) {
 | 
					 | 
				
			||||||
		mfc0_inst.r_format.opcode = spec_op;
 | 
					 | 
				
			||||||
		mfc0_inst.r_format.rd = inst.c0r_format.rt;
 | 
					 | 
				
			||||||
		mfc0_inst.r_format.func = add_op;
 | 
					 | 
				
			||||||
	} else {
 | 
					 | 
				
			||||||
		mfc0_inst.i_format.opcode = lw_op;
 | 
					 | 
				
			||||||
		mfc0_inst.i_format.rt = inst.c0r_format.rt;
 | 
					 | 
				
			||||||
		mfc0_inst.i_format.simmediate = KVM_GUEST_COMMPAGE_ADDR |
 | 
					 | 
				
			||||||
			offsetof(struct kvm_mips_commpage, cop0.reg[rd][sel]);
 | 
					 | 
				
			||||||
#ifdef CONFIG_CPU_BIG_ENDIAN
 | 
					 | 
				
			||||||
		if (sizeof(vcpu->arch.cop0->reg[0][0]) == 8)
 | 
					 | 
				
			||||||
			mfc0_inst.i_format.simmediate |= 4;
 | 
					 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return kvm_mips_trans_replace(vcpu, opc, mfc0_inst);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
int kvm_mips_trans_mtc0(union mips_instruction inst, u32 *opc,
 | 
					 | 
				
			||||||
			struct kvm_vcpu *vcpu)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	union mips_instruction mtc0_inst = { 0 };
 | 
					 | 
				
			||||||
	u32 rd, sel;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	rd = inst.c0r_format.rd;
 | 
					 | 
				
			||||||
	sel = inst.c0r_format.sel;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	mtc0_inst.i_format.opcode = sw_op;
 | 
					 | 
				
			||||||
	mtc0_inst.i_format.rt = inst.c0r_format.rt;
 | 
					 | 
				
			||||||
	mtc0_inst.i_format.simmediate = KVM_GUEST_COMMPAGE_ADDR |
 | 
					 | 
				
			||||||
		offsetof(struct kvm_mips_commpage, cop0.reg[rd][sel]);
 | 
					 | 
				
			||||||
#ifdef CONFIG_CPU_BIG_ENDIAN
 | 
					 | 
				
			||||||
	if (sizeof(vcpu->arch.cop0->reg[0][0]) == 8)
 | 
					 | 
				
			||||||
		mtc0_inst.i_format.simmediate |= 4;
 | 
					 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return kvm_mips_trans_replace(vcpu, opc, mtc0_inst);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
										
											
												File diff suppressed because it is too large
												Load diff
											
										
									
								
							| 
						 | 
					@ -305,7 +305,6 @@ static void *kvm_mips_build_enter_guest(void *addr)
 | 
				
			||||||
	UASM_i_LW(&p, T0, offsetof(struct kvm_vcpu_arch, pc), K1);
 | 
						UASM_i_LW(&p, T0, offsetof(struct kvm_vcpu_arch, pc), K1);
 | 
				
			||||||
	UASM_i_MTC0(&p, T0, C0_EPC);
 | 
						UASM_i_MTC0(&p, T0, C0_EPC);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
	/* Save normal linux process pgd (VZ guarantees pgd_reg is set) */
 | 
						/* Save normal linux process pgd (VZ guarantees pgd_reg is set) */
 | 
				
			||||||
	if (cpu_has_ldpte)
 | 
						if (cpu_has_ldpte)
 | 
				
			||||||
		UASM_i_MFC0(&p, K0, C0_PWBASE);
 | 
							UASM_i_MFC0(&p, K0, C0_PWBASE);
 | 
				
			||||||
| 
						 | 
					@ -367,21 +366,6 @@ static void *kvm_mips_build_enter_guest(void *addr)
 | 
				
			||||||
	/* Set the root ASID for the Guest */
 | 
						/* Set the root ASID for the Guest */
 | 
				
			||||||
	UASM_i_ADDIU(&p, T1, S0,
 | 
						UASM_i_ADDIU(&p, T1, S0,
 | 
				
			||||||
		     offsetof(struct kvm, arch.gpa_mm.context.asid));
 | 
							     offsetof(struct kvm, arch.gpa_mm.context.asid));
 | 
				
			||||||
#else
 | 
					 | 
				
			||||||
	/* Set the ASID for the Guest Kernel or User */
 | 
					 | 
				
			||||||
	UASM_i_LW(&p, T0, offsetof(struct kvm_vcpu_arch, cop0), K1);
 | 
					 | 
				
			||||||
	UASM_i_LW(&p, T0, offsetof(struct mips_coproc, reg[MIPS_CP0_STATUS][0]),
 | 
					 | 
				
			||||||
		  T0);
 | 
					 | 
				
			||||||
	uasm_i_andi(&p, T0, T0, KSU_USER | ST0_ERL | ST0_EXL);
 | 
					 | 
				
			||||||
	uasm_i_xori(&p, T0, T0, KSU_USER);
 | 
					 | 
				
			||||||
	uasm_il_bnez(&p, &r, T0, label_kernel_asid);
 | 
					 | 
				
			||||||
	 UASM_i_ADDIU(&p, T1, K1, offsetof(struct kvm_vcpu_arch,
 | 
					 | 
				
			||||||
					   guest_kernel_mm.context.asid));
 | 
					 | 
				
			||||||
	/* else user */
 | 
					 | 
				
			||||||
	UASM_i_ADDIU(&p, T1, K1, offsetof(struct kvm_vcpu_arch,
 | 
					 | 
				
			||||||
					  guest_user_mm.context.asid));
 | 
					 | 
				
			||||||
	uasm_l_kernel_asid(&l, p);
 | 
					 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* t1: contains the base of the ASID array, need to get the cpu id  */
 | 
						/* t1: contains the base of the ASID array, need to get the cpu id  */
 | 
				
			||||||
	/* smp_processor_id */
 | 
						/* smp_processor_id */
 | 
				
			||||||
| 
						 | 
					@ -406,24 +390,9 @@ static void *kvm_mips_build_enter_guest(void *addr)
 | 
				
			||||||
	uasm_i_andi(&p, K0, K0, MIPS_ENTRYHI_ASID);
 | 
						uasm_i_andi(&p, K0, K0, MIPS_ENTRYHI_ASID);
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifndef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
	/*
 | 
					 | 
				
			||||||
	 * Set up KVM T&E GVA pgd.
 | 
					 | 
				
			||||||
	 * This does roughly the same as TLBMISS_HANDLER_SETUP_PGD():
 | 
					 | 
				
			||||||
	 * - call tlbmiss_handler_setup_pgd(mm->pgd)
 | 
					 | 
				
			||||||
	 * - but skips write into CP0_PWBase for now
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
	UASM_i_LW(&p, A0, (int)offsetof(struct mm_struct, pgd) -
 | 
					 | 
				
			||||||
			  (int)offsetof(struct mm_struct, context.asid), T1);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	UASM_i_LA(&p, T9, (unsigned long)tlbmiss_handler_setup_pgd);
 | 
					 | 
				
			||||||
	uasm_i_jalr(&p, RA, T9);
 | 
					 | 
				
			||||||
	 uasm_i_mtc0(&p, K0, C0_ENTRYHI);
 | 
					 | 
				
			||||||
#else
 | 
					 | 
				
			||||||
	/* Set up KVM VZ root ASID (!guestid) */
 | 
						/* Set up KVM VZ root ASID (!guestid) */
 | 
				
			||||||
	uasm_i_mtc0(&p, K0, C0_ENTRYHI);
 | 
						uasm_i_mtc0(&p, K0, C0_ENTRYHI);
 | 
				
			||||||
skip_asid_restore:
 | 
					skip_asid_restore:
 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
	uasm_i_ehb(&p);
 | 
						uasm_i_ehb(&p);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* Disable RDHWR access */
 | 
						/* Disable RDHWR access */
 | 
				
			||||||
| 
						 | 
					@ -720,7 +689,6 @@ void *kvm_mips_build_exit(void *addr)
 | 
				
			||||||
		uasm_l_msa_1(&l, p);
 | 
							uasm_l_msa_1(&l, p);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
	/* Restore host ASID */
 | 
						/* Restore host ASID */
 | 
				
			||||||
	if (!cpu_has_guestid) {
 | 
						if (!cpu_has_guestid) {
 | 
				
			||||||
		UASM_i_LW(&p, K0, offsetof(struct kvm_vcpu_arch, host_entryhi),
 | 
							UASM_i_LW(&p, K0, offsetof(struct kvm_vcpu_arch, host_entryhi),
 | 
				
			||||||
| 
						 | 
					@ -764,7 +732,6 @@ void *kvm_mips_build_exit(void *addr)
 | 
				
			||||||
			   MIPS_GCTL1_RID_WIDTH);
 | 
								   MIPS_GCTL1_RID_WIDTH);
 | 
				
			||||||
		uasm_i_mtc0(&p, T0, C0_GUESTCTL1);
 | 
							uasm_i_mtc0(&p, T0, C0_GUESTCTL1);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* Now that the new EBASE has been loaded, unset BEV and KSU_USER */
 | 
						/* Now that the new EBASE has been loaded, unset BEV and KSU_USER */
 | 
				
			||||||
	uasm_i_addiu(&p, AT, ZERO, ~(ST0_EXL | KSU_USER | ST0_IE));
 | 
						uasm_i_addiu(&p, AT, ZERO, ~(ST0_EXL | KSU_USER | ST0_IE));
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -21,119 +21,6 @@
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#include "interrupt.h"
 | 
					#include "interrupt.h"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void kvm_mips_queue_irq(struct kvm_vcpu *vcpu, unsigned int priority)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	set_bit(priority, &vcpu->arch.pending_exceptions);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
void kvm_mips_dequeue_irq(struct kvm_vcpu *vcpu, unsigned int priority)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	clear_bit(priority, &vcpu->arch.pending_exceptions);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
void kvm_mips_queue_timer_int_cb(struct kvm_vcpu *vcpu)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	/*
 | 
					 | 
				
			||||||
	 * Cause bits to reflect the pending timer interrupt,
 | 
					 | 
				
			||||||
	 * the EXC code will be set when we are actually
 | 
					 | 
				
			||||||
	 * delivering the interrupt:
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
	kvm_set_c0_guest_cause(vcpu->arch.cop0, (C_IRQ5 | C_TI));
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Queue up an INT exception for the core */
 | 
					 | 
				
			||||||
	kvm_mips_queue_irq(vcpu, MIPS_EXC_INT_TIMER);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
void kvm_mips_dequeue_timer_int_cb(struct kvm_vcpu *vcpu)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	kvm_clear_c0_guest_cause(vcpu->arch.cop0, (C_IRQ5 | C_TI));
 | 
					 | 
				
			||||||
	kvm_mips_dequeue_irq(vcpu, MIPS_EXC_INT_TIMER);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
void kvm_mips_queue_io_int_cb(struct kvm_vcpu *vcpu,
 | 
					 | 
				
			||||||
			      struct kvm_mips_interrupt *irq)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	int intr = (int)irq->irq;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/*
 | 
					 | 
				
			||||||
	 * Cause bits to reflect the pending IO interrupt,
 | 
					 | 
				
			||||||
	 * the EXC code will be set when we are actually
 | 
					 | 
				
			||||||
	 * delivering the interrupt:
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
	kvm_set_c0_guest_cause(vcpu->arch.cop0, 1 << (intr + 8));
 | 
					 | 
				
			||||||
	kvm_mips_queue_irq(vcpu, kvm_irq_to_priority(intr));
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
void kvm_mips_dequeue_io_int_cb(struct kvm_vcpu *vcpu,
 | 
					 | 
				
			||||||
				struct kvm_mips_interrupt *irq)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	int intr = (int)irq->irq;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	kvm_clear_c0_guest_cause(vcpu->arch.cop0, 1 << (-intr + 8));
 | 
					 | 
				
			||||||
	kvm_mips_dequeue_irq(vcpu, kvm_irq_to_priority(-intr));
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/* Deliver the interrupt of the corresponding priority, if possible. */
 | 
					 | 
				
			||||||
int kvm_mips_irq_deliver_cb(struct kvm_vcpu *vcpu, unsigned int priority,
 | 
					 | 
				
			||||||
			    u32 cause)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	int allowed = 0;
 | 
					 | 
				
			||||||
	u32 exccode, ie;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	struct kvm_vcpu_arch *arch = &vcpu->arch;
 | 
					 | 
				
			||||||
	struct mips_coproc *cop0 = vcpu->arch.cop0;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (priority == MIPS_EXC_MAX)
 | 
					 | 
				
			||||||
		return 0;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	ie = 1 << (kvm_priority_to_irq[priority] + 8);
 | 
					 | 
				
			||||||
	if ((kvm_read_c0_guest_status(cop0) & ST0_IE)
 | 
					 | 
				
			||||||
	    && (!(kvm_read_c0_guest_status(cop0) & (ST0_EXL | ST0_ERL)))
 | 
					 | 
				
			||||||
	    && (kvm_read_c0_guest_status(cop0) & ie)) {
 | 
					 | 
				
			||||||
		allowed = 1;
 | 
					 | 
				
			||||||
		exccode = EXCCODE_INT;
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Are we allowed to deliver the interrupt ??? */
 | 
					 | 
				
			||||||
	if (allowed) {
 | 
					 | 
				
			||||||
		if ((kvm_read_c0_guest_status(cop0) & ST0_EXL) == 0) {
 | 
					 | 
				
			||||||
			/* save old pc */
 | 
					 | 
				
			||||||
			kvm_write_c0_guest_epc(cop0, arch->pc);
 | 
					 | 
				
			||||||
			kvm_set_c0_guest_status(cop0, ST0_EXL);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
			if (cause & CAUSEF_BD)
 | 
					 | 
				
			||||||
				kvm_set_c0_guest_cause(cop0, CAUSEF_BD);
 | 
					 | 
				
			||||||
			else
 | 
					 | 
				
			||||||
				kvm_clear_c0_guest_cause(cop0, CAUSEF_BD);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
			kvm_debug("Delivering INT @ pc %#lx\n", arch->pc);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		} else
 | 
					 | 
				
			||||||
			kvm_err("Trying to deliver interrupt when EXL is already set\n");
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		kvm_change_c0_guest_cause(cop0, CAUSEF_EXCCODE,
 | 
					 | 
				
			||||||
					  (exccode << CAUSEB_EXCCODE));
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		/* XXXSL Set PC to the interrupt exception entry point */
 | 
					 | 
				
			||||||
		arch->pc = kvm_mips_guest_exception_base(vcpu);
 | 
					 | 
				
			||||||
		if (kvm_read_c0_guest_cause(cop0) & CAUSEF_IV)
 | 
					 | 
				
			||||||
			arch->pc += 0x200;
 | 
					 | 
				
			||||||
		else
 | 
					 | 
				
			||||||
			arch->pc += 0x180;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		clear_bit(priority, &vcpu->arch.pending_exceptions);
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return allowed;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
int kvm_mips_irq_clear_cb(struct kvm_vcpu *vcpu, unsigned int priority,
 | 
					 | 
				
			||||||
			  u32 cause)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	return 1;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
void kvm_mips_deliver_interrupts(struct kvm_vcpu *vcpu, u32 cause)
 | 
					void kvm_mips_deliver_interrupts(struct kvm_vcpu *vcpu, u32 cause)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	unsigned long *pending = &vcpu->arch.pending_exceptions;
 | 
						unsigned long *pending = &vcpu->arch.pending_exceptions;
 | 
				
			||||||
| 
						 | 
					@ -145,10 +32,7 @@ void kvm_mips_deliver_interrupts(struct kvm_vcpu *vcpu, u32 cause)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	priority = __ffs(*pending_clr);
 | 
						priority = __ffs(*pending_clr);
 | 
				
			||||||
	while (priority <= MIPS_EXC_MAX) {
 | 
						while (priority <= MIPS_EXC_MAX) {
 | 
				
			||||||
		if (kvm_mips_callbacks->irq_clear(vcpu, priority, cause)) {
 | 
							kvm_mips_callbacks->irq_clear(vcpu, priority, cause);
 | 
				
			||||||
			if (!KVM_MIPS_IRQ_CLEAR_ALL_AT_ONCE)
 | 
					 | 
				
			||||||
				break;
 | 
					 | 
				
			||||||
		}
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
		priority = find_next_bit(pending_clr,
 | 
							priority = find_next_bit(pending_clr,
 | 
				
			||||||
					 BITS_PER_BYTE * sizeof(*pending_clr),
 | 
										 BITS_PER_BYTE * sizeof(*pending_clr),
 | 
				
			||||||
| 
						 | 
					@ -157,10 +41,7 @@ void kvm_mips_deliver_interrupts(struct kvm_vcpu *vcpu, u32 cause)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	priority = __ffs(*pending);
 | 
						priority = __ffs(*pending);
 | 
				
			||||||
	while (priority <= MIPS_EXC_MAX) {
 | 
						while (priority <= MIPS_EXC_MAX) {
 | 
				
			||||||
		if (kvm_mips_callbacks->irq_deliver(vcpu, priority, cause)) {
 | 
							kvm_mips_callbacks->irq_deliver(vcpu, priority, cause);
 | 
				
			||||||
			if (!KVM_MIPS_IRQ_DELIVER_ALL_AT_ONCE)
 | 
					 | 
				
			||||||
				break;
 | 
					 | 
				
			||||||
		}
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
		priority = find_next_bit(pending,
 | 
							priority = find_next_bit(pending,
 | 
				
			||||||
					 BITS_PER_BYTE * sizeof(*pending),
 | 
										 BITS_PER_BYTE * sizeof(*pending),
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -31,29 +31,9 @@
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define C_TI        (_ULCAST_(1) << 30)
 | 
					#define C_TI        (_ULCAST_(1) << 30)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
#define KVM_MIPS_IRQ_DELIVER_ALL_AT_ONCE (1)
 | 
					 | 
				
			||||||
#define KVM_MIPS_IRQ_CLEAR_ALL_AT_ONCE   (1)
 | 
					 | 
				
			||||||
#else
 | 
					 | 
				
			||||||
#define KVM_MIPS_IRQ_DELIVER_ALL_AT_ONCE (0)
 | 
					 | 
				
			||||||
#define KVM_MIPS_IRQ_CLEAR_ALL_AT_ONCE   (0)
 | 
					 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
extern u32 *kvm_priority_to_irq;
 | 
					extern u32 *kvm_priority_to_irq;
 | 
				
			||||||
u32 kvm_irq_to_priority(u32 irq);
 | 
					u32 kvm_irq_to_priority(u32 irq);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void kvm_mips_queue_irq(struct kvm_vcpu *vcpu, unsigned int priority);
 | 
					 | 
				
			||||||
void kvm_mips_dequeue_irq(struct kvm_vcpu *vcpu, unsigned int priority);
 | 
					 | 
				
			||||||
int kvm_mips_pending_timer(struct kvm_vcpu *vcpu);
 | 
					int kvm_mips_pending_timer(struct kvm_vcpu *vcpu);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void kvm_mips_queue_timer_int_cb(struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
void kvm_mips_dequeue_timer_int_cb(struct kvm_vcpu *vcpu);
 | 
					 | 
				
			||||||
void kvm_mips_queue_io_int_cb(struct kvm_vcpu *vcpu,
 | 
					 | 
				
			||||||
			      struct kvm_mips_interrupt *irq);
 | 
					 | 
				
			||||||
void kvm_mips_dequeue_io_int_cb(struct kvm_vcpu *vcpu,
 | 
					 | 
				
			||||||
				struct kvm_mips_interrupt *irq);
 | 
					 | 
				
			||||||
int kvm_mips_irq_deliver_cb(struct kvm_vcpu *vcpu, unsigned int priority,
 | 
					 | 
				
			||||||
			    u32 cause);
 | 
					 | 
				
			||||||
int kvm_mips_irq_clear_cb(struct kvm_vcpu *vcpu, unsigned int priority,
 | 
					 | 
				
			||||||
			  u32 cause);
 | 
					 | 
				
			||||||
void kvm_mips_deliver_interrupts(struct kvm_vcpu *vcpu, u32 cause);
 | 
					void kvm_mips_deliver_interrupts(struct kvm_vcpu *vcpu, u32 cause);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -30,7 +30,6 @@
 | 
				
			||||||
#include <linux/kvm_host.h>
 | 
					#include <linux/kvm_host.h>
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#include "interrupt.h"
 | 
					#include "interrupt.h"
 | 
				
			||||||
#include "commpage.h"
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define CREATE_TRACE_POINTS
 | 
					#define CREATE_TRACE_POINTS
 | 
				
			||||||
#include "trace.h"
 | 
					#include "trace.h"
 | 
				
			||||||
| 
						 | 
					@ -58,7 +57,6 @@ struct kvm_stats_debugfs_item debugfs_entries[] = {
 | 
				
			||||||
	VCPU_STAT("fpe", fpe_exits),
 | 
						VCPU_STAT("fpe", fpe_exits),
 | 
				
			||||||
	VCPU_STAT("msa_disabled", msa_disabled_exits),
 | 
						VCPU_STAT("msa_disabled", msa_disabled_exits),
 | 
				
			||||||
	VCPU_STAT("flush_dcache", flush_dcache_exits),
 | 
						VCPU_STAT("flush_dcache", flush_dcache_exits),
 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
	VCPU_STAT("vz_gpsi", vz_gpsi_exits),
 | 
						VCPU_STAT("vz_gpsi", vz_gpsi_exits),
 | 
				
			||||||
	VCPU_STAT("vz_gsfc", vz_gsfc_exits),
 | 
						VCPU_STAT("vz_gsfc", vz_gsfc_exits),
 | 
				
			||||||
	VCPU_STAT("vz_hc", vz_hc_exits),
 | 
						VCPU_STAT("vz_hc", vz_hc_exits),
 | 
				
			||||||
| 
						 | 
					@ -69,7 +67,6 @@ struct kvm_stats_debugfs_item debugfs_entries[] = {
 | 
				
			||||||
	VCPU_STAT("vz_resvd", vz_resvd_exits),
 | 
						VCPU_STAT("vz_resvd", vz_resvd_exits),
 | 
				
			||||||
#ifdef CONFIG_CPU_LOONGSON64
 | 
					#ifdef CONFIG_CPU_LOONGSON64
 | 
				
			||||||
	VCPU_STAT("vz_cpucfg", vz_cpucfg_exits),
 | 
						VCPU_STAT("vz_cpucfg", vz_cpucfg_exits),
 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
	VCPU_STAT("halt_successful_poll", halt_successful_poll),
 | 
						VCPU_STAT("halt_successful_poll", halt_successful_poll),
 | 
				
			||||||
	VCPU_STAT("halt_attempted_poll", halt_attempted_poll),
 | 
						VCPU_STAT("halt_attempted_poll", halt_attempted_poll),
 | 
				
			||||||
| 
						 | 
					@ -139,11 +136,7 @@ int kvm_arch_init_vm(struct kvm *kvm, unsigned long type)
 | 
				
			||||||
	switch (type) {
 | 
						switch (type) {
 | 
				
			||||||
	case KVM_VM_MIPS_AUTO:
 | 
						case KVM_VM_MIPS_AUTO:
 | 
				
			||||||
		break;
 | 
							break;
 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
	case KVM_VM_MIPS_VZ:
 | 
						case KVM_VM_MIPS_VZ:
 | 
				
			||||||
#else
 | 
					 | 
				
			||||||
	case KVM_VM_MIPS_TE:
 | 
					 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
		break;
 | 
							break;
 | 
				
			||||||
	default:
 | 
						default:
 | 
				
			||||||
		/* Unsupported KVM type */
 | 
							/* Unsupported KVM type */
 | 
				
			||||||
| 
						 | 
					@ -361,7 +354,7 @@ int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* TLB refill (or XTLB refill on 64-bit VZ where KX=1) */
 | 
						/* TLB refill (or XTLB refill on 64-bit VZ where KX=1) */
 | 
				
			||||||
	refill_start = gebase;
 | 
						refill_start = gebase;
 | 
				
			||||||
	if (IS_ENABLED(CONFIG_KVM_MIPS_VZ) && IS_ENABLED(CONFIG_64BIT))
 | 
						if (IS_ENABLED(CONFIG_64BIT))
 | 
				
			||||||
		refill_start += 0x080;
 | 
							refill_start += 0x080;
 | 
				
			||||||
	refill_end = kvm_mips_build_tlb_refill_exception(refill_start, handler);
 | 
						refill_end = kvm_mips_build_tlb_refill_exception(refill_start, handler);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -397,20 +390,6 @@ int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
 | 
				
			||||||
	flush_icache_range((unsigned long)gebase,
 | 
						flush_icache_range((unsigned long)gebase,
 | 
				
			||||||
			   (unsigned long)gebase + ALIGN(size, PAGE_SIZE));
 | 
								   (unsigned long)gebase + ALIGN(size, PAGE_SIZE));
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/*
 | 
					 | 
				
			||||||
	 * Allocate comm page for guest kernel, a TLB will be reserved for
 | 
					 | 
				
			||||||
	 * mapping GVA @ 0xFFFF8000 to this page
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
	vcpu->arch.kseg0_commpage = kzalloc(PAGE_SIZE << 1, GFP_KERNEL);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (!vcpu->arch.kseg0_commpage) {
 | 
					 | 
				
			||||||
		err = -ENOMEM;
 | 
					 | 
				
			||||||
		goto out_free_gebase;
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	kvm_debug("Allocated COMM page @ %p\n", vcpu->arch.kseg0_commpage);
 | 
					 | 
				
			||||||
	kvm_mips_commpage_init(vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Init */
 | 
						/* Init */
 | 
				
			||||||
	vcpu->arch.last_sched_cpu = -1;
 | 
						vcpu->arch.last_sched_cpu = -1;
 | 
				
			||||||
	vcpu->arch.last_exec_cpu = -1;
 | 
						vcpu->arch.last_exec_cpu = -1;
 | 
				
			||||||
| 
						 | 
					@ -418,12 +397,10 @@ int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
 | 
				
			||||||
	/* Initial guest state */
 | 
						/* Initial guest state */
 | 
				
			||||||
	err = kvm_mips_callbacks->vcpu_setup(vcpu);
 | 
						err = kvm_mips_callbacks->vcpu_setup(vcpu);
 | 
				
			||||||
	if (err)
 | 
						if (err)
 | 
				
			||||||
		goto out_free_commpage;
 | 
							goto out_free_gebase;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return 0;
 | 
						return 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
out_free_commpage:
 | 
					 | 
				
			||||||
	kfree(vcpu->arch.kseg0_commpage);
 | 
					 | 
				
			||||||
out_free_gebase:
 | 
					out_free_gebase:
 | 
				
			||||||
	kfree(gebase);
 | 
						kfree(gebase);
 | 
				
			||||||
out_uninit_vcpu:
 | 
					out_uninit_vcpu:
 | 
				
			||||||
| 
						 | 
					@ -439,7 +416,6 @@ void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	kvm_mmu_free_memory_caches(vcpu);
 | 
						kvm_mmu_free_memory_caches(vcpu);
 | 
				
			||||||
	kfree(vcpu->arch.guest_ebase);
 | 
						kfree(vcpu->arch.guest_ebase);
 | 
				
			||||||
	kfree(vcpu->arch.kseg0_commpage);
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
	kvm_mips_callbacks->vcpu_uninit(vcpu);
 | 
						kvm_mips_callbacks->vcpu_uninit(vcpu);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					@ -1212,10 +1188,6 @@ int kvm_mips_handle_exit(struct kvm_vcpu *vcpu)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	vcpu->mode = OUTSIDE_GUEST_MODE;
 | 
						vcpu->mode = OUTSIDE_GUEST_MODE;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* re-enable HTW before enabling interrupts */
 | 
					 | 
				
			||||||
	if (!IS_ENABLED(CONFIG_KVM_MIPS_VZ))
 | 
					 | 
				
			||||||
		htw_start();
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Set a default exit reason */
 | 
						/* Set a default exit reason */
 | 
				
			||||||
	run->exit_reason = KVM_EXIT_UNKNOWN;
 | 
						run->exit_reason = KVM_EXIT_UNKNOWN;
 | 
				
			||||||
	run->ready_for_interrupt_injection = 1;
 | 
						run->ready_for_interrupt_injection = 1;
 | 
				
			||||||
| 
						 | 
					@ -1232,22 +1204,6 @@ int kvm_mips_handle_exit(struct kvm_vcpu *vcpu)
 | 
				
			||||||
			cause, opc, run, vcpu);
 | 
								cause, opc, run, vcpu);
 | 
				
			||||||
	trace_kvm_exit(vcpu, exccode);
 | 
						trace_kvm_exit(vcpu, exccode);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (!IS_ENABLED(CONFIG_KVM_MIPS_VZ)) {
 | 
					 | 
				
			||||||
		/*
 | 
					 | 
				
			||||||
		 * Do a privilege check, if in UM most of these exit conditions
 | 
					 | 
				
			||||||
		 * end up causing an exception to be delivered to the Guest
 | 
					 | 
				
			||||||
		 * Kernel
 | 
					 | 
				
			||||||
		 */
 | 
					 | 
				
			||||||
		er = kvm_mips_check_privilege(cause, opc, vcpu);
 | 
					 | 
				
			||||||
		if (er == EMULATE_PRIV_FAIL) {
 | 
					 | 
				
			||||||
			goto skip_emul;
 | 
					 | 
				
			||||||
		} else if (er == EMULATE_FAIL) {
 | 
					 | 
				
			||||||
			run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
 | 
					 | 
				
			||||||
			ret = RESUME_HOST;
 | 
					 | 
				
			||||||
			goto skip_emul;
 | 
					 | 
				
			||||||
		}
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	switch (exccode) {
 | 
						switch (exccode) {
 | 
				
			||||||
	case EXCCODE_INT:
 | 
						case EXCCODE_INT:
 | 
				
			||||||
		kvm_debug("[%d]EXCCODE_INT @ %p\n", vcpu->vcpu_id, opc);
 | 
							kvm_debug("[%d]EXCCODE_INT @ %p\n", vcpu->vcpu_id, opc);
 | 
				
			||||||
| 
						 | 
					@ -1357,7 +1313,6 @@ int kvm_mips_handle_exit(struct kvm_vcpu *vcpu)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
skip_emul:
 | 
					 | 
				
			||||||
	local_irq_disable();
 | 
						local_irq_disable();
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (ret == RESUME_GUEST)
 | 
						if (ret == RESUME_GUEST)
 | 
				
			||||||
| 
						 | 
					@ -1406,11 +1361,6 @@ int kvm_mips_handle_exit(struct kvm_vcpu *vcpu)
 | 
				
			||||||
		    read_c0_config5() & MIPS_CONF5_MSAEN)
 | 
							    read_c0_config5() & MIPS_CONF5_MSAEN)
 | 
				
			||||||
			__kvm_restore_msacsr(&vcpu->arch);
 | 
								__kvm_restore_msacsr(&vcpu->arch);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Disable HTW before returning to guest or host */
 | 
					 | 
				
			||||||
	if (!IS_ENABLED(CONFIG_KVM_MIPS_VZ))
 | 
					 | 
				
			||||||
		htw_stop();
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return ret;
 | 
						return ret;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1429,10 +1379,6 @@ void kvm_own_fpu(struct kvm_vcpu *vcpu)
 | 
				
			||||||
	 * FR=0 FPU state, and we don't want to hit reserved instruction
 | 
						 * FR=0 FPU state, and we don't want to hit reserved instruction
 | 
				
			||||||
	 * exceptions trying to save the MSA state later when CU=1 && FR=1, so
 | 
						 * exceptions trying to save the MSA state later when CU=1 && FR=1, so
 | 
				
			||||||
	 * play it safe and save it first.
 | 
						 * play it safe and save it first.
 | 
				
			||||||
	 *
 | 
					 | 
				
			||||||
	 * In theory we shouldn't ever hit this case since kvm_lose_fpu() should
 | 
					 | 
				
			||||||
	 * get called when guest CU1 is set, however we can't trust the guest
 | 
					 | 
				
			||||||
	 * not to clobber the status register directly via the commpage.
 | 
					 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	if (cpu_has_msa && sr & ST0_CU1 && !(sr & ST0_FR) &&
 | 
						if (cpu_has_msa && sr & ST0_CU1 && !(sr & ST0_FR) &&
 | 
				
			||||||
	    vcpu->arch.aux_inuse & KVM_MIPS_AUX_MSA)
 | 
						    vcpu->arch.aux_inuse & KVM_MIPS_AUX_MSA)
 | 
				
			||||||
| 
						 | 
					@ -1553,11 +1499,6 @@ void kvm_lose_fpu(struct kvm_vcpu *vcpu)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	preempt_disable();
 | 
						preempt_disable();
 | 
				
			||||||
	if (cpu_has_msa && vcpu->arch.aux_inuse & KVM_MIPS_AUX_MSA) {
 | 
						if (cpu_has_msa && vcpu->arch.aux_inuse & KVM_MIPS_AUX_MSA) {
 | 
				
			||||||
		if (!IS_ENABLED(CONFIG_KVM_MIPS_VZ)) {
 | 
					 | 
				
			||||||
			set_c0_config5(MIPS_CONF5_MSAEN);
 | 
					 | 
				
			||||||
			enable_fpu_hazard();
 | 
					 | 
				
			||||||
		}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		__kvm_save_msa(&vcpu->arch);
 | 
							__kvm_save_msa(&vcpu->arch);
 | 
				
			||||||
		trace_kvm_aux(vcpu, KVM_TRACE_AUX_SAVE, KVM_TRACE_AUX_FPU_MSA);
 | 
							trace_kvm_aux(vcpu, KVM_TRACE_AUX_SAVE, KVM_TRACE_AUX_FPU_MSA);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1569,11 +1510,6 @@ void kvm_lose_fpu(struct kvm_vcpu *vcpu)
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
		vcpu->arch.aux_inuse &= ~(KVM_MIPS_AUX_FPU | KVM_MIPS_AUX_MSA);
 | 
							vcpu->arch.aux_inuse &= ~(KVM_MIPS_AUX_FPU | KVM_MIPS_AUX_MSA);
 | 
				
			||||||
	} else if (vcpu->arch.aux_inuse & KVM_MIPS_AUX_FPU) {
 | 
						} else if (vcpu->arch.aux_inuse & KVM_MIPS_AUX_FPU) {
 | 
				
			||||||
		if (!IS_ENABLED(CONFIG_KVM_MIPS_VZ)) {
 | 
					 | 
				
			||||||
			set_c0_status(ST0_CU1);
 | 
					 | 
				
			||||||
			enable_fpu_hazard();
 | 
					 | 
				
			||||||
		}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		__kvm_save_fpu(&vcpu->arch);
 | 
							__kvm_save_fpu(&vcpu->arch);
 | 
				
			||||||
		vcpu->arch.aux_inuse &= ~KVM_MIPS_AUX_FPU;
 | 
							vcpu->arch.aux_inuse &= ~KVM_MIPS_AUX_FPU;
 | 
				
			||||||
		trace_kvm_aux(vcpu, KVM_TRACE_AUX_SAVE, KVM_TRACE_AUX_FPU);
 | 
							trace_kvm_aux(vcpu, KVM_TRACE_AUX_SAVE, KVM_TRACE_AUX_FPU);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -756,209 +756,6 @@ static int kvm_mips_map_page(struct kvm_vcpu *vcpu, unsigned long gpa,
 | 
				
			||||||
	return err;
 | 
						return err;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static pte_t *kvm_trap_emul_pte_for_gva(struct kvm_vcpu *vcpu,
 | 
					 | 
				
			||||||
					unsigned long addr)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	struct kvm_mmu_memory_cache *memcache = &vcpu->arch.mmu_page_cache;
 | 
					 | 
				
			||||||
	pgd_t *pgdp;
 | 
					 | 
				
			||||||
	int ret;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* We need a minimum of cached pages ready for page table creation */
 | 
					 | 
				
			||||||
	ret = kvm_mmu_topup_memory_cache(memcache, KVM_MMU_CACHE_MIN_PAGES);
 | 
					 | 
				
			||||||
	if (ret)
 | 
					 | 
				
			||||||
		return NULL;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (KVM_GUEST_KERNEL_MODE(vcpu))
 | 
					 | 
				
			||||||
		pgdp = vcpu->arch.guest_kernel_mm.pgd;
 | 
					 | 
				
			||||||
	else
 | 
					 | 
				
			||||||
		pgdp = vcpu->arch.guest_user_mm.pgd;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return kvm_mips_walk_pgd(pgdp, memcache, addr);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
void kvm_trap_emul_invalidate_gva(struct kvm_vcpu *vcpu, unsigned long addr,
 | 
					 | 
				
			||||||
				  bool user)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	pgd_t *pgdp;
 | 
					 | 
				
			||||||
	pte_t *ptep;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	addr &= PAGE_MASK << 1;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	pgdp = vcpu->arch.guest_kernel_mm.pgd;
 | 
					 | 
				
			||||||
	ptep = kvm_mips_walk_pgd(pgdp, NULL, addr);
 | 
					 | 
				
			||||||
	if (ptep) {
 | 
					 | 
				
			||||||
		ptep[0] = pfn_pte(0, __pgprot(0));
 | 
					 | 
				
			||||||
		ptep[1] = pfn_pte(0, __pgprot(0));
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (user) {
 | 
					 | 
				
			||||||
		pgdp = vcpu->arch.guest_user_mm.pgd;
 | 
					 | 
				
			||||||
		ptep = kvm_mips_walk_pgd(pgdp, NULL, addr);
 | 
					 | 
				
			||||||
		if (ptep) {
 | 
					 | 
				
			||||||
			ptep[0] = pfn_pte(0, __pgprot(0));
 | 
					 | 
				
			||||||
			ptep[1] = pfn_pte(0, __pgprot(0));
 | 
					 | 
				
			||||||
		}
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/*
 | 
					 | 
				
			||||||
 * kvm_mips_flush_gva_{pte,pmd,pud,pgd,pt}.
 | 
					 | 
				
			||||||
 * Flush a range of guest physical address space from the VM's GPA page tables.
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
static bool kvm_mips_flush_gva_pte(pte_t *pte, unsigned long start_gva,
 | 
					 | 
				
			||||||
				   unsigned long end_gva)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	int i_min = pte_index(start_gva);
 | 
					 | 
				
			||||||
	int i_max = pte_index(end_gva);
 | 
					 | 
				
			||||||
	bool safe_to_remove = (i_min == 0 && i_max == PTRS_PER_PTE - 1);
 | 
					 | 
				
			||||||
	int i;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/*
 | 
					 | 
				
			||||||
	 * There's no freeing to do, so there's no point clearing individual
 | 
					 | 
				
			||||||
	 * entries unless only part of the last level page table needs flushing.
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
	if (safe_to_remove)
 | 
					 | 
				
			||||||
		return true;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	for (i = i_min; i <= i_max; ++i) {
 | 
					 | 
				
			||||||
		if (!pte_present(pte[i]))
 | 
					 | 
				
			||||||
			continue;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		set_pte(pte + i, __pte(0));
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
	return false;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
static bool kvm_mips_flush_gva_pmd(pmd_t *pmd, unsigned long start_gva,
 | 
					 | 
				
			||||||
				   unsigned long end_gva)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	pte_t *pte;
 | 
					 | 
				
			||||||
	unsigned long end = ~0ul;
 | 
					 | 
				
			||||||
	int i_min = pmd_index(start_gva);
 | 
					 | 
				
			||||||
	int i_max = pmd_index(end_gva);
 | 
					 | 
				
			||||||
	bool safe_to_remove = (i_min == 0 && i_max == PTRS_PER_PMD - 1);
 | 
					 | 
				
			||||||
	int i;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	for (i = i_min; i <= i_max; ++i, start_gva = 0) {
 | 
					 | 
				
			||||||
		if (!pmd_present(pmd[i]))
 | 
					 | 
				
			||||||
			continue;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		pte = pte_offset_kernel(pmd + i, 0);
 | 
					 | 
				
			||||||
		if (i == i_max)
 | 
					 | 
				
			||||||
			end = end_gva;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		if (kvm_mips_flush_gva_pte(pte, start_gva, end)) {
 | 
					 | 
				
			||||||
			pmd_clear(pmd + i);
 | 
					 | 
				
			||||||
			pte_free_kernel(NULL, pte);
 | 
					 | 
				
			||||||
		} else {
 | 
					 | 
				
			||||||
			safe_to_remove = false;
 | 
					 | 
				
			||||||
		}
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
	return safe_to_remove;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
static bool kvm_mips_flush_gva_pud(pud_t *pud, unsigned long start_gva,
 | 
					 | 
				
			||||||
				   unsigned long end_gva)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	pmd_t *pmd;
 | 
					 | 
				
			||||||
	unsigned long end = ~0ul;
 | 
					 | 
				
			||||||
	int i_min = pud_index(start_gva);
 | 
					 | 
				
			||||||
	int i_max = pud_index(end_gva);
 | 
					 | 
				
			||||||
	bool safe_to_remove = (i_min == 0 && i_max == PTRS_PER_PUD - 1);
 | 
					 | 
				
			||||||
	int i;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	for (i = i_min; i <= i_max; ++i, start_gva = 0) {
 | 
					 | 
				
			||||||
		if (!pud_present(pud[i]))
 | 
					 | 
				
			||||||
			continue;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		pmd = pmd_offset(pud + i, 0);
 | 
					 | 
				
			||||||
		if (i == i_max)
 | 
					 | 
				
			||||||
			end = end_gva;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		if (kvm_mips_flush_gva_pmd(pmd, start_gva, end)) {
 | 
					 | 
				
			||||||
			pud_clear(pud + i);
 | 
					 | 
				
			||||||
			pmd_free(NULL, pmd);
 | 
					 | 
				
			||||||
		} else {
 | 
					 | 
				
			||||||
			safe_to_remove = false;
 | 
					 | 
				
			||||||
		}
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
	return safe_to_remove;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
static bool kvm_mips_flush_gva_pgd(pgd_t *pgd, unsigned long start_gva,
 | 
					 | 
				
			||||||
				   unsigned long end_gva)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	p4d_t *p4d;
 | 
					 | 
				
			||||||
	pud_t *pud;
 | 
					 | 
				
			||||||
	unsigned long end = ~0ul;
 | 
					 | 
				
			||||||
	int i_min = pgd_index(start_gva);
 | 
					 | 
				
			||||||
	int i_max = pgd_index(end_gva);
 | 
					 | 
				
			||||||
	bool safe_to_remove = (i_min == 0 && i_max == PTRS_PER_PGD - 1);
 | 
					 | 
				
			||||||
	int i;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	for (i = i_min; i <= i_max; ++i, start_gva = 0) {
 | 
					 | 
				
			||||||
		if (!pgd_present(pgd[i]))
 | 
					 | 
				
			||||||
			continue;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		p4d = p4d_offset(pgd, 0);
 | 
					 | 
				
			||||||
		pud = pud_offset(p4d + i, 0);
 | 
					 | 
				
			||||||
		if (i == i_max)
 | 
					 | 
				
			||||||
			end = end_gva;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		if (kvm_mips_flush_gva_pud(pud, start_gva, end)) {
 | 
					 | 
				
			||||||
			pgd_clear(pgd + i);
 | 
					 | 
				
			||||||
			pud_free(NULL, pud);
 | 
					 | 
				
			||||||
		} else {
 | 
					 | 
				
			||||||
			safe_to_remove = false;
 | 
					 | 
				
			||||||
		}
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
	return safe_to_remove;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
void kvm_mips_flush_gva_pt(pgd_t *pgd, enum kvm_mips_flush flags)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	if (flags & KMF_GPA) {
 | 
					 | 
				
			||||||
		/* all of guest virtual address space could be affected */
 | 
					 | 
				
			||||||
		if (flags & KMF_KERN)
 | 
					 | 
				
			||||||
			/* useg, kseg0, seg2/3 */
 | 
					 | 
				
			||||||
			kvm_mips_flush_gva_pgd(pgd, 0, 0x7fffffff);
 | 
					 | 
				
			||||||
		else
 | 
					 | 
				
			||||||
			/* useg */
 | 
					 | 
				
			||||||
			kvm_mips_flush_gva_pgd(pgd, 0, 0x3fffffff);
 | 
					 | 
				
			||||||
	} else {
 | 
					 | 
				
			||||||
		/* useg */
 | 
					 | 
				
			||||||
		kvm_mips_flush_gva_pgd(pgd, 0, 0x3fffffff);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		/* kseg2/3 */
 | 
					 | 
				
			||||||
		if (flags & KMF_KERN)
 | 
					 | 
				
			||||||
			kvm_mips_flush_gva_pgd(pgd, 0x60000000, 0x7fffffff);
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
static pte_t kvm_mips_gpa_pte_to_gva_unmapped(pte_t pte)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	/*
 | 
					 | 
				
			||||||
	 * Don't leak writeable but clean entries from GPA page tables. We don't
 | 
					 | 
				
			||||||
	 * want the normal Linux tlbmod handler to handle dirtying when KVM
 | 
					 | 
				
			||||||
	 * accesses guest memory.
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
	if (!pte_dirty(pte))
 | 
					 | 
				
			||||||
		pte = pte_wrprotect(pte);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return pte;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
static pte_t kvm_mips_gpa_pte_to_gva_mapped(pte_t pte, long entrylo)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	/* Guest EntryLo overrides host EntryLo */
 | 
					 | 
				
			||||||
	if (!(entrylo & ENTRYLO_D))
 | 
					 | 
				
			||||||
		pte = pte_mkclean(pte);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return kvm_mips_gpa_pte_to_gva_unmapped(pte);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
int kvm_mips_handle_vz_root_tlb_fault(unsigned long badvaddr,
 | 
					int kvm_mips_handle_vz_root_tlb_fault(unsigned long badvaddr,
 | 
				
			||||||
				      struct kvm_vcpu *vcpu,
 | 
									      struct kvm_vcpu *vcpu,
 | 
				
			||||||
				      bool write_fault)
 | 
									      bool write_fault)
 | 
				
			||||||
| 
						 | 
					@ -972,125 +769,6 @@ int kvm_mips_handle_vz_root_tlb_fault(unsigned long badvaddr,
 | 
				
			||||||
	/* Invalidate this entry in the TLB */
 | 
						/* Invalidate this entry in the TLB */
 | 
				
			||||||
	return kvm_vz_host_tlb_inv(vcpu, badvaddr);
 | 
						return kvm_vz_host_tlb_inv(vcpu, badvaddr);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/* XXXKYMA: Must be called with interrupts disabled */
 | 
					 | 
				
			||||||
int kvm_mips_handle_kseg0_tlb_fault(unsigned long badvaddr,
 | 
					 | 
				
			||||||
				    struct kvm_vcpu *vcpu,
 | 
					 | 
				
			||||||
				    bool write_fault)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	unsigned long gpa;
 | 
					 | 
				
			||||||
	pte_t pte_gpa[2], *ptep_gva;
 | 
					 | 
				
			||||||
	int idx;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (KVM_GUEST_KSEGX(badvaddr) != KVM_GUEST_KSEG0) {
 | 
					 | 
				
			||||||
		kvm_err("%s: Invalid BadVaddr: %#lx\n", __func__, badvaddr);
 | 
					 | 
				
			||||||
		kvm_mips_dump_host_tlbs();
 | 
					 | 
				
			||||||
		return -1;
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Get the GPA page table entry */
 | 
					 | 
				
			||||||
	gpa = KVM_GUEST_CPHYSADDR(badvaddr);
 | 
					 | 
				
			||||||
	idx = (badvaddr >> PAGE_SHIFT) & 1;
 | 
					 | 
				
			||||||
	if (kvm_mips_map_page(vcpu, gpa, write_fault, &pte_gpa[idx],
 | 
					 | 
				
			||||||
			      &pte_gpa[!idx]) < 0)
 | 
					 | 
				
			||||||
		return -1;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Get the GVA page table entry */
 | 
					 | 
				
			||||||
	ptep_gva = kvm_trap_emul_pte_for_gva(vcpu, badvaddr & ~PAGE_SIZE);
 | 
					 | 
				
			||||||
	if (!ptep_gva) {
 | 
					 | 
				
			||||||
		kvm_err("No ptep for gva %lx\n", badvaddr);
 | 
					 | 
				
			||||||
		return -1;
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Copy a pair of entries from GPA page table to GVA page table */
 | 
					 | 
				
			||||||
	ptep_gva[0] = kvm_mips_gpa_pte_to_gva_unmapped(pte_gpa[0]);
 | 
					 | 
				
			||||||
	ptep_gva[1] = kvm_mips_gpa_pte_to_gva_unmapped(pte_gpa[1]);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Invalidate this entry in the TLB, guest kernel ASID only */
 | 
					 | 
				
			||||||
	kvm_mips_host_tlb_inv(vcpu, badvaddr, false, true);
 | 
					 | 
				
			||||||
	return 0;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
int kvm_mips_handle_mapped_seg_tlb_fault(struct kvm_vcpu *vcpu,
 | 
					 | 
				
			||||||
					 struct kvm_mips_tlb *tlb,
 | 
					 | 
				
			||||||
					 unsigned long gva,
 | 
					 | 
				
			||||||
					 bool write_fault)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	struct kvm *kvm = vcpu->kvm;
 | 
					 | 
				
			||||||
	long tlb_lo[2];
 | 
					 | 
				
			||||||
	pte_t pte_gpa[2], *ptep_buddy, *ptep_gva;
 | 
					 | 
				
			||||||
	unsigned int idx = TLB_LO_IDX(*tlb, gva);
 | 
					 | 
				
			||||||
	bool kernel = KVM_GUEST_KERNEL_MODE(vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	tlb_lo[0] = tlb->tlb_lo[0];
 | 
					 | 
				
			||||||
	tlb_lo[1] = tlb->tlb_lo[1];
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/*
 | 
					 | 
				
			||||||
	 * The commpage address must not be mapped to anything else if the guest
 | 
					 | 
				
			||||||
	 * TLB contains entries nearby, or commpage accesses will break.
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
	if (!((gva ^ KVM_GUEST_COMMPAGE_ADDR) & VPN2_MASK & (PAGE_MASK << 1)))
 | 
					 | 
				
			||||||
		tlb_lo[TLB_LO_IDX(*tlb, KVM_GUEST_COMMPAGE_ADDR)] = 0;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Get the GPA page table entry */
 | 
					 | 
				
			||||||
	if (kvm_mips_map_page(vcpu, mips3_tlbpfn_to_paddr(tlb_lo[idx]),
 | 
					 | 
				
			||||||
			      write_fault, &pte_gpa[idx], NULL) < 0)
 | 
					 | 
				
			||||||
		return -1;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* And its GVA buddy's GPA page table entry if it also exists */
 | 
					 | 
				
			||||||
	pte_gpa[!idx] = pfn_pte(0, __pgprot(0));
 | 
					 | 
				
			||||||
	if (tlb_lo[!idx] & ENTRYLO_V) {
 | 
					 | 
				
			||||||
		spin_lock(&kvm->mmu_lock);
 | 
					 | 
				
			||||||
		ptep_buddy = kvm_mips_pte_for_gpa(kvm, NULL,
 | 
					 | 
				
			||||||
					mips3_tlbpfn_to_paddr(tlb_lo[!idx]));
 | 
					 | 
				
			||||||
		if (ptep_buddy)
 | 
					 | 
				
			||||||
			pte_gpa[!idx] = *ptep_buddy;
 | 
					 | 
				
			||||||
		spin_unlock(&kvm->mmu_lock);
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Get the GVA page table entry pair */
 | 
					 | 
				
			||||||
	ptep_gva = kvm_trap_emul_pte_for_gva(vcpu, gva & ~PAGE_SIZE);
 | 
					 | 
				
			||||||
	if (!ptep_gva) {
 | 
					 | 
				
			||||||
		kvm_err("No ptep for gva %lx\n", gva);
 | 
					 | 
				
			||||||
		return -1;
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Copy a pair of entries from GPA page table to GVA page table */
 | 
					 | 
				
			||||||
	ptep_gva[0] = kvm_mips_gpa_pte_to_gva_mapped(pte_gpa[0], tlb_lo[0]);
 | 
					 | 
				
			||||||
	ptep_gva[1] = kvm_mips_gpa_pte_to_gva_mapped(pte_gpa[1], tlb_lo[1]);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Invalidate this entry in the TLB, current guest mode ASID only */
 | 
					 | 
				
			||||||
	kvm_mips_host_tlb_inv(vcpu, gva, !kernel, kernel);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	kvm_debug("@ %#lx tlb_lo0: 0x%08lx tlb_lo1: 0x%08lx\n", vcpu->arch.pc,
 | 
					 | 
				
			||||||
		  tlb->tlb_lo[0], tlb->tlb_lo[1]);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return 0;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
int kvm_mips_handle_commpage_tlb_fault(unsigned long badvaddr,
 | 
					 | 
				
			||||||
				       struct kvm_vcpu *vcpu)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	kvm_pfn_t pfn;
 | 
					 | 
				
			||||||
	pte_t *ptep;
 | 
					 | 
				
			||||||
	pgprot_t prot;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	ptep = kvm_trap_emul_pte_for_gva(vcpu, badvaddr);
 | 
					 | 
				
			||||||
	if (!ptep) {
 | 
					 | 
				
			||||||
		kvm_err("No ptep for commpage %lx\n", badvaddr);
 | 
					 | 
				
			||||||
		return -1;
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	pfn = PFN_DOWN(virt_to_phys(vcpu->arch.kseg0_commpage));
 | 
					 | 
				
			||||||
	/* Also set valid and dirty, so refill handler doesn't have to */
 | 
					 | 
				
			||||||
	prot = vm_get_page_prot(VM_READ|VM_WRITE|VM_SHARED);
 | 
					 | 
				
			||||||
	*ptep = pte_mkyoung(pte_mkdirty(pfn_pte(pfn, prot)));
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* Invalidate this entry in the TLB, guest kernel ASID only */
 | 
					 | 
				
			||||||
	kvm_mips_host_tlb_inv(vcpu, badvaddr, false, true);
 | 
					 | 
				
			||||||
	return 0;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
/**
 | 
					/**
 | 
				
			||||||
 * kvm_mips_migrate_count() - Migrate timer.
 | 
					 * kvm_mips_migrate_count() - Migrate timer.
 | 
				
			||||||
| 
						 | 
					@ -1153,86 +831,3 @@ void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	local_irq_restore(flags);
 | 
						local_irq_restore(flags);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					 | 
				
			||||||
/**
 | 
					 | 
				
			||||||
 * kvm_trap_emul_gva_fault() - Safely attempt to handle a GVA access fault.
 | 
					 | 
				
			||||||
 * @vcpu:	Virtual CPU.
 | 
					 | 
				
			||||||
 * @gva:	Guest virtual address to be accessed.
 | 
					 | 
				
			||||||
 * @write:	True if write attempted (must be dirtied and made writable).
 | 
					 | 
				
			||||||
 *
 | 
					 | 
				
			||||||
 * Safely attempt to handle a GVA fault, mapping GVA pages if necessary, and
 | 
					 | 
				
			||||||
 * dirtying the page if @write so that guest instructions can be modified.
 | 
					 | 
				
			||||||
 *
 | 
					 | 
				
			||||||
 * Returns:	KVM_MIPS_MAPPED on success.
 | 
					 | 
				
			||||||
 *		KVM_MIPS_GVA if bad guest virtual address.
 | 
					 | 
				
			||||||
 *		KVM_MIPS_GPA if bad guest physical address.
 | 
					 | 
				
			||||||
 *		KVM_MIPS_TLB if guest TLB not present.
 | 
					 | 
				
			||||||
 *		KVM_MIPS_TLBINV if guest TLB present but not valid.
 | 
					 | 
				
			||||||
 *		KVM_MIPS_TLBMOD if guest TLB read only.
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
enum kvm_mips_fault_result kvm_trap_emul_gva_fault(struct kvm_vcpu *vcpu,
 | 
					 | 
				
			||||||
						   unsigned long gva,
 | 
					 | 
				
			||||||
						   bool write)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	struct mips_coproc *cop0 = vcpu->arch.cop0;
 | 
					 | 
				
			||||||
	struct kvm_mips_tlb *tlb;
 | 
					 | 
				
			||||||
	int index;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (KVM_GUEST_KSEGX(gva) == KVM_GUEST_KSEG0) {
 | 
					 | 
				
			||||||
		if (kvm_mips_handle_kseg0_tlb_fault(gva, vcpu, write) < 0)
 | 
					 | 
				
			||||||
			return KVM_MIPS_GPA;
 | 
					 | 
				
			||||||
	} else if ((KVM_GUEST_KSEGX(gva) < KVM_GUEST_KSEG0) ||
 | 
					 | 
				
			||||||
		   KVM_GUEST_KSEGX(gva) == KVM_GUEST_KSEG23) {
 | 
					 | 
				
			||||||
		/* Address should be in the guest TLB */
 | 
					 | 
				
			||||||
		index = kvm_mips_guest_tlb_lookup(vcpu, (gva & VPN2_MASK) |
 | 
					 | 
				
			||||||
			  (kvm_read_c0_guest_entryhi(cop0) & KVM_ENTRYHI_ASID));
 | 
					 | 
				
			||||||
		if (index < 0)
 | 
					 | 
				
			||||||
			return KVM_MIPS_TLB;
 | 
					 | 
				
			||||||
		tlb = &vcpu->arch.guest_tlb[index];
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		/* Entry should be valid, and dirty for writes */
 | 
					 | 
				
			||||||
		if (!TLB_IS_VALID(*tlb, gva))
 | 
					 | 
				
			||||||
			return KVM_MIPS_TLBINV;
 | 
					 | 
				
			||||||
		if (write && !TLB_IS_DIRTY(*tlb, gva))
 | 
					 | 
				
			||||||
			return KVM_MIPS_TLBMOD;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		if (kvm_mips_handle_mapped_seg_tlb_fault(vcpu, tlb, gva, write))
 | 
					 | 
				
			||||||
			return KVM_MIPS_GPA;
 | 
					 | 
				
			||||||
	} else {
 | 
					 | 
				
			||||||
		return KVM_MIPS_GVA;
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return KVM_MIPS_MAPPED;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
int kvm_get_inst(u32 *opc, struct kvm_vcpu *vcpu, u32 *out)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	int err;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (WARN(IS_ENABLED(CONFIG_KVM_MIPS_VZ),
 | 
					 | 
				
			||||||
		 "Expect BadInstr/BadInstrP registers to be used with VZ\n"))
 | 
					 | 
				
			||||||
		return -EINVAL;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
retry:
 | 
					 | 
				
			||||||
	kvm_trap_emul_gva_lockless_begin(vcpu);
 | 
					 | 
				
			||||||
	err = get_user(*out, opc);
 | 
					 | 
				
			||||||
	kvm_trap_emul_gva_lockless_end(vcpu);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (unlikely(err)) {
 | 
					 | 
				
			||||||
		/*
 | 
					 | 
				
			||||||
		 * Try to handle the fault, maybe we just raced with a GVA
 | 
					 | 
				
			||||||
		 * invalidation.
 | 
					 | 
				
			||||||
		 */
 | 
					 | 
				
			||||||
		err = kvm_trap_emul_gva_fault(vcpu, (unsigned long)opc,
 | 
					 | 
				
			||||||
					      false);
 | 
					 | 
				
			||||||
		if (unlikely(err)) {
 | 
					 | 
				
			||||||
			kvm_err("%s: illegal address: %p\n",
 | 
					 | 
				
			||||||
				__func__, opc);
 | 
					 | 
				
			||||||
			return -EFAULT;
 | 
					 | 
				
			||||||
		}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		/* Hopefully it'll work now */
 | 
					 | 
				
			||||||
		goto retry;
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
	return 0;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -30,10 +30,6 @@
 | 
				
			||||||
#include <asm/r4kcache.h>
 | 
					#include <asm/r4kcache.h>
 | 
				
			||||||
#define CONFIG_MIPS_MT
 | 
					#define CONFIG_MIPS_MT
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define KVM_GUEST_PC_TLB    0
 | 
					 | 
				
			||||||
#define KVM_GUEST_SP_TLB    1
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
unsigned long GUESTID_MASK;
 | 
					unsigned long GUESTID_MASK;
 | 
				
			||||||
EXPORT_SYMBOL_GPL(GUESTID_MASK);
 | 
					EXPORT_SYMBOL_GPL(GUESTID_MASK);
 | 
				
			||||||
unsigned long GUESTID_FIRST_VERSION;
 | 
					unsigned long GUESTID_FIRST_VERSION;
 | 
				
			||||||
| 
						 | 
					@ -50,91 +46,6 @@ static u32 kvm_mips_get_root_asid(struct kvm_vcpu *vcpu)
 | 
				
			||||||
	else
 | 
						else
 | 
				
			||||||
		return cpu_asid(smp_processor_id(), gpa_mm);
 | 
							return cpu_asid(smp_processor_id(), gpa_mm);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
static u32 kvm_mips_get_kernel_asid(struct kvm_vcpu *vcpu)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	struct mm_struct *kern_mm = &vcpu->arch.guest_kernel_mm;
 | 
					 | 
				
			||||||
	int cpu = smp_processor_id();
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return cpu_asid(cpu, kern_mm);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
static u32 kvm_mips_get_user_asid(struct kvm_vcpu *vcpu)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	struct mm_struct *user_mm = &vcpu->arch.guest_user_mm;
 | 
					 | 
				
			||||||
	int cpu = smp_processor_id();
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return cpu_asid(cpu, user_mm);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/* Structure defining an tlb entry data set. */
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
void kvm_mips_dump_host_tlbs(void)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	unsigned long flags;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	local_irq_save(flags);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	kvm_info("HOST TLBs:\n");
 | 
					 | 
				
			||||||
	dump_tlb_regs();
 | 
					 | 
				
			||||||
	pr_info("\n");
 | 
					 | 
				
			||||||
	dump_tlb_all();
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	local_irq_restore(flags);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
EXPORT_SYMBOL_GPL(kvm_mips_dump_host_tlbs);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
void kvm_mips_dump_guest_tlbs(struct kvm_vcpu *vcpu)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	struct mips_coproc *cop0 = vcpu->arch.cop0;
 | 
					 | 
				
			||||||
	struct kvm_mips_tlb tlb;
 | 
					 | 
				
			||||||
	int i;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	kvm_info("Guest TLBs:\n");
 | 
					 | 
				
			||||||
	kvm_info("Guest EntryHi: %#lx\n", kvm_read_c0_guest_entryhi(cop0));
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	for (i = 0; i < KVM_MIPS_GUEST_TLB_SIZE; i++) {
 | 
					 | 
				
			||||||
		tlb = vcpu->arch.guest_tlb[i];
 | 
					 | 
				
			||||||
		kvm_info("TLB%c%3d Hi 0x%08lx ",
 | 
					 | 
				
			||||||
			 (tlb.tlb_lo[0] | tlb.tlb_lo[1]) & ENTRYLO_V
 | 
					 | 
				
			||||||
							? ' ' : '*',
 | 
					 | 
				
			||||||
			 i, tlb.tlb_hi);
 | 
					 | 
				
			||||||
		kvm_info("Lo0=0x%09llx %c%c attr %lx ",
 | 
					 | 
				
			||||||
			 (u64) mips3_tlbpfn_to_paddr(tlb.tlb_lo[0]),
 | 
					 | 
				
			||||||
			 (tlb.tlb_lo[0] & ENTRYLO_D) ? 'D' : ' ',
 | 
					 | 
				
			||||||
			 (tlb.tlb_lo[0] & ENTRYLO_G) ? 'G' : ' ',
 | 
					 | 
				
			||||||
			 (tlb.tlb_lo[0] & ENTRYLO_C) >> ENTRYLO_C_SHIFT);
 | 
					 | 
				
			||||||
		kvm_info("Lo1=0x%09llx %c%c attr %lx sz=%lx\n",
 | 
					 | 
				
			||||||
			 (u64) mips3_tlbpfn_to_paddr(tlb.tlb_lo[1]),
 | 
					 | 
				
			||||||
			 (tlb.tlb_lo[1] & ENTRYLO_D) ? 'D' : ' ',
 | 
					 | 
				
			||||||
			 (tlb.tlb_lo[1] & ENTRYLO_G) ? 'G' : ' ',
 | 
					 | 
				
			||||||
			 (tlb.tlb_lo[1] & ENTRYLO_C) >> ENTRYLO_C_SHIFT,
 | 
					 | 
				
			||||||
			 tlb.tlb_mask);
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
EXPORT_SYMBOL_GPL(kvm_mips_dump_guest_tlbs);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
int kvm_mips_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long entryhi)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	int i;
 | 
					 | 
				
			||||||
	int index = -1;
 | 
					 | 
				
			||||||
	struct kvm_mips_tlb *tlb = vcpu->arch.guest_tlb;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	for (i = 0; i < KVM_MIPS_GUEST_TLB_SIZE; i++) {
 | 
					 | 
				
			||||||
		if (TLB_HI_VPN2_HIT(tlb[i], entryhi) &&
 | 
					 | 
				
			||||||
		    TLB_HI_ASID_HIT(tlb[i], entryhi)) {
 | 
					 | 
				
			||||||
			index = i;
 | 
					 | 
				
			||||||
			break;
 | 
					 | 
				
			||||||
		}
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	kvm_debug("%s: entryhi: %#lx, index: %d lo0: %#lx, lo1: %#lx\n",
 | 
					 | 
				
			||||||
		  __func__, entryhi, index, tlb[i].tlb_lo[0], tlb[i].tlb_lo[1]);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return index;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
EXPORT_SYMBOL_GPL(kvm_mips_guest_tlb_lookup);
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
static int _kvm_mips_host_tlb_inv(unsigned long entryhi)
 | 
					static int _kvm_mips_host_tlb_inv(unsigned long entryhi)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
| 
						 | 
					@ -163,54 +74,6 @@ static int _kvm_mips_host_tlb_inv(unsigned long entryhi)
 | 
				
			||||||
	return idx;
 | 
						return idx;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int kvm_mips_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long va,
 | 
					 | 
				
			||||||
			  bool user, bool kernel)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	/*
 | 
					 | 
				
			||||||
	 * Initialize idx_user and idx_kernel to workaround bogus
 | 
					 | 
				
			||||||
	 * maybe-initialized warning when using GCC 6.
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
	int idx_user = 0, idx_kernel = 0;
 | 
					 | 
				
			||||||
	unsigned long flags, old_entryhi;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	local_irq_save(flags);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	old_entryhi = read_c0_entryhi();
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (user)
 | 
					 | 
				
			||||||
		idx_user = _kvm_mips_host_tlb_inv((va & VPN2_MASK) |
 | 
					 | 
				
			||||||
						  kvm_mips_get_user_asid(vcpu));
 | 
					 | 
				
			||||||
	if (kernel)
 | 
					 | 
				
			||||||
		idx_kernel = _kvm_mips_host_tlb_inv((va & VPN2_MASK) |
 | 
					 | 
				
			||||||
						kvm_mips_get_kernel_asid(vcpu));
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	write_c0_entryhi(old_entryhi);
 | 
					 | 
				
			||||||
	mtc0_tlbw_hazard();
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	local_irq_restore(flags);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/*
 | 
					 | 
				
			||||||
	 * We don't want to get reserved instruction exceptions for missing tlb
 | 
					 | 
				
			||||||
	 * entries.
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
	if (cpu_has_vtag_icache)
 | 
					 | 
				
			||||||
		flush_icache_all();
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (user && idx_user >= 0)
 | 
					 | 
				
			||||||
		kvm_debug("%s: Invalidated guest user entryhi %#lx @ idx %d\n",
 | 
					 | 
				
			||||||
			  __func__, (va & VPN2_MASK) |
 | 
					 | 
				
			||||||
				    kvm_mips_get_user_asid(vcpu), idx_user);
 | 
					 | 
				
			||||||
	if (kernel && idx_kernel >= 0)
 | 
					 | 
				
			||||||
		kvm_debug("%s: Invalidated guest kernel entryhi %#lx @ idx %d\n",
 | 
					 | 
				
			||||||
			  __func__, (va & VPN2_MASK) |
 | 
					 | 
				
			||||||
				    kvm_mips_get_kernel_asid(vcpu), idx_kernel);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return 0;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
EXPORT_SYMBOL_GPL(kvm_mips_host_tlb_inv);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
#ifdef CONFIG_KVM_MIPS_VZ
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/* GuestID management */
 | 
					/* GuestID management */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/**
 | 
					/**
 | 
				
			||||||
| 
						 | 
					@ -661,40 +524,3 @@ void kvm_loongson_clear_guest_ftlb(void)
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL_GPL(kvm_loongson_clear_guest_ftlb);
 | 
					EXPORT_SYMBOL_GPL(kvm_loongson_clear_guest_ftlb);
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
 | 
					 | 
				
			||||||
#endif
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/**
 | 
					 | 
				
			||||||
 * kvm_mips_suspend_mm() - Suspend the active mm.
 | 
					 | 
				
			||||||
 * @cpu		The CPU we're running on.
 | 
					 | 
				
			||||||
 *
 | 
					 | 
				
			||||||
 * Suspend the active_mm, ready for a switch to a KVM guest virtual address
 | 
					 | 
				
			||||||
 * space. This is left active for the duration of guest context, including time
 | 
					 | 
				
			||||||
 * with interrupts enabled, so we need to be careful not to confuse e.g. cache
 | 
					 | 
				
			||||||
 * management IPIs.
 | 
					 | 
				
			||||||
 *
 | 
					 | 
				
			||||||
 * kvm_mips_resume_mm() should be called before context switching to a different
 | 
					 | 
				
			||||||
 * process so we don't need to worry about reference counting.
 | 
					 | 
				
			||||||
 *
 | 
					 | 
				
			||||||
 * This needs to be in static kernel code to avoid exporting init_mm.
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
void kvm_mips_suspend_mm(int cpu)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	cpumask_clear_cpu(cpu, mm_cpumask(current->active_mm));
 | 
					 | 
				
			||||||
	current->active_mm = &init_mm;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
EXPORT_SYMBOL_GPL(kvm_mips_suspend_mm);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/**
 | 
					 | 
				
			||||||
 * kvm_mips_resume_mm() - Resume the current process mm.
 | 
					 | 
				
			||||||
 * @cpu		The CPU we're running on.
 | 
					 | 
				
			||||||
 *
 | 
					 | 
				
			||||||
 * Resume the mm of the current process, after a switch back from a KVM guest
 | 
					 | 
				
			||||||
 * virtual address space (see kvm_mips_suspend_mm()).
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
void kvm_mips_resume_mm(int cpu)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	cpumask_set_cpu(cpu, mm_cpumask(current->mm));
 | 
					 | 
				
			||||||
	current->active_mm = current->mm;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
EXPORT_SYMBOL_GPL(kvm_mips_resume_mm);
 | 
					 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
										
											
												File diff suppressed because it is too large
												Load diff
											
										
									
								
							| 
						 | 
					@ -292,9 +292,8 @@ static int kvm_vz_irq_clear_cb(struct kvm_vcpu *vcpu, unsigned int priority,
 | 
				
			||||||
	switch (priority) {
 | 
						switch (priority) {
 | 
				
			||||||
	case MIPS_EXC_INT_TIMER:
 | 
						case MIPS_EXC_INT_TIMER:
 | 
				
			||||||
		/*
 | 
							/*
 | 
				
			||||||
		 * Call to kvm_write_c0_guest_compare() clears Cause.TI in
 | 
							 * Explicitly clear irq associated with Cause.IP[IPTI]
 | 
				
			||||||
		 * kvm_mips_emulate_CP0(). Explicitly clear irq associated with
 | 
							 * if GuestCtl2 virtual interrupt register not
 | 
				
			||||||
		 * Cause.IP[IPTI] if GuestCtl2 virtual interrupt register not
 | 
					 | 
				
			||||||
		 * supported or if not using GuestCtl2 Hardware Clear.
 | 
							 * supported or if not using GuestCtl2 Hardware Clear.
 | 
				
			||||||
		 */
 | 
							 */
 | 
				
			||||||
		if (cpu_has_guestctl2) {
 | 
							if (cpu_has_guestctl2) {
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue