| /* | 
 |  * This file is subject to the terms and conditions of the GNU General Public | 
 |  * License.  See the file "COPYING" in the main directory of this archive | 
 |  * for more details. | 
 |  * | 
 |  * KVM/MIPS TLB handling, this file is part of the Linux host kernel so that | 
 |  * TLB handlers run from KSEG0 | 
 |  * | 
 |  * Copyright (C) 2012  MIPS Technologies, Inc.  All rights reserved. | 
 |  * Authors: Sanjay Lal <sanjayl@kymasys.com> | 
 |  */ | 
 |  | 
 | #include <linux/sched.h> | 
 | #include <linux/smp.h> | 
 | #include <linux/mm.h> | 
 | #include <linux/delay.h> | 
 | #include <linux/export.h> | 
 | #include <linux/kvm_host.h> | 
 | #include <linux/srcu.h> | 
 |  | 
 | #include <asm/cpu.h> | 
 | #include <asm/bootinfo.h> | 
 | #include <asm/mmu_context.h> | 
 | #include <asm/pgtable.h> | 
 | #include <asm/cacheflush.h> | 
 | #include <asm/tlb.h> | 
 | #include <asm/tlbdebug.h> | 
 |  | 
 | #undef CONFIG_MIPS_MT | 
 | #include <asm/r4kcache.h> | 
 | #define CONFIG_MIPS_MT | 
 |  | 
 | #define KVM_GUEST_PC_TLB    0 | 
 | #define KVM_GUEST_SP_TLB    1 | 
 |  | 
 | #ifdef CONFIG_KVM_MIPS_VZ | 
 | unsigned long GUESTID_MASK; | 
 | EXPORT_SYMBOL_GPL(GUESTID_MASK); | 
 | unsigned long GUESTID_FIRST_VERSION; | 
 | EXPORT_SYMBOL_GPL(GUESTID_FIRST_VERSION); | 
 | unsigned long GUESTID_VERSION_MASK; | 
 | EXPORT_SYMBOL_GPL(GUESTID_VERSION_MASK); | 
 |  | 
 | static u32 kvm_mips_get_root_asid(struct kvm_vcpu *vcpu) | 
 | { | 
 | 	struct mm_struct *gpa_mm = &vcpu->kvm->arch.gpa_mm; | 
 |  | 
 | 	if (cpu_has_guestid) | 
 | 		return 0; | 
 | 	else | 
 | 		return cpu_asid(smp_processor_id(), gpa_mm); | 
 | } | 
 | #endif | 
 |  | 
 | static u32 kvm_mips_get_kernel_asid(struct kvm_vcpu *vcpu) | 
 | { | 
 | 	struct mm_struct *kern_mm = &vcpu->arch.guest_kernel_mm; | 
 | 	int cpu = smp_processor_id(); | 
 |  | 
 | 	return cpu_asid(cpu, kern_mm); | 
 | } | 
 |  | 
 | static u32 kvm_mips_get_user_asid(struct kvm_vcpu *vcpu) | 
 | { | 
 | 	struct mm_struct *user_mm = &vcpu->arch.guest_user_mm; | 
 | 	int cpu = smp_processor_id(); | 
 |  | 
 | 	return cpu_asid(cpu, user_mm); | 
 | } | 
 |  | 
 | /* Structure defining an tlb entry data set. */ | 
 |  | 
 | void kvm_mips_dump_host_tlbs(void) | 
 | { | 
 | 	unsigned long flags; | 
 |  | 
 | 	local_irq_save(flags); | 
 |  | 
 | 	kvm_info("HOST TLBs:\n"); | 
 | 	dump_tlb_regs(); | 
 | 	pr_info("\n"); | 
 | 	dump_tlb_all(); | 
 |  | 
 | 	local_irq_restore(flags); | 
 | } | 
 | EXPORT_SYMBOL_GPL(kvm_mips_dump_host_tlbs); | 
 |  | 
 | void kvm_mips_dump_guest_tlbs(struct kvm_vcpu *vcpu) | 
 | { | 
 | 	struct mips_coproc *cop0 = vcpu->arch.cop0; | 
 | 	struct kvm_mips_tlb tlb; | 
 | 	int i; | 
 |  | 
 | 	kvm_info("Guest TLBs:\n"); | 
 | 	kvm_info("Guest EntryHi: %#lx\n", kvm_read_c0_guest_entryhi(cop0)); | 
 |  | 
 | 	for (i = 0; i < KVM_MIPS_GUEST_TLB_SIZE; i++) { | 
 | 		tlb = vcpu->arch.guest_tlb[i]; | 
 | 		kvm_info("TLB%c%3d Hi 0x%08lx ", | 
 | 			 (tlb.tlb_lo[0] | tlb.tlb_lo[1]) & ENTRYLO_V | 
 | 							? ' ' : '*', | 
 | 			 i, tlb.tlb_hi); | 
 | 		kvm_info("Lo0=0x%09llx %c%c attr %lx ", | 
 | 			 (u64) mips3_tlbpfn_to_paddr(tlb.tlb_lo[0]), | 
 | 			 (tlb.tlb_lo[0] & ENTRYLO_D) ? 'D' : ' ', | 
 | 			 (tlb.tlb_lo[0] & ENTRYLO_G) ? 'G' : ' ', | 
 | 			 (tlb.tlb_lo[0] & ENTRYLO_C) >> ENTRYLO_C_SHIFT); | 
 | 		kvm_info("Lo1=0x%09llx %c%c attr %lx sz=%lx\n", | 
 | 			 (u64) mips3_tlbpfn_to_paddr(tlb.tlb_lo[1]), | 
 | 			 (tlb.tlb_lo[1] & ENTRYLO_D) ? 'D' : ' ', | 
 | 			 (tlb.tlb_lo[1] & ENTRYLO_G) ? 'G' : ' ', | 
 | 			 (tlb.tlb_lo[1] & ENTRYLO_C) >> ENTRYLO_C_SHIFT, | 
 | 			 tlb.tlb_mask); | 
 | 	} | 
 | } | 
 | EXPORT_SYMBOL_GPL(kvm_mips_dump_guest_tlbs); | 
 |  | 
 | int kvm_mips_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long entryhi) | 
 | { | 
 | 	int i; | 
 | 	int index = -1; | 
 | 	struct kvm_mips_tlb *tlb = vcpu->arch.guest_tlb; | 
 |  | 
 | 	for (i = 0; i < KVM_MIPS_GUEST_TLB_SIZE; i++) { | 
 | 		if (TLB_HI_VPN2_HIT(tlb[i], entryhi) && | 
 | 		    TLB_HI_ASID_HIT(tlb[i], entryhi)) { | 
 | 			index = i; | 
 | 			break; | 
 | 		} | 
 | 	} | 
 |  | 
 | 	kvm_debug("%s: entryhi: %#lx, index: %d lo0: %#lx, lo1: %#lx\n", | 
 | 		  __func__, entryhi, index, tlb[i].tlb_lo[0], tlb[i].tlb_lo[1]); | 
 |  | 
 | 	return index; | 
 | } | 
 | EXPORT_SYMBOL_GPL(kvm_mips_guest_tlb_lookup); | 
 |  | 
 | static int _kvm_mips_host_tlb_inv(unsigned long entryhi) | 
 | { | 
 | 	int idx; | 
 |  | 
 | 	write_c0_entryhi(entryhi); | 
 | 	mtc0_tlbw_hazard(); | 
 |  | 
 | 	tlb_probe(); | 
 | 	tlb_probe_hazard(); | 
 | 	idx = read_c0_index(); | 
 |  | 
 | 	if (idx >= current_cpu_data.tlbsize) | 
 | 		BUG(); | 
 |  | 
 | 	if (idx >= 0) { | 
 | 		write_c0_entryhi(UNIQUE_ENTRYHI(idx)); | 
 | 		write_c0_entrylo0(0); | 
 | 		write_c0_entrylo1(0); | 
 | 		mtc0_tlbw_hazard(); | 
 |  | 
 | 		tlb_write_indexed(); | 
 | 		tlbw_use_hazard(); | 
 | 	} | 
 |  | 
 | 	return idx; | 
 | } | 
 |  | 
 | int kvm_mips_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long va, | 
 | 			  bool user, bool kernel) | 
 | { | 
 | 	/* | 
 | 	 * Initialize idx_user and idx_kernel to workaround bogus | 
 | 	 * maybe-initialized warning when using GCC 6. | 
 | 	 */ | 
 | 	int idx_user = 0, idx_kernel = 0; | 
 | 	unsigned long flags, old_entryhi; | 
 |  | 
 | 	local_irq_save(flags); | 
 |  | 
 | 	old_entryhi = read_c0_entryhi(); | 
 |  | 
 | 	if (user) | 
 | 		idx_user = _kvm_mips_host_tlb_inv((va & VPN2_MASK) | | 
 | 						  kvm_mips_get_user_asid(vcpu)); | 
 | 	if (kernel) | 
 | 		idx_kernel = _kvm_mips_host_tlb_inv((va & VPN2_MASK) | | 
 | 						kvm_mips_get_kernel_asid(vcpu)); | 
 |  | 
 | 	write_c0_entryhi(old_entryhi); | 
 | 	mtc0_tlbw_hazard(); | 
 |  | 
 | 	local_irq_restore(flags); | 
 |  | 
 | 	/* | 
 | 	 * We don't want to get reserved instruction exceptions for missing tlb | 
 | 	 * entries. | 
 | 	 */ | 
 | 	if (cpu_has_vtag_icache) | 
 | 		flush_icache_all(); | 
 |  | 
 | 	if (user && idx_user >= 0) | 
 | 		kvm_debug("%s: Invalidated guest user entryhi %#lx @ idx %d\n", | 
 | 			  __func__, (va & VPN2_MASK) | | 
 | 				    kvm_mips_get_user_asid(vcpu), idx_user); | 
 | 	if (kernel && idx_kernel >= 0) | 
 | 		kvm_debug("%s: Invalidated guest kernel entryhi %#lx @ idx %d\n", | 
 | 			  __func__, (va & VPN2_MASK) | | 
 | 				    kvm_mips_get_kernel_asid(vcpu), idx_kernel); | 
 |  | 
 | 	return 0; | 
 | } | 
 | EXPORT_SYMBOL_GPL(kvm_mips_host_tlb_inv); | 
 |  | 
 | #ifdef CONFIG_KVM_MIPS_VZ | 
 |  | 
 | /* GuestID management */ | 
 |  | 
 | /** | 
 |  * clear_root_gid() - Set GuestCtl1.RID for normal root operation. | 
 |  */ | 
 | static inline void clear_root_gid(void) | 
 | { | 
 | 	if (cpu_has_guestid) { | 
 | 		clear_c0_guestctl1(MIPS_GCTL1_RID); | 
 | 		mtc0_tlbw_hazard(); | 
 | 	} | 
 | } | 
 |  | 
 | /** | 
 |  * set_root_gid_to_guest_gid() - Set GuestCtl1.RID to match GuestCtl1.ID. | 
 |  * | 
 |  * Sets the root GuestID to match the current guest GuestID, for TLB operation | 
 |  * on the GPA->RPA mappings in the root TLB. | 
 |  * | 
 |  * The caller must be sure to disable HTW while the root GID is set, and | 
 |  * possibly longer if TLB registers are modified. | 
 |  */ | 
 | static inline void set_root_gid_to_guest_gid(void) | 
 | { | 
 | 	unsigned int guestctl1; | 
 |  | 
 | 	if (cpu_has_guestid) { | 
 | 		back_to_back_c0_hazard(); | 
 | 		guestctl1 = read_c0_guestctl1(); | 
 | 		guestctl1 = (guestctl1 & ~MIPS_GCTL1_RID) | | 
 | 			((guestctl1 & MIPS_GCTL1_ID) >> MIPS_GCTL1_ID_SHIFT) | 
 | 						     << MIPS_GCTL1_RID_SHIFT; | 
 | 		write_c0_guestctl1(guestctl1); | 
 | 		mtc0_tlbw_hazard(); | 
 | 	} | 
 | } | 
 |  | 
 | int kvm_vz_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long va) | 
 | { | 
 | 	int idx; | 
 | 	unsigned long flags, old_entryhi; | 
 |  | 
 | 	local_irq_save(flags); | 
 | 	htw_stop(); | 
 |  | 
 | 	/* Set root GuestID for root probe and write of guest TLB entry */ | 
 | 	set_root_gid_to_guest_gid(); | 
 |  | 
 | 	old_entryhi = read_c0_entryhi(); | 
 |  | 
 | 	idx = _kvm_mips_host_tlb_inv((va & VPN2_MASK) | | 
 | 				     kvm_mips_get_root_asid(vcpu)); | 
 |  | 
 | 	write_c0_entryhi(old_entryhi); | 
 | 	clear_root_gid(); | 
 | 	mtc0_tlbw_hazard(); | 
 |  | 
 | 	htw_start(); | 
 | 	local_irq_restore(flags); | 
 |  | 
 | 	/* | 
 | 	 * We don't want to get reserved instruction exceptions for missing tlb | 
 | 	 * entries. | 
 | 	 */ | 
 | 	if (cpu_has_vtag_icache) | 
 | 		flush_icache_all(); | 
 |  | 
 | 	if (idx > 0) | 
 | 		kvm_debug("%s: Invalidated root entryhi %#lx @ idx %d\n", | 
 | 			  __func__, (va & VPN2_MASK) | | 
 | 				    kvm_mips_get_root_asid(vcpu), idx); | 
 |  | 
 | 	return 0; | 
 | } | 
 | EXPORT_SYMBOL_GPL(kvm_vz_host_tlb_inv); | 
 |  | 
 | /** | 
 |  * kvm_vz_guest_tlb_lookup() - Lookup a guest VZ TLB mapping. | 
 |  * @vcpu:	KVM VCPU pointer. | 
 |  * @gpa:	Guest virtual address in a TLB mapped guest segment. | 
 |  * @gpa:	Ponter to output guest physical address it maps to. | 
 |  * | 
 |  * Converts a guest virtual address in a guest TLB mapped segment to a guest | 
 |  * physical address, by probing the guest TLB. | 
 |  * | 
 |  * Returns:	0 if guest TLB mapping exists for @gva. *@gpa will have been | 
 |  *		written. | 
 |  *		-EFAULT if no guest TLB mapping exists for @gva. *@gpa may not | 
 |  *		have been written. | 
 |  */ | 
 | int kvm_vz_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long gva, | 
 | 			    unsigned long *gpa) | 
 | { | 
 | 	unsigned long o_entryhi, o_entrylo[2], o_pagemask; | 
 | 	unsigned int o_index; | 
 | 	unsigned long entrylo[2], pagemask, pagemaskbit, pa; | 
 | 	unsigned long flags; | 
 | 	int index; | 
 |  | 
 | 	/* Probe the guest TLB for a mapping */ | 
 | 	local_irq_save(flags); | 
 | 	/* Set root GuestID for root probe of guest TLB entry */ | 
 | 	htw_stop(); | 
 | 	set_root_gid_to_guest_gid(); | 
 |  | 
 | 	o_entryhi = read_gc0_entryhi(); | 
 | 	o_index = read_gc0_index(); | 
 |  | 
 | 	write_gc0_entryhi((o_entryhi & 0x3ff) | (gva & ~0xfffl)); | 
 | 	mtc0_tlbw_hazard(); | 
 | 	guest_tlb_probe(); | 
 | 	tlb_probe_hazard(); | 
 |  | 
 | 	index = read_gc0_index(); | 
 | 	if (index < 0) { | 
 | 		/* No match, fail */ | 
 | 		write_gc0_entryhi(o_entryhi); | 
 | 		write_gc0_index(o_index); | 
 |  | 
 | 		clear_root_gid(); | 
 | 		htw_start(); | 
 | 		local_irq_restore(flags); | 
 | 		return -EFAULT; | 
 | 	} | 
 |  | 
 | 	/* Match! read the TLB entry */ | 
 | 	o_entrylo[0] = read_gc0_entrylo0(); | 
 | 	o_entrylo[1] = read_gc0_entrylo1(); | 
 | 	o_pagemask = read_gc0_pagemask(); | 
 |  | 
 | 	mtc0_tlbr_hazard(); | 
 | 	guest_tlb_read(); | 
 | 	tlb_read_hazard(); | 
 |  | 
 | 	entrylo[0] = read_gc0_entrylo0(); | 
 | 	entrylo[1] = read_gc0_entrylo1(); | 
 | 	pagemask = ~read_gc0_pagemask() & ~0x1fffl; | 
 |  | 
 | 	write_gc0_entryhi(o_entryhi); | 
 | 	write_gc0_index(o_index); | 
 | 	write_gc0_entrylo0(o_entrylo[0]); | 
 | 	write_gc0_entrylo1(o_entrylo[1]); | 
 | 	write_gc0_pagemask(o_pagemask); | 
 |  | 
 | 	clear_root_gid(); | 
 | 	htw_start(); | 
 | 	local_irq_restore(flags); | 
 |  | 
 | 	/* Select one of the EntryLo values and interpret the GPA */ | 
 | 	pagemaskbit = (pagemask ^ (pagemask & (pagemask - 1))) >> 1; | 
 | 	pa = entrylo[!!(gva & pagemaskbit)]; | 
 |  | 
 | 	/* | 
 | 	 * TLB entry may have become invalid since TLB probe if physical FTLB | 
 | 	 * entries are shared between threads (e.g. I6400). | 
 | 	 */ | 
 | 	if (!(pa & ENTRYLO_V)) | 
 | 		return -EFAULT; | 
 |  | 
 | 	/* | 
 | 	 * Note, this doesn't take guest MIPS32 XPA into account, where PFN is | 
 | 	 * split with XI/RI in the middle. | 
 | 	 */ | 
 | 	pa = (pa << 6) & ~0xfffl; | 
 | 	pa |= gva & ~(pagemask | pagemaskbit); | 
 |  | 
 | 	*gpa = pa; | 
 | 	return 0; | 
 | } | 
 | EXPORT_SYMBOL_GPL(kvm_vz_guest_tlb_lookup); | 
 |  | 
 | /** | 
 |  * kvm_vz_local_flush_roottlb_all_guests() - Flush all root TLB entries for | 
 |  * guests. | 
 |  * | 
 |  * Invalidate all entries in root tlb which are GPA mappings. | 
 |  */ | 
 | void kvm_vz_local_flush_roottlb_all_guests(void) | 
 | { | 
 | 	unsigned long flags; | 
 | 	unsigned long old_entryhi, old_pagemask, old_guestctl1; | 
 | 	int entry; | 
 |  | 
 | 	if (WARN_ON(!cpu_has_guestid)) | 
 | 		return; | 
 |  | 
 | 	local_irq_save(flags); | 
 | 	htw_stop(); | 
 |  | 
 | 	/* TLBR may clobber EntryHi.ASID, PageMask, and GuestCtl1.RID */ | 
 | 	old_entryhi = read_c0_entryhi(); | 
 | 	old_pagemask = read_c0_pagemask(); | 
 | 	old_guestctl1 = read_c0_guestctl1(); | 
 |  | 
 | 	/* | 
 | 	 * Invalidate guest entries in root TLB while leaving root entries | 
 | 	 * intact when possible. | 
 | 	 */ | 
 | 	for (entry = 0; entry < current_cpu_data.tlbsize; entry++) { | 
 | 		write_c0_index(entry); | 
 | 		mtc0_tlbw_hazard(); | 
 | 		tlb_read(); | 
 | 		tlb_read_hazard(); | 
 |  | 
 | 		/* Don't invalidate non-guest (RVA) mappings in the root TLB */ | 
 | 		if (!(read_c0_guestctl1() & MIPS_GCTL1_RID)) | 
 | 			continue; | 
 |  | 
 | 		/* Make sure all entries differ. */ | 
 | 		write_c0_entryhi(UNIQUE_ENTRYHI(entry)); | 
 | 		write_c0_entrylo0(0); | 
 | 		write_c0_entrylo1(0); | 
 | 		write_c0_guestctl1(0); | 
 | 		mtc0_tlbw_hazard(); | 
 | 		tlb_write_indexed(); | 
 | 	} | 
 |  | 
 | 	write_c0_entryhi(old_entryhi); | 
 | 	write_c0_pagemask(old_pagemask); | 
 | 	write_c0_guestctl1(old_guestctl1); | 
 | 	tlbw_use_hazard(); | 
 |  | 
 | 	htw_start(); | 
 | 	local_irq_restore(flags); | 
 | } | 
 | EXPORT_SYMBOL_GPL(kvm_vz_local_flush_roottlb_all_guests); | 
 |  | 
 | /** | 
 |  * kvm_vz_local_flush_guesttlb_all() - Flush all guest TLB entries. | 
 |  * | 
 |  * Invalidate all entries in guest tlb irrespective of guestid. | 
 |  */ | 
 | void kvm_vz_local_flush_guesttlb_all(void) | 
 | { | 
 | 	unsigned long flags; | 
 | 	unsigned long old_index; | 
 | 	unsigned long old_entryhi; | 
 | 	unsigned long old_entrylo[2]; | 
 | 	unsigned long old_pagemask; | 
 | 	int entry; | 
 | 	u64 cvmmemctl2 = 0; | 
 |  | 
 | 	local_irq_save(flags); | 
 |  | 
 | 	/* Preserve all clobbered guest registers */ | 
 | 	old_index = read_gc0_index(); | 
 | 	old_entryhi = read_gc0_entryhi(); | 
 | 	old_entrylo[0] = read_gc0_entrylo0(); | 
 | 	old_entrylo[1] = read_gc0_entrylo1(); | 
 | 	old_pagemask = read_gc0_pagemask(); | 
 |  | 
 | 	switch (current_cpu_type()) { | 
 | 	case CPU_CAVIUM_OCTEON3: | 
 | 		/* Inhibit machine check due to multiple matching TLB entries */ | 
 | 		cvmmemctl2 = read_c0_cvmmemctl2(); | 
 | 		cvmmemctl2 |= CVMMEMCTL2_INHIBITTS; | 
 | 		write_c0_cvmmemctl2(cvmmemctl2); | 
 | 		break; | 
 | 	}; | 
 |  | 
 | 	/* Invalidate guest entries in guest TLB */ | 
 | 	write_gc0_entrylo0(0); | 
 | 	write_gc0_entrylo1(0); | 
 | 	write_gc0_pagemask(0); | 
 | 	for (entry = 0; entry < current_cpu_data.guest.tlbsize; entry++) { | 
 | 		/* Make sure all entries differ. */ | 
 | 		write_gc0_index(entry); | 
 | 		write_gc0_entryhi(UNIQUE_GUEST_ENTRYHI(entry)); | 
 | 		mtc0_tlbw_hazard(); | 
 | 		guest_tlb_write_indexed(); | 
 | 	} | 
 |  | 
 | 	if (cvmmemctl2) { | 
 | 		cvmmemctl2 &= ~CVMMEMCTL2_INHIBITTS; | 
 | 		write_c0_cvmmemctl2(cvmmemctl2); | 
 | 	}; | 
 |  | 
 | 	write_gc0_index(old_index); | 
 | 	write_gc0_entryhi(old_entryhi); | 
 | 	write_gc0_entrylo0(old_entrylo[0]); | 
 | 	write_gc0_entrylo1(old_entrylo[1]); | 
 | 	write_gc0_pagemask(old_pagemask); | 
 | 	tlbw_use_hazard(); | 
 |  | 
 | 	local_irq_restore(flags); | 
 | } | 
 | EXPORT_SYMBOL_GPL(kvm_vz_local_flush_guesttlb_all); | 
 |  | 
 | /** | 
 |  * kvm_vz_save_guesttlb() - Save a range of guest TLB entries. | 
 |  * @buf:	Buffer to write TLB entries into. | 
 |  * @index:	Start index. | 
 |  * @count:	Number of entries to save. | 
 |  * | 
 |  * Save a range of guest TLB entries. The caller must ensure interrupts are | 
 |  * disabled. | 
 |  */ | 
 | void kvm_vz_save_guesttlb(struct kvm_mips_tlb *buf, unsigned int index, | 
 | 			  unsigned int count) | 
 | { | 
 | 	unsigned int end = index + count; | 
 | 	unsigned long old_entryhi, old_entrylo0, old_entrylo1, old_pagemask; | 
 | 	unsigned int guestctl1 = 0; | 
 | 	int old_index, i; | 
 |  | 
 | 	/* Save registers we're about to clobber */ | 
 | 	old_index = read_gc0_index(); | 
 | 	old_entryhi = read_gc0_entryhi(); | 
 | 	old_entrylo0 = read_gc0_entrylo0(); | 
 | 	old_entrylo1 = read_gc0_entrylo1(); | 
 | 	old_pagemask = read_gc0_pagemask(); | 
 |  | 
 | 	/* Set root GuestID for root probe */ | 
 | 	htw_stop(); | 
 | 	set_root_gid_to_guest_gid(); | 
 | 	if (cpu_has_guestid) | 
 | 		guestctl1 = read_c0_guestctl1(); | 
 |  | 
 | 	/* Read each entry from guest TLB */ | 
 | 	for (i = index; i < end; ++i, ++buf) { | 
 | 		write_gc0_index(i); | 
 |  | 
 | 		mtc0_tlbr_hazard(); | 
 | 		guest_tlb_read(); | 
 | 		tlb_read_hazard(); | 
 |  | 
 | 		if (cpu_has_guestid && | 
 | 		    (read_c0_guestctl1() ^ guestctl1) & MIPS_GCTL1_RID) { | 
 | 			/* Entry invalid or belongs to another guest */ | 
 | 			buf->tlb_hi = UNIQUE_GUEST_ENTRYHI(i); | 
 | 			buf->tlb_lo[0] = 0; | 
 | 			buf->tlb_lo[1] = 0; | 
 | 			buf->tlb_mask = 0; | 
 | 		} else { | 
 | 			/* Entry belongs to the right guest */ | 
 | 			buf->tlb_hi = read_gc0_entryhi(); | 
 | 			buf->tlb_lo[0] = read_gc0_entrylo0(); | 
 | 			buf->tlb_lo[1] = read_gc0_entrylo1(); | 
 | 			buf->tlb_mask = read_gc0_pagemask(); | 
 | 		} | 
 | 	} | 
 |  | 
 | 	/* Clear root GuestID again */ | 
 | 	clear_root_gid(); | 
 | 	htw_start(); | 
 |  | 
 | 	/* Restore clobbered registers */ | 
 | 	write_gc0_index(old_index); | 
 | 	write_gc0_entryhi(old_entryhi); | 
 | 	write_gc0_entrylo0(old_entrylo0); | 
 | 	write_gc0_entrylo1(old_entrylo1); | 
 | 	write_gc0_pagemask(old_pagemask); | 
 |  | 
 | 	tlbw_use_hazard(); | 
 | } | 
 | EXPORT_SYMBOL_GPL(kvm_vz_save_guesttlb); | 
 |  | 
 | /** | 
 |  * kvm_vz_load_guesttlb() - Save a range of guest TLB entries. | 
 |  * @buf:	Buffer to read TLB entries from. | 
 |  * @index:	Start index. | 
 |  * @count:	Number of entries to load. | 
 |  * | 
 |  * Load a range of guest TLB entries. The caller must ensure interrupts are | 
 |  * disabled. | 
 |  */ | 
 | void kvm_vz_load_guesttlb(const struct kvm_mips_tlb *buf, unsigned int index, | 
 | 			  unsigned int count) | 
 | { | 
 | 	unsigned int end = index + count; | 
 | 	unsigned long old_entryhi, old_entrylo0, old_entrylo1, old_pagemask; | 
 | 	int old_index, i; | 
 |  | 
 | 	/* Save registers we're about to clobber */ | 
 | 	old_index = read_gc0_index(); | 
 | 	old_entryhi = read_gc0_entryhi(); | 
 | 	old_entrylo0 = read_gc0_entrylo0(); | 
 | 	old_entrylo1 = read_gc0_entrylo1(); | 
 | 	old_pagemask = read_gc0_pagemask(); | 
 |  | 
 | 	/* Set root GuestID for root probe */ | 
 | 	htw_stop(); | 
 | 	set_root_gid_to_guest_gid(); | 
 |  | 
 | 	/* Write each entry to guest TLB */ | 
 | 	for (i = index; i < end; ++i, ++buf) { | 
 | 		write_gc0_index(i); | 
 | 		write_gc0_entryhi(buf->tlb_hi); | 
 | 		write_gc0_entrylo0(buf->tlb_lo[0]); | 
 | 		write_gc0_entrylo1(buf->tlb_lo[1]); | 
 | 		write_gc0_pagemask(buf->tlb_mask); | 
 |  | 
 | 		mtc0_tlbw_hazard(); | 
 | 		guest_tlb_write_indexed(); | 
 | 	} | 
 |  | 
 | 	/* Clear root GuestID again */ | 
 | 	clear_root_gid(); | 
 | 	htw_start(); | 
 |  | 
 | 	/* Restore clobbered registers */ | 
 | 	write_gc0_index(old_index); | 
 | 	write_gc0_entryhi(old_entryhi); | 
 | 	write_gc0_entrylo0(old_entrylo0); | 
 | 	write_gc0_entrylo1(old_entrylo1); | 
 | 	write_gc0_pagemask(old_pagemask); | 
 |  | 
 | 	tlbw_use_hazard(); | 
 | } | 
 | EXPORT_SYMBOL_GPL(kvm_vz_load_guesttlb); | 
 |  | 
 | #endif | 
 |  | 
 | /** | 
 |  * kvm_mips_suspend_mm() - Suspend the active mm. | 
 |  * @cpu		The CPU we're running on. | 
 |  * | 
 |  * Suspend the active_mm, ready for a switch to a KVM guest virtual address | 
 |  * space. This is left active for the duration of guest context, including time | 
 |  * with interrupts enabled, so we need to be careful not to confuse e.g. cache | 
 |  * management IPIs. | 
 |  * | 
 |  * kvm_mips_resume_mm() should be called before context switching to a different | 
 |  * process so we don't need to worry about reference counting. | 
 |  * | 
 |  * This needs to be in static kernel code to avoid exporting init_mm. | 
 |  */ | 
 | void kvm_mips_suspend_mm(int cpu) | 
 | { | 
 | 	cpumask_clear_cpu(cpu, mm_cpumask(current->active_mm)); | 
 | 	current->active_mm = &init_mm; | 
 | } | 
 | EXPORT_SYMBOL_GPL(kvm_mips_suspend_mm); | 
 |  | 
 | /** | 
 |  * kvm_mips_resume_mm() - Resume the current process mm. | 
 |  * @cpu		The CPU we're running on. | 
 |  * | 
 |  * Resume the mm of the current process, after a switch back from a KVM guest | 
 |  * virtual address space (see kvm_mips_suspend_mm()). | 
 |  */ | 
 | void kvm_mips_resume_mm(int cpu) | 
 | { | 
 | 	cpumask_set_cpu(cpu, mm_cpumask(current->mm)); | 
 | 	current->active_mm = current->mm; | 
 | } | 
 | EXPORT_SYMBOL_GPL(kvm_mips_resume_mm); |