This eliminates needing to have an arch operation to get cr0.
Signed-off-by: Avi Kivity <[email protected]>
Index: linux-2.6/drivers/kvm/kvm.h
===================================================================
--- linux-2.6.orig/drivers/kvm/kvm.h
+++ linux-2.6/drivers/kvm/kvm.h
@@ -41,7 +41,6 @@
| CR0_NW_MASK | CR0_CD_MASK)
#define KVM_VM_CR0_ALWAYS_ON \
(CR0_PG_MASK | CR0_PE_MASK | CR0_WP_MASK | CR0_NE_MASK)
-
#define KVM_GUEST_CR4_MASK \
(CR4_PSE_MASK | CR4_PAE_MASK | CR4_PGE_MASK | CR4_VMXE_MASK | CR4_VME_MASK)
#define KVM_PMODE_VM_CR4_ALWAYS_ON (CR4_VMXE_MASK | CR4_PAE_MASK)
@@ -166,6 +165,7 @@ struct kvm_vcpu {
unsigned long regs[NR_VCPU_REGS]; /* for rsp: vcpu_load_rsp_rip() */
unsigned long rip; /* needs vcpu_load_rsp_rip() */
+ unsigned long cr0;
unsigned long cr2;
unsigned long cr3;
unsigned long cr4;
@@ -346,20 +346,14 @@ static inline int is_pse(struct kvm_vcpu
return vcpu->cr4 & CR4_PSE_MASK;
}
-static inline unsigned long guest_cr0(void)
-{
- return (vmcs_readl(CR0_READ_SHADOW) & KVM_GUEST_CR0_MASK) |
- (vmcs_readl(GUEST_CR0) & ~KVM_GUEST_CR0_MASK);
-}
-
static inline unsigned guest_cpl(void)
{
return vmcs_read16(GUEST_CS_SELECTOR) & SELECTOR_RPL_MASK;
}
-static inline int is_paging(void)
+static inline int is_paging(struct kvm_vcpu *vcpu)
{
- return guest_cr0() & CR0_PG_MASK;
+ return vcpu->cr0 & CR0_PG_MASK;
}
static inline int is_page_fault(u32 intr_info)
Index: linux-2.6/drivers/kvm/kvm_main.c
===================================================================
--- linux-2.6.orig/drivers/kvm/kvm_main.c
+++ linux-2.6/drivers/kvm/kvm_main.c
@@ -875,9 +875,9 @@ static void __set_cr0(struct kvm_vcpu *v
#ifdef __x86_64__
if (vcpu->shadow_efer & EFER_LME) {
- if (!is_paging() && (cr0 & CR0_PG_MASK))
+ if (!is_paging(vcpu) && (cr0 & CR0_PG_MASK))
enter_lmode(vcpu);
- if (is_paging() && !(cr0 & CR0_PG_MASK))
+ if (is_paging(vcpu) && !(cr0 & CR0_PG_MASK))
exit_lmode(vcpu);
}
#endif
@@ -885,6 +885,7 @@ static void __set_cr0(struct kvm_vcpu *v
vmcs_writel(CR0_READ_SHADOW, cr0);
vmcs_writel(GUEST_CR0,
(cr0 & ~KVM_GUEST_CR0_MASK) | KVM_VM_CR0_ALWAYS_ON);
+ vcpu->cr0 = cr0;
}
static int pdptrs_have_reserved_bits_set(struct kvm_vcpu *vcpu,
@@ -918,7 +919,7 @@ static void set_cr0(struct kvm_vcpu *vcp
{
if (cr0 & CR0_RESEVED_BITS) {
printk(KERN_DEBUG "set_cr0: 0x%lx #GP, reserved bits 0x%lx\n",
- cr0, guest_cr0());
+ cr0, vcpu->cr0);
inject_gp(vcpu);
return;
}
@@ -936,7 +937,7 @@ static void set_cr0(struct kvm_vcpu *vcp
return;
}
- if (!is_paging() && (cr0 & CR0_PG_MASK)) {
+ if (!is_paging(vcpu) && (cr0 & CR0_PG_MASK)) {
#ifdef __x86_64__
if ((vcpu->shadow_efer & EFER_LME)) {
u32 guest_cs_ar;
@@ -975,7 +976,7 @@ static void set_cr0(struct kvm_vcpu *vcp
static void lmsw(struct kvm_vcpu *vcpu, unsigned long msw)
{
- unsigned long cr0 = guest_cr0();
+ unsigned long cr0 = vcpu->cr0;
if ((msw & CR0_PE_MASK) && !(cr0 & CR0_PE_MASK)) {
enter_pmode(vcpu);
@@ -986,6 +987,7 @@ static void lmsw(struct kvm_vcpu *vcpu,
vmcs_writel(GUEST_CR0, (vmcs_readl(GUEST_CR0) & ~LMSW_GUEST_MASK)
| (msw & LMSW_GUEST_MASK));
+ vcpu->cr0 = (vcpu->cr0 & ~0xfffful) | msw;
}
static void __set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
@@ -1011,7 +1013,7 @@ static void set_cr4(struct kvm_vcpu *vcp
inject_gp(vcpu);
return;
}
- } else if (is_paging() && !is_pae(vcpu) && (cr4 & CR4_PAE_MASK)
+ } else if (is_paging(vcpu) && !is_pae(vcpu) && (cr4 & CR4_PAE_MASK)
&& pdptrs_have_reserved_bits_set(vcpu, vcpu->cr3)) {
printk(KERN_DEBUG "set_cr4: #GP, pdptrs reserved bits\n");
inject_gp(vcpu);
@@ -1042,7 +1044,7 @@ static void set_cr3(struct kvm_vcpu *vcp
inject_gp(vcpu);
return;
}
- if (is_paging() && is_pae(vcpu) &&
+ if (is_paging(vcpu) && is_pae(vcpu) &&
pdptrs_have_reserved_bits_set(vcpu, cr3)) {
printk(KERN_DEBUG "set_cr3: #GP, pdptrs "
"reserved bits\n");
@@ -1897,7 +1899,7 @@ unsigned long realmode_get_cr(struct kvm
{
switch (cr) {
case 0:
- return guest_cr0();
+ return vcpu->cr0;
case 2:
return vcpu->cr2;
case 3:
@@ -1915,7 +1917,7 @@ void realmode_set_cr(struct kvm_vcpu *vc
{
switch (cr) {
case 0:
- set_cr0(vcpu, mk_cr_64(guest_cr0(), val));
+ set_cr0(vcpu, mk_cr_64(vcpu->cr0, val));
*rflags = vmcs_readl(GUEST_RFLAGS);
break;
case 2:
@@ -2264,7 +2266,8 @@ void set_efer(struct kvm_vcpu *vcpu, u64
return;
}
- if (is_paging() && (vcpu->shadow_efer & EFER_LME) != (efer & EFER_LME)) {
+ if (is_paging(vcpu)
+ && (vcpu->shadow_efer & EFER_LME) != (efer & EFER_LME)) {
printk(KERN_DEBUG "set_efer: #GP, change LME while paging\n");
inject_gp(vcpu);
return;
@@ -2842,7 +2845,7 @@ static int kvm_dev_ioctl_get_sregs(struc
get_dtable(gdt, GDTR);
#undef get_dtable
- sregs->cr0 = guest_cr0();
+ sregs->cr0 = vcpu->cr0;
sregs->cr2 = vcpu->cr2;
sregs->cr3 = vcpu->cr3;
sregs->cr4 = vcpu->cr4;
@@ -2906,12 +2909,13 @@ static int kvm_dev_ioctl_set_sregs(struc
#endif
vcpu->apic_base = sregs->apic_base;
- mmu_reset_needed |= guest_cr0() != sregs->cr0;
+ mmu_reset_needed |= vcpu->cr0 != sregs->cr0;
vcpu->rmode.active = ((sregs->cr0 & CR0_PE_MASK) == 0);
update_exception_bitmap(vcpu);
vmcs_writel(CR0_READ_SHADOW, sregs->cr0);
vmcs_writel(GUEST_CR0,
(sregs->cr0 & ~KVM_GUEST_CR0_MASK) | KVM_VM_CR0_ALWAYS_ON);
+ vcpu->cr0 = sregs->cr0;
mmu_reset_needed |= vcpu->cr4 != sregs->cr4;
__set_cr4(vcpu, sregs->cr4);
Index: linux-2.6/drivers/kvm/mmu.c
===================================================================
--- linux-2.6.orig/drivers/kvm/mmu.c
+++ linux-2.6/drivers/kvm/mmu.c
@@ -138,9 +138,9 @@
#define PT_DIRECTORY_LEVEL 2
#define PT_PAGE_TABLE_LEVEL 1
-static int is_write_protection(void)
+static int is_write_protection(struct kvm_vcpu *vcpu)
{
- return guest_cr0() & CR0_WP_MASK;
+ return vcpu->cr0 & CR0_WP_MASK;
}
static int is_cpuid_PSE36(void)
@@ -321,7 +321,7 @@ static void nonpaging_flush(struct kvm_v
root = kvm_mmu_alloc_page(vcpu, 0);
ASSERT(VALID_PAGE(root));
vcpu->mmu.root_hpa = root;
- if (is_paging())
+ if (is_paging(vcpu))
root |= (vcpu->cr3 & (CR3_PCD_MASK | CR3_WPT_MASK));
vmcs_writel(GUEST_CR3, root);
}
@@ -614,7 +614,7 @@ static int init_kvm_mmu(struct kvm_vcpu
ASSERT(vcpu);
ASSERT(!VALID_PAGE(vcpu->mmu.root_hpa));
- if (!is_paging())
+ if (!is_paging(vcpu))
return nonpaging_init_context(vcpu);
else if (is_long_mode())
return paging64_init_context(vcpu);
Index: linux-2.6/drivers/kvm/paging_tmpl.h
===================================================================
--- linux-2.6.orig/drivers/kvm/paging_tmpl.h
+++ linux-2.6/drivers/kvm/paging_tmpl.h
@@ -251,7 +251,7 @@ static int FNAME(fix_write_pf)(struct kv
* supervisor write protection is enabled.
*/
if (!writable_shadow) {
- if (is_write_protection())
+ if (is_write_protection(vcpu))
return 0;
*shadow_ent &= ~PT_USER_MASK;
}
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]