Commit 25c0de2c authored by Avi Kivity's avatar Avi Kivity Committed by Linus Torvalds

[PATCH] KVM: MMU: Make kvm_mmu_alloc_page() return a kvm_mmu_page pointer

This allows further manipulation on the shadow page table.
Signed-off-by: default avatarAvi Kivity <avi@qumranet.com>
Acked-by: default avatarIngo Molnar <mingo@elte.hu>
Signed-off-by: default avatarAndrew Morton <akpm@osdl.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@osdl.org>
parent aef3d3fe
...@@ -292,12 +292,13 @@ static int is_empty_shadow_page(hpa_t page_hpa) ...@@ -292,12 +292,13 @@ static int is_empty_shadow_page(hpa_t page_hpa)
return 1; return 1;
} }
static hpa_t kvm_mmu_alloc_page(struct kvm_vcpu *vcpu, u64 *parent_pte) static struct kvm_mmu_page *kvm_mmu_alloc_page(struct kvm_vcpu *vcpu,
u64 *parent_pte)
{ {
struct kvm_mmu_page *page; struct kvm_mmu_page *page;
if (list_empty(&vcpu->free_pages)) if (list_empty(&vcpu->free_pages))
return INVALID_PAGE; return NULL;
page = list_entry(vcpu->free_pages.next, struct kvm_mmu_page, link); page = list_entry(vcpu->free_pages.next, struct kvm_mmu_page, link);
list_del(&page->link); list_del(&page->link);
...@@ -306,7 +307,7 @@ static hpa_t kvm_mmu_alloc_page(struct kvm_vcpu *vcpu, u64 *parent_pte) ...@@ -306,7 +307,7 @@ static hpa_t kvm_mmu_alloc_page(struct kvm_vcpu *vcpu, u64 *parent_pte)
page->slot_bitmap = 0; page->slot_bitmap = 0;
page->global = 1; page->global = 1;
page->parent_pte = parent_pte; page->parent_pte = parent_pte;
return page->page_hpa; return page;
} }
static void page_header_update_slot(struct kvm *kvm, void *pte, gpa_t gpa) static void page_header_update_slot(struct kvm *kvm, void *pte, gpa_t gpa)
...@@ -402,19 +403,16 @@ static int nonpaging_map(struct kvm_vcpu *vcpu, gva_t v, hpa_t p) ...@@ -402,19 +403,16 @@ static int nonpaging_map(struct kvm_vcpu *vcpu, gva_t v, hpa_t p)
} }
if (table[index] == 0) { if (table[index] == 0) {
hpa_t new_table = kvm_mmu_alloc_page(vcpu, struct kvm_mmu_page *new_table;
&table[index]);
if (!VALID_PAGE(new_table)) { new_table = kvm_mmu_alloc_page(vcpu, &table[index]);
if (!new_table) {
pgprintk("nonpaging_map: ENOMEM\n"); pgprintk("nonpaging_map: ENOMEM\n");
return -ENOMEM; return -ENOMEM;
} }
if (level == PT32E_ROOT_LEVEL) table[index] = new_table->page_hpa | PT_PRESENT_MASK
table[index] = new_table | PT_PRESENT_MASK; | PT_WRITABLE_MASK | PT_USER_MASK;
else
table[index] = new_table | PT_PRESENT_MASK |
PT_WRITABLE_MASK | PT_USER_MASK;
} }
table_addr = table[index] & PT64_BASE_ADDR_MASK; table_addr = table[index] & PT64_BASE_ADDR_MASK;
} }
...@@ -454,7 +452,7 @@ static void mmu_alloc_roots(struct kvm_vcpu *vcpu) ...@@ -454,7 +452,7 @@ static void mmu_alloc_roots(struct kvm_vcpu *vcpu)
hpa_t root = vcpu->mmu.root_hpa; hpa_t root = vcpu->mmu.root_hpa;
ASSERT(!VALID_PAGE(root)); ASSERT(!VALID_PAGE(root));
root = kvm_mmu_alloc_page(vcpu, NULL); root = kvm_mmu_alloc_page(vcpu, NULL)->page_hpa;
vcpu->mmu.root_hpa = root; vcpu->mmu.root_hpa = root;
return; return;
} }
...@@ -463,7 +461,7 @@ static void mmu_alloc_roots(struct kvm_vcpu *vcpu) ...@@ -463,7 +461,7 @@ static void mmu_alloc_roots(struct kvm_vcpu *vcpu)
hpa_t root = vcpu->mmu.pae_root[i]; hpa_t root = vcpu->mmu.pae_root[i];
ASSERT(!VALID_PAGE(root)); ASSERT(!VALID_PAGE(root));
root = kvm_mmu_alloc_page(vcpu, NULL); root = kvm_mmu_alloc_page(vcpu, NULL)->page_hpa;
vcpu->mmu.pae_root[i] = root | PT_PRESENT_MASK; vcpu->mmu.pae_root[i] = root | PT_PRESENT_MASK;
} }
vcpu->mmu.root_hpa = __pa(vcpu->mmu.pae_root); vcpu->mmu.root_hpa = __pa(vcpu->mmu.pae_root);
......
...@@ -179,6 +179,7 @@ static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr, ...@@ -179,6 +179,7 @@ static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr,
for (; ; level--) { for (; ; level--) {
u32 index = SHADOW_PT_INDEX(addr, level); u32 index = SHADOW_PT_INDEX(addr, level);
u64 *shadow_ent = ((u64 *)__va(shadow_addr)) + index; u64 *shadow_ent = ((u64 *)__va(shadow_addr)) + index;
struct kvm_mmu_page *shadow_page;
u64 shadow_pte; u64 shadow_pte;
if (is_present_pte(*shadow_ent) || is_io_pte(*shadow_ent)) { if (is_present_pte(*shadow_ent) || is_io_pte(*shadow_ent)) {
...@@ -204,9 +205,10 @@ static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr, ...@@ -204,9 +205,10 @@ static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr,
return shadow_ent; return shadow_ent;
} }
shadow_addr = kvm_mmu_alloc_page(vcpu, shadow_ent); shadow_page = kvm_mmu_alloc_page(vcpu, shadow_ent);
if (!VALID_PAGE(shadow_addr)) if (!shadow_page)
return ERR_PTR(-ENOMEM); return ERR_PTR(-ENOMEM);
shadow_addr = shadow_page->page_hpa;
shadow_pte = shadow_addr | PT_PRESENT_MASK | PT_ACCESSED_MASK shadow_pte = shadow_addr | PT_PRESENT_MASK | PT_ACCESSED_MASK
| PT_WRITABLE_MASK | PT_USER_MASK; | PT_WRITABLE_MASK | PT_USER_MASK;
*shadow_ent = shadow_pte; *shadow_ent = shadow_pte;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment