Commit 8d4e1288 authored by Anthony Liguori's avatar Anthony Liguori Committed by Avi Kivity

KVM: Allocate userspace memory for older userspace

Allocate a userspace buffer for older userspaces.  Also eliminate phys_mem
buffer.  The memset() in kvmctl really kills initial memory usage but swapping
works even with old userspaces.

A side effect is that maximum guest side is reduced for older userspace on
i386.
Signed-off-by: default avatarAnthony Liguori <aliguori@us.ibm.com>
Signed-off-by: default avatarAvi Kivity <avi@qumranet.com>
parent e56a7a28
...@@ -406,10 +406,8 @@ struct kvm_memory_slot { ...@@ -406,10 +406,8 @@ struct kvm_memory_slot {
gfn_t base_gfn; gfn_t base_gfn;
unsigned long npages; unsigned long npages;
unsigned long flags; unsigned long flags;
struct page **phys_mem;
unsigned long *rmap; unsigned long *rmap;
unsigned long *dirty_bitmap; unsigned long *dirty_bitmap;
int user_alloc; /* user allocated memory */
unsigned long userspace_addr; unsigned long userspace_addr;
}; };
......
...@@ -42,6 +42,7 @@ ...@@ -42,6 +42,7 @@
#include <linux/profile.h> #include <linux/profile.h>
#include <linux/kvm_para.h> #include <linux/kvm_para.h>
#include <linux/pagemap.h> #include <linux/pagemap.h>
#include <linux/mman.h>
#include <asm/processor.h> #include <asm/processor.h>
#include <asm/msr.h> #include <asm/msr.h>
...@@ -300,36 +301,21 @@ static struct kvm *kvm_create_vm(void) ...@@ -300,36 +301,21 @@ static struct kvm *kvm_create_vm(void)
return kvm; return kvm;
} }
static void kvm_free_kernel_physmem(struct kvm_memory_slot *free)
{
int i;
for (i = 0; i < free->npages; ++i)
if (free->phys_mem[i])
__free_page(free->phys_mem[i]);
}
/* /*
* Free any memory in @free but not in @dont. * Free any memory in @free but not in @dont.
*/ */
static void kvm_free_physmem_slot(struct kvm_memory_slot *free, static void kvm_free_physmem_slot(struct kvm_memory_slot *free,
struct kvm_memory_slot *dont) struct kvm_memory_slot *dont)
{ {
if (!dont || free->phys_mem != dont->phys_mem)
if (free->phys_mem) {
if (!free->user_alloc)
kvm_free_kernel_physmem(free);
vfree(free->phys_mem);
}
if (!dont || free->rmap != dont->rmap) if (!dont || free->rmap != dont->rmap)
vfree(free->rmap); vfree(free->rmap);
if (!dont || free->dirty_bitmap != dont->dirty_bitmap) if (!dont || free->dirty_bitmap != dont->dirty_bitmap)
vfree(free->dirty_bitmap); vfree(free->dirty_bitmap);
free->phys_mem = NULL;
free->npages = 0; free->npages = 0;
free->dirty_bitmap = NULL; free->dirty_bitmap = NULL;
free->rmap = NULL;
} }
static void kvm_free_physmem(struct kvm *kvm) static void kvm_free_physmem(struct kvm *kvm)
...@@ -712,10 +698,6 @@ static int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, ...@@ -712,10 +698,6 @@ static int kvm_vm_ioctl_set_memory_region(struct kvm *kvm,
goto out_unlock; goto out_unlock;
} }
/* Deallocate if slot is being removed */
if (!npages)
new.phys_mem = NULL;
/* Free page dirty bitmap if unneeded */ /* Free page dirty bitmap if unneeded */
if (!(new.flags & KVM_MEM_LOG_DIRTY_PAGES)) if (!(new.flags & KVM_MEM_LOG_DIRTY_PAGES))
new.dirty_bitmap = NULL; new.dirty_bitmap = NULL;
...@@ -723,29 +705,27 @@ static int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, ...@@ -723,29 +705,27 @@ static int kvm_vm_ioctl_set_memory_region(struct kvm *kvm,
r = -ENOMEM; r = -ENOMEM;
/* Allocate if a slot is being created */ /* Allocate if a slot is being created */
if (npages && !new.phys_mem) { if (npages && !new.rmap) {
new.phys_mem = vmalloc(npages * sizeof(struct page *));
if (!new.phys_mem)
goto out_unlock;
new.rmap = vmalloc(npages * sizeof(struct page *)); new.rmap = vmalloc(npages * sizeof(struct page *));
if (!new.rmap) if (!new.rmap)
goto out_unlock; goto out_unlock;
memset(new.phys_mem, 0, npages * sizeof(struct page *));
memset(new.rmap, 0, npages * sizeof(*new.rmap)); memset(new.rmap, 0, npages * sizeof(*new.rmap));
if (user_alloc) {
new.user_alloc = 1; if (user_alloc)
new.userspace_addr = mem->userspace_addr; new.userspace_addr = mem->userspace_addr;
} else { else {
for (i = 0; i < npages; ++i) { down_write(&current->mm->mmap_sem);
new.phys_mem[i] = alloc_page(GFP_HIGHUSER new.userspace_addr = do_mmap(NULL, 0,
| __GFP_ZERO); npages * PAGE_SIZE,
if (!new.phys_mem[i]) PROT_READ | PROT_WRITE,
goto out_unlock; MAP_SHARED | MAP_ANONYMOUS,
} 0);
up_write(&current->mm->mmap_sem);
if (IS_ERR((void *)new.userspace_addr))
goto out_unlock;
} }
} }
...@@ -1010,6 +990,8 @@ struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) ...@@ -1010,6 +990,8 @@ struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn)
struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn) struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn)
{ {
struct kvm_memory_slot *slot; struct kvm_memory_slot *slot;
struct page *page[1];
int npages;
gfn = unalias_gfn(kvm, gfn); gfn = unalias_gfn(kvm, gfn);
slot = __gfn_to_memslot(kvm, gfn); slot = __gfn_to_memslot(kvm, gfn);
...@@ -1017,24 +999,19 @@ struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn) ...@@ -1017,24 +999,19 @@ struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn)
get_page(bad_page); get_page(bad_page);
return bad_page; return bad_page;
} }
if (slot->user_alloc) {
struct page *page[1]; down_read(&current->mm->mmap_sem);
int npages; npages = get_user_pages(current, current->mm,
slot->userspace_addr
down_read(&current->mm->mmap_sem); + (gfn - slot->base_gfn) * PAGE_SIZE, 1,
npages = get_user_pages(current, current->mm, 1, 1, page, NULL);
slot->userspace_addr up_read(&current->mm->mmap_sem);
+ (gfn - slot->base_gfn) * PAGE_SIZE, 1, if (npages != 1) {
1, 1, page, NULL); get_page(bad_page);
up_read(&current->mm->mmap_sem); return bad_page;
if (npages != 1) {
get_page(bad_page);
return bad_page;
}
return page[0];
} }
get_page(slot->phys_mem[gfn - slot->base_gfn]);
return slot->phys_mem[gfn - slot->base_gfn]; return page[0];
} }
EXPORT_SYMBOL_GPL(gfn_to_page); EXPORT_SYMBOL_GPL(gfn_to_page);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment