Commit c40f6f8b authored by Linus Torvalds's avatar Linus Torvalds

Merge git://git.kernel.org/pub/scm/linux/kernel/git/dhowells/linux-2.6-nommu

* git://git.kernel.org/pub/scm/linux/kernel/git/dhowells/linux-2.6-nommu:
  NOMMU: Support XIP on initramfs
  NOMMU: Teach kobjsize() about VMA regions.
  FLAT: Don't attempt to expand the userspace stack to fill the space allocated
  FDPIC: Don't attempt to expand the userspace stack to fill the space allocated
  NOMMU: Improve procfs output using per-MM VMAs
  NOMMU: Make mmap allocation page trimming behaviour configurable.
  NOMMU: Make VMAs per MM as for MMU-mode linux
  NOMMU: Delete askedalloc and realalloc variables
  NOMMU: Rename ARM's struct vm_region
  NOMMU: Fix cleanup handling in ramfs_nommu_get_umapped_area()
parents 1a7d0f0b cb6ff208
...@@ -109,12 +109,18 @@ and it's also much more restricted in the latter case: ...@@ -109,12 +109,18 @@ and it's also much more restricted in the latter case:
FURTHER NOTES ON NO-MMU MMAP FURTHER NOTES ON NO-MMU MMAP
============================ ============================
(*) A request for a private mapping of less than a page in size may not return (*) A request for a private mapping of a file may return a buffer that is not
a page-aligned buffer. This is because the kernel calls kmalloc() to page-aligned. This is because XIP may take place, and the data may not be
allocate the buffer, not get_free_page(). paged aligned in the backing store.
(*) A list of all the mappings on the system is visible through /proc/maps in (*) A request for an anonymous mapping will always be page aligned. If
no-MMU mode. possible the size of the request should be a power of two otherwise some
of the space may be wasted as the kernel must allocate a power-of-2
granule but will only discard the excess if appropriately configured as
this has an effect on fragmentation.
(*) A list of all the private copy and anonymous mappings on the system is
visible through /proc/maps in no-MMU mode.
(*) A list of all the mappings in use by a process is visible through (*) A list of all the mappings in use by a process is visible through
/proc/<pid>/maps in no-MMU mode. /proc/<pid>/maps in no-MMU mode.
...@@ -242,3 +248,18 @@ PROVIDING SHAREABLE BLOCK DEVICE SUPPORT ...@@ -242,3 +248,18 @@ PROVIDING SHAREABLE BLOCK DEVICE SUPPORT
Provision of shared mappings on block device files is exactly the same as for Provision of shared mappings on block device files is exactly the same as for
character devices. If there isn't a real device underneath, then the driver character devices. If there isn't a real device underneath, then the driver
should allocate sufficient contiguous memory to honour any supported mapping. should allocate sufficient contiguous memory to honour any supported mapping.
=================================
ADJUSTING PAGE TRIMMING BEHAVIOUR
=================================
NOMMU mmap automatically rounds up to the nearest power-of-2 number of pages
when performing an allocation. This can have adverse effects on memory
fragmentation, and as such, is left configurable. The default behaviour is to
aggressively trim allocations and discard any excess pages back in to the page
allocator. In order to retain finer-grained control over fragmentation, this
behaviour can either be disabled completely, or bumped up to a higher page
watermark where trimming begins.
Page trimming behaviour is configurable via the sysctl `vm.nr_trim_pages'.
...@@ -38,6 +38,7 @@ Currently, these files are in /proc/sys/vm: ...@@ -38,6 +38,7 @@ Currently, these files are in /proc/sys/vm:
- numa_zonelist_order - numa_zonelist_order
- nr_hugepages - nr_hugepages
- nr_overcommit_hugepages - nr_overcommit_hugepages
- nr_trim_pages (only if CONFIG_MMU=n)
============================================================== ==============================================================
...@@ -348,3 +349,20 @@ Change the maximum size of the hugepage pool. The maximum is ...@@ -348,3 +349,20 @@ Change the maximum size of the hugepage pool. The maximum is
nr_hugepages + nr_overcommit_hugepages. nr_hugepages + nr_overcommit_hugepages.
See Documentation/vm/hugetlbpage.txt See Documentation/vm/hugetlbpage.txt
==============================================================
nr_trim_pages
This is available only on NOMMU kernels.
This value adjusts the excess page trimming behaviour of power-of-2 aligned
NOMMU mmap allocations.
A value of 0 disables trimming of allocations entirely, while a value of 1
trims excess pages aggressively. Any value >= 1 acts as the watermark where
trimming of allocations is initiated.
The default value is 1.
See Documentation/nommu-mmap.txt for more information.
...@@ -24,7 +24,6 @@ typedef struct { ...@@ -24,7 +24,6 @@ typedef struct {
* modified for 2.6 by Hyok S. Choi <hyok.choi@samsung.com> * modified for 2.6 by Hyok S. Choi <hyok.choi@samsung.com>
*/ */
typedef struct { typedef struct {
struct vm_list_struct *vmlist;
unsigned long end_brk; unsigned long end_brk;
} mm_context_t; } mm_context_t;
......
...@@ -71,7 +71,7 @@ static DEFINE_SPINLOCK(consistent_lock); ...@@ -71,7 +71,7 @@ static DEFINE_SPINLOCK(consistent_lock);
* the amount of RAM found at boot time.) I would imagine that get_vm_area() * the amount of RAM found at boot time.) I would imagine that get_vm_area()
* would have to initialise this each time prior to calling vm_region_alloc(). * would have to initialise this each time prior to calling vm_region_alloc().
*/ */
struct vm_region { struct arm_vm_region {
struct list_head vm_list; struct list_head vm_list;
unsigned long vm_start; unsigned long vm_start;
unsigned long vm_end; unsigned long vm_end;
...@@ -79,20 +79,20 @@ struct vm_region { ...@@ -79,20 +79,20 @@ struct vm_region {
int vm_active; int vm_active;
}; };
static struct vm_region consistent_head = { static struct arm_vm_region consistent_head = {
.vm_list = LIST_HEAD_INIT(consistent_head.vm_list), .vm_list = LIST_HEAD_INIT(consistent_head.vm_list),
.vm_start = CONSISTENT_BASE, .vm_start = CONSISTENT_BASE,
.vm_end = CONSISTENT_END, .vm_end = CONSISTENT_END,
}; };
static struct vm_region * static struct arm_vm_region *
vm_region_alloc(struct vm_region *head, size_t size, gfp_t gfp) arm_vm_region_alloc(struct arm_vm_region *head, size_t size, gfp_t gfp)
{ {
unsigned long addr = head->vm_start, end = head->vm_end - size; unsigned long addr = head->vm_start, end = head->vm_end - size;
unsigned long flags; unsigned long flags;
struct vm_region *c, *new; struct arm_vm_region *c, *new;
new = kmalloc(sizeof(struct vm_region), gfp); new = kmalloc(sizeof(struct arm_vm_region), gfp);
if (!new) if (!new)
goto out; goto out;
...@@ -127,9 +127,9 @@ vm_region_alloc(struct vm_region *head, size_t size, gfp_t gfp) ...@@ -127,9 +127,9 @@ vm_region_alloc(struct vm_region *head, size_t size, gfp_t gfp)
return NULL; return NULL;
} }
static struct vm_region *vm_region_find(struct vm_region *head, unsigned long addr) static struct arm_vm_region *arm_vm_region_find(struct arm_vm_region *head, unsigned long addr)
{ {
struct vm_region *c; struct arm_vm_region *c;
list_for_each_entry(c, &head->vm_list, vm_list) { list_for_each_entry(c, &head->vm_list, vm_list) {
if (c->vm_active && c->vm_start == addr) if (c->vm_active && c->vm_start == addr)
...@@ -149,7 +149,7 @@ __dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp, ...@@ -149,7 +149,7 @@ __dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp,
pgprot_t prot) pgprot_t prot)
{ {
struct page *page; struct page *page;
struct vm_region *c; struct arm_vm_region *c;
unsigned long order; unsigned long order;
u64 mask = ISA_DMA_THRESHOLD, limit; u64 mask = ISA_DMA_THRESHOLD, limit;
...@@ -214,7 +214,7 @@ __dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp, ...@@ -214,7 +214,7 @@ __dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp,
/* /*
* Allocate a virtual address in the consistent mapping region. * Allocate a virtual address in the consistent mapping region.
*/ */
c = vm_region_alloc(&consistent_head, size, c = arm_vm_region_alloc(&consistent_head, size,
gfp & ~(__GFP_DMA | __GFP_HIGHMEM)); gfp & ~(__GFP_DMA | __GFP_HIGHMEM));
if (c) { if (c) {
pte_t *pte; pte_t *pte;
...@@ -311,13 +311,13 @@ static int dma_mmap(struct device *dev, struct vm_area_struct *vma, ...@@ -311,13 +311,13 @@ static int dma_mmap(struct device *dev, struct vm_area_struct *vma,
void *cpu_addr, dma_addr_t dma_addr, size_t size) void *cpu_addr, dma_addr_t dma_addr, size_t size)
{ {
unsigned long flags, user_size, kern_size; unsigned long flags, user_size, kern_size;
struct vm_region *c; struct arm_vm_region *c;
int ret = -ENXIO; int ret = -ENXIO;
user_size = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; user_size = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT;
spin_lock_irqsave(&consistent_lock, flags); spin_lock_irqsave(&consistent_lock, flags);
c = vm_region_find(&consistent_head, (unsigned long)cpu_addr); c = arm_vm_region_find(&consistent_head, (unsigned long)cpu_addr);
spin_unlock_irqrestore(&consistent_lock, flags); spin_unlock_irqrestore(&consistent_lock, flags);
if (c) { if (c) {
...@@ -359,7 +359,7 @@ EXPORT_SYMBOL(dma_mmap_writecombine); ...@@ -359,7 +359,7 @@ EXPORT_SYMBOL(dma_mmap_writecombine);
*/ */
void dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, dma_addr_t handle) void dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, dma_addr_t handle)
{ {
struct vm_region *c; struct arm_vm_region *c;
unsigned long flags, addr; unsigned long flags, addr;
pte_t *ptep; pte_t *ptep;
int idx; int idx;
...@@ -378,7 +378,7 @@ void dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, dma_addr ...@@ -378,7 +378,7 @@ void dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, dma_addr
size = PAGE_ALIGN(size); size = PAGE_ALIGN(size);
spin_lock_irqsave(&consistent_lock, flags); spin_lock_irqsave(&consistent_lock, flags);
c = vm_region_find(&consistent_head, (unsigned long)cpu_addr); c = arm_vm_region_find(&consistent_head, (unsigned long)cpu_addr);
if (!c) if (!c)
goto no_area; goto no_area;
......
...@@ -10,7 +10,6 @@ struct sram_list_struct { ...@@ -10,7 +10,6 @@ struct sram_list_struct {
}; };
typedef struct { typedef struct {
struct vm_list_struct *vmlist;
unsigned long end_brk; unsigned long end_brk;
unsigned long stack_start; unsigned long stack_start;
......
...@@ -160,15 +160,15 @@ put_reg(struct task_struct *task, int regno, unsigned long data) ...@@ -160,15 +160,15 @@ put_reg(struct task_struct *task, int regno, unsigned long data)
static inline int is_user_addr_valid(struct task_struct *child, static inline int is_user_addr_valid(struct task_struct *child,
unsigned long start, unsigned long len) unsigned long start, unsigned long len)
{ {
struct vm_list_struct *vml; struct vm_area_struct *vma;
struct sram_list_struct *sraml; struct sram_list_struct *sraml;
/* overflow */ /* overflow */
if (start + len < start) if (start + len < start)
return -EIO; return -EIO;
for (vml = child->mm->context.vmlist; vml; vml = vml->next) vma = find_vma(child->mm, start);
if (start >= vml->vma->vm_start && start + len < vml->vma->vm_end) if (vma && start >= vma->vm_start && start + len <= vma->vm_end)
return 0; return 0;
for (sraml = child->mm->context.sram_list; sraml; sraml = sraml->next) for (sraml = child->mm->context.sram_list; sraml; sraml = sraml->next)
......
...@@ -32,6 +32,7 @@ ...@@ -32,6 +32,7 @@
#include <linux/module.h> #include <linux/module.h>
#include <linux/kallsyms.h> #include <linux/kallsyms.h>
#include <linux/fs.h> #include <linux/fs.h>
#include <linux/rbtree.h>
#include <asm/traps.h> #include <asm/traps.h>
#include <asm/cacheflush.h> #include <asm/cacheflush.h>
#include <asm/cplb.h> #include <asm/cplb.h>
...@@ -83,6 +84,7 @@ static void decode_address(char *buf, unsigned long address) ...@@ -83,6 +84,7 @@ static void decode_address(char *buf, unsigned long address)
struct mm_struct *mm; struct mm_struct *mm;
unsigned long flags, offset; unsigned long flags, offset;
unsigned char in_atomic = (bfin_read_IPEND() & 0x10) || in_atomic(); unsigned char in_atomic = (bfin_read_IPEND() & 0x10) || in_atomic();
struct rb_node *n;
#ifdef CONFIG_KALLSYMS #ifdef CONFIG_KALLSYMS
unsigned long symsize; unsigned long symsize;
...@@ -128,9 +130,10 @@ static void decode_address(char *buf, unsigned long address) ...@@ -128,9 +130,10 @@ static void decode_address(char *buf, unsigned long address)
if (!mm) if (!mm)
continue; continue;
vml = mm->context.vmlist; for (n = rb_first(&mm->mm_rb); n; n = rb_next(n)) {
while (vml) { struct vm_area_struct *vma;
struct vm_area_struct *vma = vml->vma;
vma = rb_entry(n, struct vm_area_struct, vm_rb);
if (address >= vma->vm_start && address < vma->vm_end) { if (address >= vma->vm_start && address < vma->vm_end) {
char _tmpbuf[256]; char _tmpbuf[256];
...@@ -176,8 +179,6 @@ static void decode_address(char *buf, unsigned long address) ...@@ -176,8 +179,6 @@ static void decode_address(char *buf, unsigned long address)
goto done; goto done;
} }
vml = vml->next;
} }
if (!in_atomic) if (!in_atomic)
mmput(mm); mmput(mm);
......
...@@ -69,7 +69,8 @@ static inline int put_reg(struct task_struct *task, int regno, ...@@ -69,7 +69,8 @@ static inline int put_reg(struct task_struct *task, int regno,
} }
/* /*
* check that an address falls within the bounds of the target process's memory mappings * check that an address falls within the bounds of the target process's memory
* mappings
*/ */
static inline int is_user_addr_valid(struct task_struct *child, static inline int is_user_addr_valid(struct task_struct *child,
unsigned long start, unsigned long len) unsigned long start, unsigned long len)
...@@ -79,11 +80,11 @@ static inline int is_user_addr_valid(struct task_struct *child, ...@@ -79,11 +80,11 @@ static inline int is_user_addr_valid(struct task_struct *child,
return -EIO; return -EIO;
return 0; return 0;
#else #else
struct vm_list_struct *vml; struct vm_area_struct *vma;
for (vml = child->mm->context.vmlist; vml; vml = vml->next) vma = find_vma(child->mm, start);
if (start >= vml->vma->vm_start && start + len <= vml->vma->vm_end) if (vma && start >= vma->vm_start && start + len <= vma->vm_end)
return 0; return 0;
return -EIO; return -EIO;
#endif #endif
......
...@@ -4,7 +4,6 @@ ...@@ -4,7 +4,6 @@
/* Copyright (C) 2002, David McCullough <davidm@snapgear.com> */ /* Copyright (C) 2002, David McCullough <davidm@snapgear.com> */
typedef struct { typedef struct {
struct vm_list_struct *vmlist;
unsigned long end_brk; unsigned long end_brk;
} mm_context_t; } mm_context_t;
......
...@@ -4,7 +4,6 @@ ...@@ -4,7 +4,6 @@
/* Copyright (C) 2002, David McCullough <davidm@snapgear.com> */ /* Copyright (C) 2002, David McCullough <davidm@snapgear.com> */
typedef struct { typedef struct {
struct vm_list_struct *vmlist;
unsigned long end_brk; unsigned long end_brk;
} mm_context_t; } mm_context_t;
......
...@@ -9,7 +9,6 @@ typedef struct { ...@@ -9,7 +9,6 @@ typedef struct {
mm_context_id_t id; mm_context_id_t id;
void *vdso; void *vdso;
#else #else
struct vm_list_struct *vmlist;
unsigned long end_brk; unsigned long end_brk;
#endif #endif
#ifdef CONFIG_BINFMT_ELF_FDPIC #ifdef CONFIG_BINFMT_ELF_FDPIC
......
...@@ -168,9 +168,6 @@ static int load_elf_fdpic_binary(struct linux_binprm *bprm, ...@@ -168,9 +168,6 @@ static int load_elf_fdpic_binary(struct linux_binprm *bprm,
struct elf_fdpic_params exec_params, interp_params; struct elf_fdpic_params exec_params, interp_params;
struct elf_phdr *phdr; struct elf_phdr *phdr;
unsigned long stack_size, entryaddr; unsigned long stack_size, entryaddr;
#ifndef CONFIG_MMU
unsigned long fullsize;
#endif
#ifdef ELF_FDPIC_PLAT_INIT #ifdef ELF_FDPIC_PLAT_INIT
unsigned long dynaddr; unsigned long dynaddr;
#endif #endif
...@@ -390,11 +387,6 @@ static int load_elf_fdpic_binary(struct linux_binprm *bprm, ...@@ -390,11 +387,6 @@ static int load_elf_fdpic_binary(struct linux_binprm *bprm,
goto error_kill; goto error_kill;
} }
/* expand the stack mapping to use up the entire allocation granule */
fullsize = kobjsize((char *) current->mm->start_brk);
if (!IS_ERR_VALUE(do_mremap(current->mm->start_brk, stack_size,
fullsize, 0, 0)))
stack_size = fullsize;
up_write(&current->mm->mmap_sem); up_write(&current->mm->mmap_sem);
current->mm->brk = current->mm->start_brk; current->mm->brk = current->mm->start_brk;
...@@ -1567,11 +1559,9 @@ end_coredump: ...@@ -1567,11 +1559,9 @@ end_coredump:
static int elf_fdpic_dump_segments(struct file *file, size_t *size, static int elf_fdpic_dump_segments(struct file *file, size_t *size,
unsigned long *limit, unsigned long mm_flags) unsigned long *limit, unsigned long mm_flags)
{ {
struct vm_list_struct *vml; struct vm_area_struct *vma;
for (vml = current->mm->context.vmlist; vml; vml = vml->next) {
struct vm_area_struct *vma = vml->vma;
for (vma = current->mm->mmap; vma; vma = vma->vm_next) {
if (!maydump(vma, mm_flags)) if (!maydump(vma, mm_flags))
continue; continue;
...@@ -1617,9 +1607,6 @@ static int elf_fdpic_core_dump(long signr, struct pt_regs *regs, ...@@ -1617,9 +1607,6 @@ static int elf_fdpic_core_dump(long signr, struct pt_regs *regs,
elf_fpxregset_t *xfpu = NULL; elf_fpxregset_t *xfpu = NULL;
#endif #endif
int thread_status_size = 0; int thread_status_size = 0;
#ifndef CONFIG_MMU
struct vm_list_struct *vml;
#endif
elf_addr_t *auxv; elf_addr_t *auxv;
unsigned long mm_flags; unsigned long mm_flags;
...@@ -1685,13 +1672,7 @@ static int elf_fdpic_core_dump(long signr, struct pt_regs *regs, ...@@ -1685,13 +1672,7 @@ static int elf_fdpic_core_dump(long signr, struct pt_regs *regs,
fill_prstatus(prstatus, current, signr); fill_prstatus(prstatus, current, signr);
elf_core_copy_regs(&prstatus->pr_reg, regs); elf_core_copy_regs(&prstatus->pr_reg, regs);
#ifdef CONFIG_MMU
segs = current->mm->map_count; segs = current->mm->map_count;
#else
segs = 0;
for (vml = current->mm->context.vmlist; vml; vml = vml->next)
segs++;
#endif
#ifdef ELF_CORE_EXTRA_PHDRS #ifdef ELF_CORE_EXTRA_PHDRS
segs += ELF_CORE_EXTRA_PHDRS; segs += ELF_CORE_EXTRA_PHDRS;
#endif #endif
...@@ -1766,20 +1747,10 @@ static int elf_fdpic_core_dump(long signr, struct pt_regs *regs, ...@@ -1766,20 +1747,10 @@ static int elf_fdpic_core_dump(long signr, struct pt_regs *regs,
mm_flags = current->mm->flags; mm_flags = current->mm->flags;
/* write program headers for segments dump */ /* write program headers for segments dump */
for ( for (vma = current->mm->mmap; vma; vma = vma->vm_next) {
#ifdef CONFIG_MMU
vma = current->mm->mmap; vma; vma = vma->vm_next
#else
vml = current->mm->context.vmlist; vml; vml = vml->next
#endif
) {
struct elf_phdr phdr; struct elf_phdr phdr;
size_t sz; size_t sz;
#ifndef CONFIG_MMU
vma = vml->vma;
#endif
sz = vma->vm_end - vma->vm_start; sz = vma->vm_end - vma->vm_start;
phdr.p_type = PT_LOAD; phdr.p_type = PT_LOAD;
......
...@@ -417,8 +417,8 @@ static int load_flat_file(struct linux_binprm * bprm, ...@@ -417,8 +417,8 @@ static int load_flat_file(struct linux_binprm * bprm,
unsigned long textpos = 0, datapos = 0, result; unsigned long textpos = 0, datapos = 0, result;
unsigned long realdatastart = 0; unsigned long realdatastart = 0;
unsigned long text_len, data_len, bss_len, stack_len, flags; unsigned long text_len, data_len, bss_len, stack_len, flags;
unsigned long len, reallen, memp = 0; unsigned long len, memp = 0;
unsigned long extra, rlim; unsigned long memp_size, extra, rlim;
unsigned long *reloc = 0, *rp; unsigned long *reloc = 0, *rp;
struct inode *inode; struct inode *inode;
int i, rev, relocs = 0; int i, rev, relocs = 0;
...@@ -543,17 +543,10 @@ static int load_flat_file(struct linux_binprm * bprm, ...@@ -543,17 +543,10 @@ static int load_flat_file(struct linux_binprm * bprm,
} }
len = data_len + extra + MAX_SHARED_LIBS * sizeof(unsigned long); len = data_len + extra + MAX_SHARED_LIBS * sizeof(unsigned long);
len = PAGE_ALIGN(len);
down_write(&current->mm->mmap_sem); down_write(&current->mm->mmap_sem);
realdatastart = do_mmap(0, 0, len, realdatastart = do_mmap(0, 0, len,
PROT_READ|PROT_WRITE|PROT_EXEC, MAP_PRIVATE, 0); PROT_READ|PROT_WRITE|PROT_EXEC, MAP_PRIVATE, 0);
/* Remap to use all availabe slack region space */
if (realdatastart && (realdatastart < (unsigned long)-4096)) {
reallen = kobjsize((void *)realdatastart);
if (reallen > len) {
realdatastart = do_mremap(realdatastart, len,
reallen, MREMAP_FIXED, realdatastart);
}
}
up_write(&current->mm->mmap_sem); up_write(&current->mm->mmap_sem);
if (realdatastart == 0 || realdatastart >= (unsigned long)-4096) { if (realdatastart == 0 || realdatastart >= (unsigned long)-4096) {
...@@ -591,21 +584,14 @@ static int load_flat_file(struct linux_binprm * bprm, ...@@ -591,21 +584,14 @@ static int load_flat_file(struct linux_binprm * bprm,
reloc = (unsigned long *) (datapos+(ntohl(hdr->reloc_start)-text_len)); reloc = (unsigned long *) (datapos+(ntohl(hdr->reloc_start)-text_len));
memp = realdatastart; memp = realdatastart;
memp_size = len;
} else { } else {
len = text_len + data_len + extra + MAX_SHARED_LIBS * sizeof(unsigned long); len = text_len + data_len + extra + MAX_SHARED_LIBS * sizeof(unsigned long);
len = PAGE_ALIGN(len);
down_write(&current->mm->mmap_sem); down_write(&current->mm->mmap_sem);
textpos = do_mmap(0, 0, len, textpos = do_mmap(0, 0, len,
PROT_READ | PROT_EXEC | PROT_WRITE, MAP_PRIVATE, 0); PROT_READ | PROT_EXEC | PROT_WRITE, MAP_PRIVATE, 0);
/* Remap to use all availabe slack region space */
if (textpos && (textpos < (unsigned long) -4096)) {
reallen = kobjsize((void *)textpos);
if (reallen > len) {
textpos = do_mremap(textpos, len, reallen,
MREMAP_FIXED, textpos);
}
}
up_write(&current->mm->mmap_sem); up_write(&current->mm->mmap_sem);
if (!textpos || textpos >= (unsigned long) -4096) { if (!textpos || textpos >= (unsigned long) -4096) {
...@@ -622,7 +608,7 @@ static int load_flat_file(struct linux_binprm * bprm, ...@@ -622,7 +608,7 @@ static int load_flat_file(struct linux_binprm * bprm,
reloc = (unsigned long *) (textpos + ntohl(hdr->reloc_start) + reloc = (unsigned long *) (textpos + ntohl(hdr->reloc_start) +
MAX_SHARED_LIBS * sizeof(unsigned long)); MAX_SHARED_LIBS * sizeof(unsigned long));
memp = textpos; memp = textpos;
memp_size = len;
#ifdef CONFIG_BINFMT_ZFLAT #ifdef CONFIG_BINFMT_ZFLAT
/* /*
* load it all in and treat it like a RAM load from now on * load it all in and treat it like a RAM load from now on
...@@ -680,10 +666,12 @@ static int load_flat_file(struct linux_binprm * bprm, ...@@ -680,10 +666,12 @@ static int load_flat_file(struct linux_binprm * bprm,
* set up the brk stuff, uses any slack left in data/bss/stack * set up the brk stuff, uses any slack left in data/bss/stack
* allocation. We put the brk after the bss (between the bss * allocation. We put the brk after the bss (between the bss
* and stack) like other platforms. * and stack) like other platforms.
* Userspace code relies on the stack pointer starting out at
* an address right at the end of a page.
*/ */
current->mm->start_brk = datapos + data_len + bss_len; current->mm->start_brk = datapos + data_len + bss_len;
current->mm->brk = (current->mm->start_brk + 3) & ~3; current->mm->brk = (current->mm->start_brk + 3) & ~3;
current->mm->context.end_brk = memp + kobjsize((void *) memp) - stack_len; current->mm->context.end_brk = memp + memp_size - stack_len;
} }
if (flags & FLAT_FLAG_KTRACE) if (flags & FLAT_FLAG_KTRACE)
...@@ -790,8 +778,8 @@ static int load_flat_file(struct linux_binprm * bprm, ...@@ -790,8 +778,8 @@ static int load_flat_file(struct linux_binprm * bprm,
/* zero the BSS, BRK and stack areas */ /* zero the BSS, BRK and stack areas */
memset((void*)(datapos + data_len), 0, bss_len + memset((void*)(datapos + data_len), 0, bss_len +
(memp + kobjsize((void *) memp) - stack_len - /* end brk */ (memp + memp_size - stack_len - /* end brk */
libinfo->lib_list[id].start_brk) + /* start brk */ libinfo->lib_list[id].start_brk) + /* start brk */
stack_len); stack_len);
return 0; return 0;
......
...@@ -41,8 +41,6 @@ do { \ ...@@ -41,8 +41,6 @@ do { \
(vmi)->used = 0; \ (vmi)->used = 0; \
(vmi)->largest_chunk = 0; \ (vmi)->largest_chunk = 0; \
} while(0) } while(0)
extern int nommu_vma_show(struct seq_file *, struct vm_area_struct *);
#endif #endif
extern int proc_tid_stat(struct seq_file *m, struct pid_namespace *ns, extern int proc_tid_stat(struct seq_file *m, struct pid_namespace *ns,
......
...@@ -73,6 +73,9 @@ static int meminfo_proc_show(struct seq_file *m, void *v) ...@@ -73,6 +73,9 @@ static int meminfo_proc_show(struct seq_file *m, void *v)
"HighFree: %8lu kB\n" "HighFree: %8lu kB\n"
"LowTotal: %8lu kB\n" "LowTotal: %8lu kB\n"
"LowFree: %8lu kB\n" "LowFree: %8lu kB\n"
#endif
#ifndef CONFIG_MMU
"MmapCopy: %8lu kB\n"
#endif #endif
"SwapTotal: %8lu kB\n" "SwapTotal: %8lu kB\n"
"SwapFree: %8lu kB\n" "SwapFree: %8lu kB\n"
...@@ -115,6 +118,9 @@ static int meminfo_proc_show(struct seq_file *m, void *v) ...@@ -115,6 +118,9 @@ static int meminfo_proc_show(struct seq_file *m, void *v)
K(i.freehigh), K(i.freehigh),
K(i.totalram-i.totalhigh), K(i.totalram-i.totalhigh),
K(i.freeram-i.freehigh), K(i.freeram-i.freehigh),
#endif
#ifndef CONFIG_MMU
K((unsigned long) atomic_read(&mmap_pages_allocated)),
#endif #endif
K(i.totalswap), K(i.totalswap),
K(i.freeswap), K(i.freeswap),
......
...@@ -33,33 +33,33 @@ ...@@ -33,33 +33,33 @@
#include "internal.h" #include "internal.h"
/* /*
* display a single VMA to a sequenced file * display a single region to a sequenced file
*/ */
int nommu_vma_show(struct seq_file *m, struct vm_area_struct *vma) static int nommu_region_show(struct seq_file *m, struct vm_region *region)
{ {
unsigned long ino = 0; unsigned long ino = 0;
struct file *file; struct file *file;
dev_t dev = 0; dev_t dev = 0;
int flags, len; int flags, len;
flags = vma->vm_flags; flags = region->vm_flags;
file = vma->vm_file; file = region->vm_file;
if (file) { if (file) {
struct inode *inode = vma->vm_file->f_path.dentry->d_inode; struct inode *inode = region->vm_file->f_path.dentry->d_inode;
dev = inode->i_sb->s_dev; dev = inode->i_sb->s_dev;
ino = inode->i_ino; ino = inode->i_ino;
} }
seq_printf(m, seq_printf(m,
"%08lx-%08lx %c%c%c%c %08llx %02x:%02x %lu %n", "%08lx-%08lx %c%c%c%c %08llx %02x:%02x %lu %n",
vma->vm_start, region->vm_start,
vma->vm_end, region->vm_end,
flags & VM_READ ? 'r' : '-', flags & VM_READ ? 'r' : '-',
flags & VM_WRITE ? 'w' : '-', flags & VM_WRITE ? 'w' : '-',
flags & VM_EXEC ? 'x' : '-', flags & VM_EXEC ? 'x' : '-',
flags & VM_MAYSHARE ? flags & VM_SHARED ? 'S' : 's' : 'p', flags & VM_MAYSHARE ? flags & VM_SHARED ? 'S' : 's' : 'p',
((loff_t)vma->vm_pgoff) << PAGE_SHIFT, ((loff_t)region->vm_pgoff) << PAGE_SHIFT,
MAJOR(dev), MINOR(dev), ino, &len); MAJOR(dev), MINOR(dev), ino, &len);
if (file) { if (file) {
...@@ -75,61 +75,54 @@ int nommu_vma_show(struct seq_file *m, struct vm_area_struct *vma) ...@@ -75,61 +75,54 @@ int nommu_vma_show(struct seq_file *m, struct vm_area_struct *vma)
} }
/* /*
* display a list of all the VMAs the kernel knows about * display a list of all the REGIONs the kernel knows about
* - nommu kernals have a single flat list * - nommu kernals have a single flat list
*/ */
static int nommu_vma_list_show(struct seq_file *m, void *v) static int nommu_region_list_show(struct seq_file *m, void *_p)
{ {
struct vm_area_struct *vma; struct rb_node *p = _p;
vma = rb_entry((struct rb_node *) v, struct vm_area_struct, vm_rb); return nommu_region_show(m, rb_entry(p, struct vm_region, vm_rb));
return nommu_vma_show(m, vma);
} }
static void *nommu_vma_list_start(struct seq_file *m, loff_t *_pos) static void *nommu_region_list_start(struct seq_file *m, loff_t *_pos)
{ {
struct rb_node *_rb; struct rb_node *p;
loff_t pos = *_pos; loff_t pos = *_pos;
void *next = NULL;
down_read(&nommu_vma_sem); down_read(&nommu_region_sem);
for (_rb = rb_first(&nommu_vma_tree); _rb; _rb = rb_next(_rb)) { for (p = rb_first(&nommu_region_tree); p; p = rb_next(p))
if (pos == 0) { if (pos-- == 0)
next = _rb; return p;
break; return NULL;
}
pos--;
}
return next;
} }
static void nommu_vma_list_stop(struct seq_file *m, void *v) static void nommu_region_list_stop(struct seq_file *m, void *v)
{ {
up_read(&nommu_vma_sem); up_read(&nommu_region_sem);
} }
static void *nommu_vma_list_next(struct seq_file *m, void *v, loff_t *pos) static void *nommu_region_list_next(struct seq_file *m, void *v, loff_t *pos)
{ {
(*pos)++; (*pos)++;
return rb_next((struct rb_node *) v); return rb_next((struct rb_node *) v);
} }
static const struct seq_operations proc_nommu_vma_list_seqop = { static struct seq_operations proc_nommu_region_list_seqop = {
.start = nommu_vma_list_start, .start = nommu_region_list_start,
.next = nommu_vma_list_next, .next = nommu_region_list_next,
.stop = nommu_vma_list_stop, .stop = nommu_region_list_stop,
.show = nommu_vma_list_show .show = nommu_region_list_show
}; };
static int proc_nommu_vma_list_open(struct inode *inode, struct file *file) static int proc_nommu_region_list_open(struct inode *inode, struct file *file)
{ {
return seq_open(file, &proc_nommu_vma_list_seqop); return seq_open(file, &proc_nommu_region_list_seqop);
} }
static const struct file_operations proc_nommu_vma_list_operations = { static const struct file_operations proc_nommu_region_list_operations = {
.open = proc_nommu_vma_list_open, .open = proc_nommu_region_list_open,
.read = seq_read, .read = seq_read,
.llseek = seq_lseek, .llseek = seq_lseek,
.release = seq_release, .release = seq_release,
...@@ -137,7 +130,7 @@ static const struct file_operations proc_nommu_vma_list_operations = { ...@@ -137,7 +130,7 @@ static const struct file_operations proc_nommu_vma_list_operations = {
static int __init proc_nommu_init(void) static int __init proc_nommu_init(void)
{ {
proc_create("maps", S_IRUGO, NULL, &proc_nommu_vma_list_operations); proc_create("maps", S_IRUGO, NULL, &proc_nommu_region_list_operations);
return 0; return 0;
} }
......
...@@ -15,25 +15,32 @@ ...@@ -15,25 +15,32 @@
*/ */
void task_mem(struct seq_file *m, struct mm_struct *mm) void task_mem(struct seq_file *m, struct mm_struct *mm)
{ {
struct vm_list_struct *vml; struct vm_area_struct *vma;
unsigned long bytes = 0, sbytes = 0, slack = 0; struct vm_region *region;
struct rb_node *p;
unsigned long bytes = 0, sbytes = 0, slack = 0, size;
down_read(&mm->mmap_sem); down_read(&mm->mmap_sem);
for (vml = mm->context.vmlist; vml; vml = vml->next) { for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) {
if (!vml->vma) vma = rb_entry(p, struct vm_area_struct, vm_rb);
continue;
bytes += kobjsize(vma);
region = vma->vm_region;
if (region) {
size = kobjsize(region);
size += region->vm_end - region->vm_start;
} else {
size = vma->vm_end - vma->vm_start;
}
bytes += kobjsize(vml);
if (atomic_read(&mm->mm_count) > 1 || if (atomic_read(&mm->mm_count) > 1 ||
atomic_read(&vml->vma->vm_usage) > 1 vma->vm_flags & VM_MAYSHARE) {
) { sbytes += size;
sbytes += kobjsize((void *) vml->vma->vm_start);
sbytes += kobjsize(vml->vma);
} else { } else {
bytes += kobjsize((void *) vml->vma->vm_start); bytes += size;
bytes += kobjsize(vml->vma); if (region)
slack += kobjsize((void *) vml->vma->vm_start) - slack = region->vm_end - vma->vm_end;
(vml->vma->vm_end - vml->vma->vm_start);
} }
} }
...@@ -70,13 +77,14 @@ void task_mem(struct seq_file *m, struct mm_struct *mm) ...@@ -70,13 +77,14 @@ void task_mem(struct seq_file *m, struct mm_struct *mm)
unsigned long task_vsize(struct mm_struct *mm) unsigned long task_vsize(struct mm_struct *mm)
{ {
struct vm_list_struct *tbp; struct vm_area_struct *vma;
struct rb_node *p;
unsigned long vsize = 0; unsigned long vsize = 0;
down_read(&mm->mmap_sem); down_read(&mm->mmap_sem);
for (tbp = mm->context.vmlist; tbp; tbp = tbp->next) { for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) {
if (tbp->vma) vma = rb_entry(p, struct vm_area_struct, vm_rb);
vsize += kobjsize((void *) tbp->vma->vm_start); vsize += vma->vm_end - vma->vm_start;
} }
up_read(&mm->mmap_sem); up_read(&mm->mmap_sem);
return vsize; return vsize;
...@@ -85,15 +93,19 @@ unsigned long task_vsize(struct mm_struct *mm) ...@@ -85,15 +93,19 @@ unsigned long task_vsize(struct mm_struct *mm)
int task_statm(struct mm_struct *mm, int *shared, int *text, int task_statm(struct mm_struct *mm, int *shared, int *text,
int *data, int *resident) int *data, int *resident)
{ {
struct vm_list_struct *tbp; struct vm_area_struct *vma;
struct vm_region *region;
struct rb_node *p;
int size = kobjsize(mm); int size = kobjsize(mm);
down_read(&mm->mmap_sem); down_read(&mm->mmap_sem);
for (tbp = mm->context.vmlist; tbp; tbp = tbp->next) { for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) {
size += kobjsize(tbp); vma = rb_entry(p, struct vm_area_struct, vm_rb);
if (tbp->vma) { size += kobjsize(vma);
size += kobjsize(tbp->vma); region = vma->vm_region;
size += kobjsize((void *) tbp->vma->vm_start); if (region) {
size += kobjsize(region);
size += region->vm_end - region->vm_start;
} }
} }
...@@ -104,21 +116,63 @@ int task_statm(struct mm_struct *mm, int *shared, int *text, ...@@ -104,21 +116,63 @@ int task_statm(struct mm_struct *mm, int *shared, int *text,
return size; return size;
} }
/*
* display a single VMA to a sequenced file
*/
static int nommu_vma_show(struct seq_file *m, struct vm_area_struct *vma)
{
unsigned long ino = 0;
struct file *file;
dev_t dev = 0;
int flags, len;
flags = vma->vm_flags;
file = vma->vm_file;
if (file) {
struct inode *inode = vma->vm_file->f_path.dentry->d_inode;
dev = inode->i_sb->s_dev;
ino = inode->i_ino;
}
seq_printf(m,
"%08lx-%08lx %c%c%c%c %08lx %02x:%02x %lu %n",
vma->vm_start,
vma->vm_end,
flags & VM_READ ? 'r' : '-',
flags & VM_WRITE ? 'w' : '-',
flags & VM_EXEC ? 'x' : '-',
flags & VM_MAYSHARE ? flags & VM_SHARED ? 'S' : 's' : 'p',
vma->vm_pgoff << PAGE_SHIFT,
MAJOR(dev), MINOR(dev), ino, &len);
if (file) {
len = 25 + sizeof(void *) * 6 - len;
if (len < 1)
len = 1;
seq_printf(m, "%*c", len, ' ');
seq_path(m, &file->f_path, "");
}
seq_putc(m, '\n');
return 0;
}
/* /*
* display mapping lines for a particular process's /proc/pid/maps * display mapping lines for a particular process's /proc/pid/maps
*/ */
static int show_map(struct seq_file *m, void *_vml) static int show_map(struct seq_file *m, void *_p)
{ {
struct vm_list_struct *vml = _vml; struct rb_node *p = _p;
return nommu_vma_show(m, vml->vma); return nommu_vma_show(m, rb_entry(p, struct vm_area_struct, vm_rb));
} }
static void *m_start(struct seq_file *m, loff_t *pos) static void *m_start(struct seq_file *m, loff_t *pos)
{ {
struct proc_maps_private *priv = m->private; struct proc_maps_private *priv = m->private;
struct vm_list_struct *vml;
struct mm_struct *mm; struct mm_struct *mm;
struct rb_node *p;
loff_t n = *pos; loff_t n = *pos;
/* pin the task and mm whilst we play with them */ /* pin the task and mm whilst we play with them */
...@@ -134,9 +188,9 @@ static void *m_start(struct seq_file *m, loff_t *pos) ...@@ -134,9 +188,9 @@ static void *m_start(struct seq_file *m, loff_t *pos)
} }
/* start from the Nth VMA */ /* start from the Nth VMA */
for (vml = mm->context.vmlist; vml; vml = vml->next) for (p = rb_first(&mm->mm_rb); p; p = rb_next(p))
if (n-- == 0) if (n-- == 0)
return vml; return p;
return NULL; return NULL;
} }
...@@ -152,12 +206,12 @@ static void m_stop(struct seq_file *m, void *_vml) ...@@ -152,12 +206,12 @@ static void m_stop(struct seq_file *m, void *_vml)
} }
} }
static void *m_next(struct seq_file *m, void *_vml, loff_t *pos) static void *m_next(struct seq_file *m, void *_p, loff_t *pos)
{ {
struct vm_list_struct *vml = _vml; struct rb_node *p = _p;
(*pos)++; (*pos)++;
return vml ? vml->next : NULL; return p ? rb_next(p) : NULL;
} }
static const struct seq_operations proc_pid_maps_ops = { static const struct seq_operations proc_pid_maps_ops = {
......
...@@ -262,11 +262,11 @@ unsigned long ramfs_nommu_get_unmapped_area(struct file *file, ...@@ -262,11 +262,11 @@ unsigned long ramfs_nommu_get_unmapped_area(struct file *file,
ret = -ENOMEM; ret = -ENOMEM;
pages = kzalloc(lpages * sizeof(struct page *), GFP_KERNEL); pages = kzalloc(lpages * sizeof(struct page *), GFP_KERNEL);
if (!pages) if (!pages)
goto out; goto out_free;
nr = find_get_pages(inode->i_mapping, pgoff, lpages, pages); nr = find_get_pages(inode->i_mapping, pgoff, lpages, pages);
if (nr != lpages) if (nr != lpages)
goto out; /* leave if some pages were missing */ goto out_free_pages; /* leave if some pages were missing */
/* check the pages for physical adjacency */ /* check the pages for physical adjacency */
ptr = pages; ptr = pages;
...@@ -274,19 +274,18 @@ unsigned long ramfs_nommu_get_unmapped_area(struct file *file, ...@@ -274,19 +274,18 @@ unsigned long ramfs_nommu_get_unmapped_area(struct file *file,
page++; page++;
for (loop = lpages; loop > 1; loop--) for (loop = lpages; loop > 1; loop--)
if (*ptr++ != page++) if (*ptr++ != page++)
goto out; goto out_free_pages;
/* okay - all conditions fulfilled */ /* okay - all conditions fulfilled */
ret = (unsigned long) page_address(pages[0]); ret = (unsigned long) page_address(pages[0]);
out: out_free_pages:
if (pages) { ptr = pages;
ptr = pages; for (loop = nr; loop > 0; loop--)
for (loop = lpages; loop > 0; loop--) put_page(*ptr++);
put_page(*ptr++); out_free:
kfree(pages); kfree(pages);
} out:
return ret; return ret;
} }
......
...@@ -22,7 +22,6 @@ typedef struct { ...@@ -22,7 +22,6 @@ typedef struct {
unsigned long dtlb_ptd_mapping; /* [DAMR5] PTD mapping for dtlb cached PGE */ unsigned long dtlb_ptd_mapping; /* [DAMR5] PTD mapping for dtlb cached PGE */
#else #else
struct vm_list_struct *vmlist;
unsigned long end_brk; unsigned long end_brk;
#endif #endif
......
...@@ -4,7 +4,6 @@ ...@@ -4,7 +4,6 @@
#if !defined(CONFIG_MMU) #if !defined(CONFIG_MMU)
typedef struct { typedef struct {
struct vm_list_struct *vmlist;
unsigned long end_brk; unsigned long end_brk;
} mm_context_t; } mm_context_t;
......
...@@ -56,19 +56,9 @@ extern unsigned long mmap_min_addr; ...@@ -56,19 +56,9 @@ extern unsigned long mmap_min_addr;
extern struct kmem_cache *vm_area_cachep; extern struct kmem_cache *vm_area_cachep;
/*
* This struct defines the per-mm list of VMAs for uClinux. If CONFIG_MMU is
* disabled, then there's a single shared list of VMAs maintained by the
* system, and mm's subscribe to these individually
*/
struct vm_list_struct {
struct vm_list_struct *next;
struct vm_area_struct *vma;
};
#ifndef CONFIG_MMU #ifndef CONFIG_MMU
extern struct rb_root nommu_vma_tree; extern struct rb_root nommu_region_tree;
extern struct rw_semaphore nommu_vma_sem; extern struct rw_semaphore nommu_region_sem;
extern unsigned int kobjsize(const void *objp); extern unsigned int kobjsize(const void *objp);
#endif #endif
...@@ -1061,6 +1051,7 @@ extern void memmap_init_zone(unsigned long, int, unsigned long, ...@@ -1061,6 +1051,7 @@ extern void memmap_init_zone(unsigned long, int, unsigned long,
unsigned long, enum memmap_context); unsigned long, enum memmap_context);
extern void setup_per_zone_pages_min(void); extern void setup_per_zone_pages_min(void);
extern void mem_init(void); extern void mem_init(void);
extern void __init mmap_init(void);
extern void show_mem(void); extern void show_mem(void);
extern void si_meminfo(struct sysinfo * val); extern void si_meminfo(struct sysinfo * val);
extern void si_meminfo_node(struct sysinfo *val, int nid); extern void si_meminfo_node(struct sysinfo *val, int nid);
...@@ -1072,6 +1063,9 @@ extern void setup_per_cpu_pageset(void); ...@@ -1072,6 +1063,9 @@ extern void setup_per_cpu_pageset(void);
static inline void setup_per_cpu_pageset(void) {} static inline void setup_per_cpu_pageset(void) {}
#endif #endif
/* nommu.c */
extern atomic_t mmap_pages_allocated;
/* prio_tree.c */ /* prio_tree.c */
void vma_prio_tree_add(struct vm_area_struct *, struct vm_area_struct *old); void vma_prio_tree_add(struct vm_area_struct *, struct vm_area_struct *old);
void vma_prio_tree_insert(struct vm_area_struct *, struct prio_tree_root *); void vma_prio_tree_insert(struct vm_area_struct *, struct prio_tree_root *);
......
...@@ -96,6 +96,23 @@ struct page { ...@@ -96,6 +96,23 @@ struct page {
#endif /* WANT_PAGE_VIRTUAL */ #endif /* WANT_PAGE_VIRTUAL */
}; };
/*
* A region containing a mapping of a non-memory backed file under NOMMU
* conditions. These are held in a global tree and are pinned by the VMAs that
* map parts of them.
*/
struct vm_region {
struct rb_node vm_rb; /* link in global region tree */
unsigned long vm_flags; /* VMA vm_flags */
unsigned long vm_start; /* start address of region */
unsigned long vm_end; /* region initialised to here */
unsigned long vm_top; /* region allocated to here */
unsigned long vm_pgoff; /* the offset in vm_file corresponding to vm_start */
struct file *vm_file; /* the backing file or NULL */
atomic_t vm_usage; /* region usage count */
};
/* /*
* This struct defines a memory VMM memory area. There is one of these * This struct defines a memory VMM memory area. There is one of these
* per VM-area/task. A VM area is any part of the process virtual memory * per VM-area/task. A VM area is any part of the process virtual memory
...@@ -152,7 +169,7 @@ struct vm_area_struct { ...@@ -152,7 +169,7 @@ struct vm_area_struct {
unsigned long vm_truncate_count;/* truncate_count or restart_addr */ unsigned long vm_truncate_count;/* truncate_count or restart_addr */
#ifndef CONFIG_MMU #ifndef CONFIG_MMU
atomic_t vm_usage; /* refcount (VMAs shared if !MMU) */ struct vm_region *vm_region; /* NOMMU mapping region */
#endif #endif
#ifdef CONFIG_NUMA #ifdef CONFIG_NUMA
struct mempolicy *vm_policy; /* NUMA policy for the VMA */ struct mempolicy *vm_policy; /* NUMA policy for the VMA */
......
...@@ -317,6 +317,7 @@ static int __init do_name(void) ...@@ -317,6 +317,7 @@ static int __init do_name(void)
if (wfd >= 0) { if (wfd >= 0) {
sys_fchown(wfd, uid, gid); sys_fchown(wfd, uid, gid);
sys_fchmod(wfd, mode); sys_fchmod(wfd, mode);
sys_ftruncate(wfd, body_len);
vcollected = kstrdup(collected, GFP_KERNEL); vcollected = kstrdup(collected, GFP_KERNEL);
state = CopyFile; state = CopyFile;
} }
......
...@@ -990,6 +990,7 @@ asmlinkage long sys_shmdt(char __user *shmaddr) ...@@ -990,6 +990,7 @@ asmlinkage long sys_shmdt(char __user *shmaddr)
*/ */
vma = find_vma(mm, addr); vma = find_vma(mm, addr);
#ifdef CONFIG_MMU
while (vma) { while (vma) {
next = vma->vm_next; next = vma->vm_next;
...@@ -1034,6 +1035,17 @@ asmlinkage long sys_shmdt(char __user *shmaddr) ...@@ -1034,6 +1035,17 @@ asmlinkage long sys_shmdt(char __user *shmaddr)
vma = next; vma = next;
} }
#else /* CONFIG_MMU */
/* under NOMMU conditions, the exact address to be destroyed must be
* given */
retval = -EINVAL;
if (vma->vm_start == addr && vma->vm_ops == &shm_vm_ops) {
do_munmap(mm, vma->vm_start, vma->vm_end - vma->vm_start);
retval = 0;
}
#endif
up_write(&mm->mmap_sem); up_write(&mm->mmap_sem);
return retval; return retval;
} }
......
...@@ -1481,12 +1481,10 @@ void __init proc_caches_init(void) ...@@ -1481,12 +1481,10 @@ void __init proc_caches_init(void)
fs_cachep = kmem_cache_create("fs_cache", fs_cachep = kmem_cache_create("fs_cache",
sizeof(struct fs_struct), 0, sizeof(struct fs_struct), 0,
SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL); SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL);
vm_area_cachep = kmem_cache_create("vm_area_struct",
sizeof(struct vm_area_struct), 0,
SLAB_PANIC, NULL);
mm_cachep = kmem_cache_create("mm_struct", mm_cachep = kmem_cache_create("mm_struct",
sizeof(struct mm_struct), ARCH_MIN_MMSTRUCT_ALIGN, sizeof(struct mm_struct), ARCH_MIN_MMSTRUCT_ALIGN,
SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL); SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL);
mmap_init();
} }
/* /*
......
...@@ -82,6 +82,9 @@ extern int percpu_pagelist_fraction; ...@@ -82,6 +82,9 @@ extern int percpu_pagelist_fraction;
extern int compat_log; extern int compat_log;
extern int latencytop_enabled; extern int latencytop_enabled;
extern int sysctl_nr_open_min, sysctl_nr_open_max; extern int sysctl_nr_open_min, sysctl_nr_open_max;
#ifndef CONFIG_MMU
extern int sysctl_nr_trim_pages;
#endif
#ifdef CONFIG_RCU_TORTURE_TEST #ifdef CONFIG_RCU_TORTURE_TEST
extern int rcutorture_runnable; extern int rcutorture_runnable;
#endif /* #ifdef CONFIG_RCU_TORTURE_TEST */ #endif /* #ifdef CONFIG_RCU_TORTURE_TEST */
...@@ -1102,6 +1105,17 @@ static struct ctl_table vm_table[] = { ...@@ -1102,6 +1105,17 @@ static struct ctl_table vm_table[] = {
.mode = 0644, .mode = 0644,
.proc_handler = &proc_dointvec .proc_handler = &proc_dointvec
}, },
#else
{
.ctl_name = CTL_UNNUMBERED,
.procname = "nr_trim_pages",
.data = &sysctl_nr_trim_pages,
.maxlen = sizeof(sysctl_nr_trim_pages),
.mode = 0644,
.proc_handler = &proc_dointvec_minmax,
.strategy = &sysctl_intvec,
.extra1 = &zero,
},
#endif #endif
{ {
.ctl_name = VM_LAPTOP_MODE, .ctl_name = VM_LAPTOP_MODE,
......
...@@ -512,6 +512,13 @@ config DEBUG_VIRTUAL ...@@ -512,6 +512,13 @@ config DEBUG_VIRTUAL
If unsure, say N. If unsure, say N.
config DEBUG_NOMMU_REGIONS
bool "Debug the global anon/private NOMMU mapping region tree"
depends on DEBUG_KERNEL && !MMU
help
This option causes the global tree of anonymous and private mapping
regions to be regularly checked for invalid topology.
config DEBUG_WRITECOUNT config DEBUG_WRITECOUNT
bool "Debug filesystem writers count" bool "Debug filesystem writers count"
depends on DEBUG_KERNEL depends on DEBUG_KERNEL
......
...@@ -2472,3 +2472,13 @@ void mm_drop_all_locks(struct mm_struct *mm) ...@@ -2472,3 +2472,13 @@ void mm_drop_all_locks(struct mm_struct *mm)
mutex_unlock(&mm_all_locks_mutex); mutex_unlock(&mm_all_locks_mutex);
} }
/*
* initialise the VMA slab
*/
void __init mmap_init(void)
{
vm_area_cachep = kmem_cache_create("vm_area_struct",
sizeof(struct vm_area_struct), 0,
SLAB_PANIC, NULL);
}
This diff is collapsed.
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment