Commit f56975da authored by Hugh Dickins's avatar Hugh Dickins Committed by james toy

Free up a pointer in struct rmap_item, by making the mm_slot's rmap_list a

singly-linked list: we always traverse that list sequentially, and we
don't even lose any prefetches (but should consider adding a few later). 
Name it rmap_list throughout.

Do we need to free up that pointer?  Not immediately, and in the end, we
could continue to avoid it with a union; but having done the conversion,
let's keep it this way, since there's no downside, and maybe we'll want
more in future (struct rmap_item is a cache-friendly 32 bytes on 32-bit
and 64 bytes on 64-bit, so we shall want to avoid expanding it).
Signed-off-by: default avatarHugh Dickins <hugh.dickins@tiscali.co.uk>
Cc: Izik Eidus <ieidus@redhat.com>
Cc: Andrea Arcangeli <aarcange@redhat.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
parent 4c0e3707
...@@ -79,13 +79,13 @@ ...@@ -79,13 +79,13 @@
* struct mm_slot - ksm information per mm that is being scanned * struct mm_slot - ksm information per mm that is being scanned
* @link: link to the mm_slots hash list * @link: link to the mm_slots hash list
* @mm_list: link into the mm_slots list, rooted in ksm_mm_head * @mm_list: link into the mm_slots list, rooted in ksm_mm_head
* @rmap_list: head for this mm_slot's list of rmap_items * @rmap_list: head for this mm_slot's singly-linked list of rmap_items
* @mm: the mm that this information is valid for * @mm: the mm that this information is valid for
*/ */
struct mm_slot { struct mm_slot {
struct hlist_node link; struct hlist_node link;
struct list_head mm_list; struct list_head mm_list;
struct list_head rmap_list; struct rmap_item *rmap_list;
struct mm_struct *mm; struct mm_struct *mm;
}; };
...@@ -93,7 +93,7 @@ struct mm_slot { ...@@ -93,7 +93,7 @@ struct mm_slot {
* struct ksm_scan - cursor for scanning * struct ksm_scan - cursor for scanning
* @mm_slot: the current mm_slot we are scanning * @mm_slot: the current mm_slot we are scanning
* @address: the next address inside that to be scanned * @address: the next address inside that to be scanned
* @rmap_item: the current rmap that we are scanning inside the rmap_list * @rmap_list: link to the next rmap to be scanned in the rmap_list
* @seqnr: count of completed full scans (needed when removing unstable node) * @seqnr: count of completed full scans (needed when removing unstable node)
* *
* There is only the one ksm_scan instance of this cursor structure. * There is only the one ksm_scan instance of this cursor structure.
...@@ -101,13 +101,14 @@ struct mm_slot { ...@@ -101,13 +101,14 @@ struct mm_slot {
struct ksm_scan { struct ksm_scan {
struct mm_slot *mm_slot; struct mm_slot *mm_slot;
unsigned long address; unsigned long address;
struct rmap_item *rmap_item; struct rmap_item **rmap_list;
unsigned long seqnr; unsigned long seqnr;
}; };
/** /**
* struct rmap_item - reverse mapping item for virtual addresses * struct rmap_item - reverse mapping item for virtual addresses
* @link: link into mm_slot's rmap_list (rmap_list is per mm) * @rmap_list: next rmap_item in mm_slot's singly-linked rmap_list
* @filler: unused space we're making available in this patch
* @mm: the memory structure this rmap_item is pointing into * @mm: the memory structure this rmap_item is pointing into
* @address: the virtual address this rmap_item tracks (+ flags in low bits) * @address: the virtual address this rmap_item tracks (+ flags in low bits)
* @oldchecksum: previous checksum of the page at that virtual address * @oldchecksum: previous checksum of the page at that virtual address
...@@ -116,7 +117,8 @@ struct ksm_scan { ...@@ -116,7 +117,8 @@ struct ksm_scan {
* @prev: previous rmap_item hanging off the same node of the stable tree * @prev: previous rmap_item hanging off the same node of the stable tree
*/ */
struct rmap_item { struct rmap_item {
struct list_head link; struct rmap_item *rmap_list;
unsigned long filler;
struct mm_struct *mm; struct mm_struct *mm;
unsigned long address; /* + low bits used for flags below */ unsigned long address; /* + low bits used for flags below */
union { union {
...@@ -275,7 +277,6 @@ static void insert_to_mm_slots_hash(struct mm_struct *mm, ...@@ -275,7 +277,6 @@ static void insert_to_mm_slots_hash(struct mm_struct *mm,
bucket = &mm_slots_hash[((unsigned long)mm / sizeof(struct mm_struct)) bucket = &mm_slots_hash[((unsigned long)mm / sizeof(struct mm_struct))
% MM_SLOTS_HASH_HEADS]; % MM_SLOTS_HASH_HEADS];
mm_slot->mm = mm; mm_slot->mm = mm;
INIT_LIST_HEAD(&mm_slot->rmap_list);
hlist_add_head(&mm_slot->link, bucket); hlist_add_head(&mm_slot->link, bucket);
} }
...@@ -479,15 +480,12 @@ static void remove_rmap_item_from_tree(struct rmap_item *rmap_item) ...@@ -479,15 +480,12 @@ static void remove_rmap_item_from_tree(struct rmap_item *rmap_item)
} }
static void remove_trailing_rmap_items(struct mm_slot *mm_slot, static void remove_trailing_rmap_items(struct mm_slot *mm_slot,
struct list_head *cur) struct rmap_item **rmap_list)
{ {
struct rmap_item *rmap_item; while (*rmap_list) {
struct rmap_item *rmap_item = *rmap_list;
while (cur != &mm_slot->rmap_list) { *rmap_list = rmap_item->rmap_list;
rmap_item = list_entry(cur, struct rmap_item, link);
cur = cur->next;
remove_rmap_item_from_tree(rmap_item); remove_rmap_item_from_tree(rmap_item);
list_del(&rmap_item->link);
free_rmap_item(rmap_item); free_rmap_item(rmap_item);
} }
} }
...@@ -553,7 +551,7 @@ static int unmerge_and_remove_all_rmap_items(void) ...@@ -553,7 +551,7 @@ static int unmerge_and_remove_all_rmap_items(void)
goto error; goto error;
} }
remove_trailing_rmap_items(mm_slot, mm_slot->rmap_list.next); remove_trailing_rmap_items(mm_slot, &mm_slot->rmap_list);
spin_lock(&ksm_mmlist_lock); spin_lock(&ksm_mmlist_lock);
ksm_scan.mm_slot = list_entry(mm_slot->mm_list.next, ksm_scan.mm_slot = list_entry(mm_slot->mm_list.next,
...@@ -1141,20 +1139,19 @@ static void cmp_and_merge_page(struct page *page, struct rmap_item *rmap_item) ...@@ -1141,20 +1139,19 @@ static void cmp_and_merge_page(struct page *page, struct rmap_item *rmap_item)
} }
static struct rmap_item *get_next_rmap_item(struct mm_slot *mm_slot, static struct rmap_item *get_next_rmap_item(struct mm_slot *mm_slot,
struct list_head *cur, struct rmap_item **rmap_list,
unsigned long addr) unsigned long addr)
{ {
struct rmap_item *rmap_item; struct rmap_item *rmap_item;
while (cur != &mm_slot->rmap_list) { while (*rmap_list) {
rmap_item = list_entry(cur, struct rmap_item, link); rmap_item = *rmap_list;
if ((rmap_item->address & PAGE_MASK) == addr) if ((rmap_item->address & PAGE_MASK) == addr)
return rmap_item; return rmap_item;
if (rmap_item->address > addr) if (rmap_item->address > addr)
break; break;
cur = cur->next; *rmap_list = rmap_item->rmap_list;
remove_rmap_item_from_tree(rmap_item); remove_rmap_item_from_tree(rmap_item);
list_del(&rmap_item->link);
free_rmap_item(rmap_item); free_rmap_item(rmap_item);
} }
...@@ -1163,7 +1160,8 @@ static struct rmap_item *get_next_rmap_item(struct mm_slot *mm_slot, ...@@ -1163,7 +1160,8 @@ static struct rmap_item *get_next_rmap_item(struct mm_slot *mm_slot,
/* It has already been zeroed */ /* It has already been zeroed */
rmap_item->mm = mm_slot->mm; rmap_item->mm = mm_slot->mm;
rmap_item->address = addr; rmap_item->address = addr;
list_add_tail(&rmap_item->link, cur); rmap_item->rmap_list = *rmap_list;
*rmap_list = rmap_item;
} }
return rmap_item; return rmap_item;
} }
...@@ -1188,8 +1186,7 @@ static struct rmap_item *scan_get_next_rmap_item(struct page **page) ...@@ -1188,8 +1186,7 @@ static struct rmap_item *scan_get_next_rmap_item(struct page **page)
spin_unlock(&ksm_mmlist_lock); spin_unlock(&ksm_mmlist_lock);
next_mm: next_mm:
ksm_scan.address = 0; ksm_scan.address = 0;
ksm_scan.rmap_item = list_entry(&slot->rmap_list, ksm_scan.rmap_list = &slot->rmap_list;
struct rmap_item, link);
} }
mm = slot->mm; mm = slot->mm;
...@@ -1215,10 +1212,10 @@ next_mm: ...@@ -1215,10 +1212,10 @@ next_mm:
flush_anon_page(vma, *page, ksm_scan.address); flush_anon_page(vma, *page, ksm_scan.address);
flush_dcache_page(*page); flush_dcache_page(*page);
rmap_item = get_next_rmap_item(slot, rmap_item = get_next_rmap_item(slot,
ksm_scan.rmap_item->link.next, ksm_scan.rmap_list, ksm_scan.address);
ksm_scan.address);
if (rmap_item) { if (rmap_item) {
ksm_scan.rmap_item = rmap_item; ksm_scan.rmap_list =
&rmap_item->rmap_list;
ksm_scan.address += PAGE_SIZE; ksm_scan.address += PAGE_SIZE;
} else } else
put_page(*page); put_page(*page);
...@@ -1234,14 +1231,13 @@ next_mm: ...@@ -1234,14 +1231,13 @@ next_mm:
if (ksm_test_exit(mm)) { if (ksm_test_exit(mm)) {
ksm_scan.address = 0; ksm_scan.address = 0;
ksm_scan.rmap_item = list_entry(&slot->rmap_list, ksm_scan.rmap_list = &slot->rmap_list;
struct rmap_item, link);
} }
/* /*
* Nuke all the rmap_items that are above this current rmap: * Nuke all the rmap_items that are above this current rmap:
* because there were no VM_MERGEABLE vmas with such addresses. * because there were no VM_MERGEABLE vmas with such addresses.
*/ */
remove_trailing_rmap_items(slot, ksm_scan.rmap_item->link.next); remove_trailing_rmap_items(slot, ksm_scan.rmap_list);
spin_lock(&ksm_mmlist_lock); spin_lock(&ksm_mmlist_lock);
ksm_scan.mm_slot = list_entry(slot->mm_list.next, ksm_scan.mm_slot = list_entry(slot->mm_list.next,
...@@ -1423,7 +1419,7 @@ void __ksm_exit(struct mm_struct *mm) ...@@ -1423,7 +1419,7 @@ void __ksm_exit(struct mm_struct *mm)
spin_lock(&ksm_mmlist_lock); spin_lock(&ksm_mmlist_lock);
mm_slot = get_mm_slot(mm); mm_slot = get_mm_slot(mm);
if (mm_slot && ksm_scan.mm_slot != mm_slot) { if (mm_slot && ksm_scan.mm_slot != mm_slot) {
if (list_empty(&mm_slot->rmap_list)) { if (!mm_slot->rmap_list) {
hlist_del(&mm_slot->link); hlist_del(&mm_slot->link);
list_del(&mm_slot->mm_list); list_del(&mm_slot->mm_list);
easy_to_free = 1; easy_to_free = 1;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment