Commit ac97991e authored by Andreas Herrmann's avatar Andreas Herrmann Committed by Ingo Molnar

x86: pat.c choose more crisp variable names

 - parse/ml => entry (within list_for_each and friends)
 - new_entry => new
 - ret_type => new_type (to avoid confusion with req_type)

(... to make it more readable...)
Signed-off-by: default avatarAndreas Herrmann <andreas.herrmann3@amd.com>
Cc: Venkatesh Pallipadi <venkatesh.pallipadi@intel.com>
Cc: Suresh B Siddha <suresh.b.siddha@intel.com>
Signed-off-by: default avatarIngo Molnar <mingo@elte.hu>
parent bcc643dc
...@@ -188,37 +188,34 @@ static unsigned long pat_x_mtrr_type(u64 start, u64 end, unsigned long req_type) ...@@ -188,37 +188,34 @@ static unsigned long pat_x_mtrr_type(u64 start, u64 end, unsigned long req_type)
* req_type will have a special case value '-1', when requester want to inherit * req_type will have a special case value '-1', when requester want to inherit
* the memory type from mtrr (if WB), existing PAT, defaulting to UC_MINUS. * the memory type from mtrr (if WB), existing PAT, defaulting to UC_MINUS.
* *
* If ret_type is NULL, function will return an error if it cannot reserve the * If new_type is NULL, function will return an error if it cannot reserve the
* region with req_type. If ret_type is non-null, function will return * region with req_type. If new_type is non-NULL, function will return
* available type in ret_type in case of no error. In case of any error * available type in new_type in case of no error. In case of any error
* it will return a negative return value. * it will return a negative return value.
*/ */
int reserve_memtype(u64 start, u64 end, unsigned long req_type, int reserve_memtype(u64 start, u64 end, unsigned long req_type,
unsigned long *ret_type) unsigned long *new_type)
{ {
struct memtype *new_entry = NULL; struct memtype *new, *entry;
struct memtype *parse;
unsigned long actual_type; unsigned long actual_type;
int err = 0; int err = 0;
/* Only track when pat_enabled */ /* Only track when pat_enabled */
if (!pat_enabled) { if (!pat_enabled) {
/* This is identical to page table setting without PAT */ /* This is identical to page table setting without PAT */
if (ret_type) { if (new_type) {
if (req_type == -1) { if (req_type == -1)
*ret_type = _PAGE_CACHE_WB; *new_type = _PAGE_CACHE_WB;
} else { else
*ret_type = req_type & _PAGE_CACHE_MASK; *new_type = req_type & _PAGE_CACHE_MASK;
}
} }
return 0; return 0;
} }
/* Low ISA region is always mapped WB in page table. No need to track */ /* Low ISA region is always mapped WB in page table. No need to track */
if (is_ISA_range(start, end - 1)) { if (is_ISA_range(start, end - 1)) {
if (ret_type) if (new_type)
*ret_type = _PAGE_CACHE_WB; *new_type = _PAGE_CACHE_WB;
return 0; return 0;
} }
...@@ -243,65 +240,65 @@ int reserve_memtype(u64 start, u64 end, unsigned long req_type, ...@@ -243,65 +240,65 @@ int reserve_memtype(u64 start, u64 end, unsigned long req_type,
actual_type = pat_x_mtrr_type(start, end, req_type); actual_type = pat_x_mtrr_type(start, end, req_type);
} }
new_entry = kmalloc(sizeof(struct memtype), GFP_KERNEL); new = kmalloc(sizeof(struct memtype), GFP_KERNEL);
if (!new_entry) if (!new)
return -ENOMEM; return -ENOMEM;
new_entry->start = start; new->start = start;
new_entry->end = end; new->end = end;
new_entry->type = actual_type; new->type = actual_type;
if (ret_type) if (new_type)
*ret_type = actual_type; *new_type = actual_type;
spin_lock(&memtype_lock); spin_lock(&memtype_lock);
/* Search for existing mapping that overlaps the current range */ /* Search for existing mapping that overlaps the current range */
list_for_each_entry(parse, &memtype_list, nd) { list_for_each_entry(entry, &memtype_list, nd) {
struct memtype *saved_ptr; struct memtype *saved_ptr;
if (parse->start >= end) { if (entry->start >= end) {
dprintk("New Entry\n"); dprintk("New Entry\n");
list_add(&new_entry->nd, parse->nd.prev); list_add(&new->nd, entry->nd.prev);
new_entry = NULL; new = NULL;
break; break;
} }
if (start <= parse->start && end >= parse->start) { if (start <= entry->start && end >= entry->start) {
if (actual_type != parse->type && ret_type) { if (actual_type != entry->type && new_type) {
actual_type = parse->type; actual_type = entry->type;
*ret_type = actual_type; *new_type = actual_type;
new_entry->type = actual_type; new->type = actual_type;
} }
if (actual_type != parse->type) { if (actual_type != entry->type) {
printk( printk(
KERN_INFO "%s:%d conflicting memory types %Lx-%Lx %s<->%s\n", KERN_INFO "%s:%d conflicting memory types %Lx-%Lx %s<->%s\n",
current->comm, current->pid, current->comm, current->pid,
start, end, start, end,
cattr_name(actual_type), cattr_name(actual_type),
cattr_name(parse->type)); cattr_name(entry->type));
err = -EBUSY; err = -EBUSY;
break; break;
} }
saved_ptr = parse; saved_ptr = entry;
/* /*
* Check to see whether the request overlaps more * Check to see whether the request overlaps more
* than one entry in the list * than one entry in the list
*/ */
list_for_each_entry_continue(parse, &memtype_list, nd) { list_for_each_entry_continue(entry, &memtype_list, nd) {
if (end <= parse->start) { if (end <= entry->start) {
break; break;
} }
if (actual_type != parse->type) { if (actual_type != entry->type) {
printk( printk(
KERN_INFO "%s:%d conflicting memory types %Lx-%Lx %s<->%s\n", KERN_INFO "%s:%d conflicting memory types %Lx-%Lx %s<->%s\n",
current->comm, current->pid, current->comm, current->pid,
start, end, start, end,
cattr_name(actual_type), cattr_name(actual_type),
cattr_name(parse->type)); cattr_name(entry->type));
err = -EBUSY; err = -EBUSY;
break; break;
} }
...@@ -314,46 +311,46 @@ int reserve_memtype(u64 start, u64 end, unsigned long req_type, ...@@ -314,46 +311,46 @@ int reserve_memtype(u64 start, u64 end, unsigned long req_type,
dprintk("Overlap at 0x%Lx-0x%Lx\n", dprintk("Overlap at 0x%Lx-0x%Lx\n",
saved_ptr->start, saved_ptr->end); saved_ptr->start, saved_ptr->end);
/* No conflict. Go ahead and add this new entry */ /* No conflict. Go ahead and add this new entry */
list_add(&new_entry->nd, saved_ptr->nd.prev); list_add(&new->nd, saved_ptr->nd.prev);
new_entry = NULL; new = NULL;
break; break;
} }
if (start < parse->end) { if (start < entry->end) {
if (actual_type != parse->type && ret_type) { if (actual_type != entry->type && new_type) {
actual_type = parse->type; actual_type = entry->type;
*ret_type = actual_type; *new_type = actual_type;
new_entry->type = actual_type; new->type = actual_type;
} }
if (actual_type != parse->type) { if (actual_type != entry->type) {
printk( printk(
KERN_INFO "%s:%d conflicting memory types %Lx-%Lx %s<->%s\n", KERN_INFO "%s:%d conflicting memory types %Lx-%Lx %s<->%s\n",
current->comm, current->pid, current->comm, current->pid,
start, end, start, end,
cattr_name(actual_type), cattr_name(actual_type),
cattr_name(parse->type)); cattr_name(entry->type));
err = -EBUSY; err = -EBUSY;
break; break;
} }
saved_ptr = parse; saved_ptr = entry;
/* /*
* Check to see whether the request overlaps more * Check to see whether the request overlaps more
* than one entry in the list * than one entry in the list
*/ */
list_for_each_entry_continue(parse, &memtype_list, nd) { list_for_each_entry_continue(entry, &memtype_list, nd) {
if (end <= parse->start) { if (end <= entry->start) {
break; break;
} }
if (actual_type != parse->type) { if (actual_type != entry->type) {
printk( printk(
KERN_INFO "%s:%d conflicting memory types %Lx-%Lx %s<->%s\n", KERN_INFO "%s:%d conflicting memory types %Lx-%Lx %s<->%s\n",
current->comm, current->pid, current->comm, current->pid,
start, end, start, end,
cattr_name(actual_type), cattr_name(actual_type),
cattr_name(parse->type)); cattr_name(entry->type));
err = -EBUSY; err = -EBUSY;
break; break;
} }
...@@ -366,8 +363,8 @@ int reserve_memtype(u64 start, u64 end, unsigned long req_type, ...@@ -366,8 +363,8 @@ int reserve_memtype(u64 start, u64 end, unsigned long req_type,
dprintk("Overlap at 0x%Lx-0x%Lx\n", dprintk("Overlap at 0x%Lx-0x%Lx\n",
saved_ptr->start, saved_ptr->end); saved_ptr->start, saved_ptr->end);
/* No conflict. Go ahead and add this new entry */ /* No conflict. Go ahead and add this new entry */
list_add(&new_entry->nd, &saved_ptr->nd); list_add(&new->nd, &saved_ptr->nd);
new_entry = NULL; new = NULL;
break; break;
} }
} }
...@@ -375,24 +372,24 @@ int reserve_memtype(u64 start, u64 end, unsigned long req_type, ...@@ -375,24 +372,24 @@ int reserve_memtype(u64 start, u64 end, unsigned long req_type,
if (err) { if (err) {
printk(KERN_INFO printk(KERN_INFO
"reserve_memtype failed 0x%Lx-0x%Lx, track %s, req %s\n", "reserve_memtype failed 0x%Lx-0x%Lx, track %s, req %s\n",
start, end, cattr_name(new_entry->type), start, end, cattr_name(new->type),
cattr_name(req_type)); cattr_name(req_type));
kfree(new_entry); kfree(new);
spin_unlock(&memtype_lock); spin_unlock(&memtype_lock);
return err; return err;
} }
if (new_entry) { if (new) {
/* No conflict. Not yet added to the list. Add to the tail */ /* No conflict. Not yet added to the list. Add to the tail */
list_add_tail(&new_entry->nd, &memtype_list); list_add_tail(&new->nd, &memtype_list);
dprintk("New Entry\n"); dprintk("New Entry\n");
} }
if (ret_type) { if (new_type) {
dprintk( dprintk(
"reserve_memtype added 0x%Lx-0x%Lx, track %s, req %s, ret %s\n", "reserve_memtype added 0x%Lx-0x%Lx, track %s, req %s, ret %s\n",
start, end, cattr_name(actual_type), start, end, cattr_name(actual_type),
cattr_name(req_type), cattr_name(*ret_type)); cattr_name(req_type), cattr_name(*new_type));
} else { } else {
dprintk( dprintk(
"reserve_memtype added 0x%Lx-0x%Lx, track %s, req %s\n", "reserve_memtype added 0x%Lx-0x%Lx, track %s, req %s\n",
...@@ -406,7 +403,7 @@ int reserve_memtype(u64 start, u64 end, unsigned long req_type, ...@@ -406,7 +403,7 @@ int reserve_memtype(u64 start, u64 end, unsigned long req_type,
int free_memtype(u64 start, u64 end) int free_memtype(u64 start, u64 end)
{ {
struct memtype *ml; struct memtype *entry;
int err = -EINVAL; int err = -EINVAL;
/* Only track when pat_enabled */ /* Only track when pat_enabled */
...@@ -419,10 +416,10 @@ int free_memtype(u64 start, u64 end) ...@@ -419,10 +416,10 @@ int free_memtype(u64 start, u64 end)
return 0; return 0;
spin_lock(&memtype_lock); spin_lock(&memtype_lock);
list_for_each_entry(ml, &memtype_list, nd) { list_for_each_entry(entry, &memtype_list, nd) {
if (ml->start == start && ml->end == end) { if (entry->start == start && entry->end == end) {
list_del(&ml->nd); list_del(&entry->nd);
kfree(ml); kfree(entry);
err = 0; err = 0;
break; break;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment