Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
L
linux-davinci
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Redmine
Redmine
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Operations
Operations
Metrics
Environments
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
linux
linux-davinci
Commits
9b4006dc
Commit
9b4006dc
authored
Mar 18, 2006
by
David S. Miller
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
[SPARC64]: Use SLAB caches for TSB tables.
Signed-off-by:
David S. Miller
<
davem@davemloft.net
>
parent
b52439c2
Changes
3
Show whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
69 additions
and
25 deletions
+69
-25
arch/sparc64/Kconfig
arch/sparc64/Kconfig
+3
-0
arch/sparc64/mm/init.c
arch/sparc64/mm/init.c
+4
-1
arch/sparc64/mm/tsb.c
arch/sparc64/mm/tsb.c
+62
-24
No files found.
arch/sparc64/Kconfig
View file @
9b4006dc
...
@@ -192,6 +192,9 @@ config ARCH_SPARSEMEM_ENABLE
...
@@ -192,6 +192,9 @@ config ARCH_SPARSEMEM_ENABLE
config ARCH_SPARSEMEM_DEFAULT
config ARCH_SPARSEMEM_DEFAULT
def_bool y
def_bool y
config LARGE_ALLOCS
def_bool y
source "mm/Kconfig"
source "mm/Kconfig"
config GENERIC_ISA_DMA
config GENERIC_ISA_DMA
...
...
arch/sparc64/mm/init.c
View file @
9b4006dc
...
@@ -165,6 +165,8 @@ static void zero_ctor(void *addr, kmem_cache_t *cache, unsigned long flags)
...
@@ -165,6 +165,8 @@ static void zero_ctor(void *addr, kmem_cache_t *cache, unsigned long flags)
clear_page
(
addr
);
clear_page
(
addr
);
}
}
extern
void
tsb_cache_init
(
void
);
void
pgtable_cache_init
(
void
)
void
pgtable_cache_init
(
void
)
{
{
pgtable_cache
=
kmem_cache_create
(
"pgtable_cache"
,
pgtable_cache
=
kmem_cache_create
(
"pgtable_cache"
,
...
@@ -174,9 +176,10 @@ void pgtable_cache_init(void)
...
@@ -174,9 +176,10 @@ void pgtable_cache_init(void)
zero_ctor
,
zero_ctor
,
NULL
);
NULL
);
if
(
!
pgtable_cache
)
{
if
(
!
pgtable_cache
)
{
prom_printf
(
"
pgtable_cache_init(): Could not create!
\n
"
);
prom_printf
(
"
Could not create pgtable_cache
\n
"
);
prom_halt
();
prom_halt
();
}
}
tsb_cache_init
();
}
}
#ifdef CONFIG_DEBUG_DCFLUSH
#ifdef CONFIG_DEBUG_DCFLUSH
...
...
arch/sparc64/mm/tsb.c
View file @
9b4006dc
...
@@ -11,6 +11,7 @@
...
@@ -11,6 +11,7 @@
#include <asm/mmu_context.h>
#include <asm/mmu_context.h>
#include <asm/pgtable.h>
#include <asm/pgtable.h>
#include <asm/tsb.h>
#include <asm/tsb.h>
#include <asm/oplib.h>
extern
struct
tsb
swapper_tsb
[
KERNEL_TSB_NENTRIES
];
extern
struct
tsb
swapper_tsb
[
KERNEL_TSB_NENTRIES
];
...
@@ -207,6 +208,39 @@ static void setup_tsb_params(struct mm_struct *mm, unsigned long tsb_bytes)
...
@@ -207,6 +208,39 @@ static void setup_tsb_params(struct mm_struct *mm, unsigned long tsb_bytes)
}
}
}
}
static
kmem_cache_t
*
tsb_caches
[
8
]
__read_mostly
;
static
const
char
*
tsb_cache_names
[
8
]
=
{
"tsb_8KB"
,
"tsb_16KB"
,
"tsb_32KB"
,
"tsb_64KB"
,
"tsb_128KB"
,
"tsb_256KB"
,
"tsb_512KB"
,
"tsb_1MB"
,
};
void
__init
tsb_cache_init
(
void
)
{
unsigned
long
i
;
for
(
i
=
0
;
i
<
8
;
i
++
)
{
unsigned
long
size
=
8192
<<
i
;
const
char
*
name
=
tsb_cache_names
[
i
];
tsb_caches
[
i
]
=
kmem_cache_create
(
name
,
size
,
size
,
SLAB_HWCACHE_ALIGN
|
SLAB_MUST_HWCACHE_ALIGN
,
NULL
,
NULL
);
if
(
!
tsb_caches
[
i
])
{
prom_printf
(
"Could not create %s cache
\n
"
,
name
);
prom_halt
();
}
}
}
/* When the RSS of an address space exceeds mm->context.tsb_rss_limit,
/* When the RSS of an address space exceeds mm->context.tsb_rss_limit,
* do_sparc64_fault() invokes this routine to try and grow the TSB.
* do_sparc64_fault() invokes this routine to try and grow the TSB.
*
*
...
@@ -226,45 +260,48 @@ static void setup_tsb_params(struct mm_struct *mm, unsigned long tsb_bytes)
...
@@ -226,45 +260,48 @@ static void setup_tsb_params(struct mm_struct *mm, unsigned long tsb_bytes)
void
tsb_grow
(
struct
mm_struct
*
mm
,
unsigned
long
rss
)
void
tsb_grow
(
struct
mm_struct
*
mm
,
unsigned
long
rss
)
{
{
unsigned
long
max_tsb_size
=
1
*
1024
*
1024
;
unsigned
long
max_tsb_size
=
1
*
1024
*
1024
;
unsigned
long
size
,
old_size
,
flags
;
unsigned
long
new_size
,
old_size
,
flags
;
struct
page
*
page
;
struct
tsb
*
old_tsb
,
*
new_tsb
;
struct
tsb
*
old_tsb
,
*
new_tsb
;
unsigned
long
order
,
new_rss_limit
;
unsigned
long
new_cache_index
,
old_cache_index
;
unsigned
long
new_rss_limit
;
gfp_t
gfp_flags
;
gfp_t
gfp_flags
;
if
(
max_tsb_size
>
(
PAGE_SIZE
<<
MAX_ORDER
))
if
(
max_tsb_size
>
(
PAGE_SIZE
<<
MAX_ORDER
))
max_tsb_size
=
(
PAGE_SIZE
<<
MAX_ORDER
);
max_tsb_size
=
(
PAGE_SIZE
<<
MAX_ORDER
);
for
(
size
=
PAGE_SIZE
;
size
<
max_tsb_size
;
size
<<=
1UL
)
{
new_cache_index
=
0
;
unsigned
long
n_entries
=
size
/
sizeof
(
struct
tsb
);
for
(
new_size
=
8192
;
new_size
<
max_tsb_size
;
new_size
<<=
1UL
)
{
unsigned
long
n_entries
=
new_size
/
sizeof
(
struct
tsb
);
n_entries
=
(
n_entries
*
3
)
/
4
;
n_entries
=
(
n_entries
*
3
)
/
4
;
if
(
n_entries
>
rss
)
if
(
n_entries
>
rss
)
break
;
break
;
new_cache_index
++
;
}
}
if
(
size
==
max_tsb_size
)
if
(
new_
size
==
max_tsb_size
)
new_rss_limit
=
~
0UL
;
new_rss_limit
=
~
0UL
;
else
else
new_rss_limit
=
((
size
/
sizeof
(
struct
tsb
))
*
3
)
/
4
;
new_rss_limit
=
((
new_
size
/
sizeof
(
struct
tsb
))
*
3
)
/
4
;
retry_page_alloc:
retry_tsb_alloc:
order
=
get_order
(
size
);
gfp_flags
=
GFP_KERNEL
;
gfp_flags
=
GFP_KERNEL
;
if
(
order
>
1
)
if
(
new_size
>
(
PAGE_SIZE
*
2
)
)
gfp_flags
=
__GFP_NOWARN
|
__GFP_NORETRY
;
gfp_flags
=
__GFP_NOWARN
|
__GFP_NORETRY
;
page
=
alloc_pages
(
gfp_flags
,
order
);
new_tsb
=
kmem_cache_alloc
(
tsb_caches
[
new_cache_index
],
gfp_flags
);
if
(
unlikely
(
!
page
))
{
if
(
unlikely
(
!
new_tsb
))
{
/* Not being able to fork due to a high-order TSB
/* Not being able to fork due to a high-order TSB
* allocation failure is very bad behavior. Just back
* allocation failure is very bad behavior. Just back
* down to a 0-order allocation and force no TSB
* down to a 0-order allocation and force no TSB
* growing for this address space.
* growing for this address space.
*/
*/
if
(
mm
->
context
.
tsb
==
NULL
&&
order
>
0
)
{
if
(
mm
->
context
.
tsb
==
NULL
&&
new_cache_index
>
0
)
{
size
=
PAGE_SIZE
;
new_cache_index
=
0
;
new_size
=
8192
;
new_rss_limit
=
~
0UL
;
new_rss_limit
=
~
0UL
;
goto
retry_
page
_alloc
;
goto
retry_
tsb
_alloc
;
}
}
/* If we failed on a TSB grow, we are under serious
/* If we failed on a TSB grow, we are under serious
...
@@ -276,8 +313,7 @@ retry_page_alloc:
...
@@ -276,8 +313,7 @@ retry_page_alloc:
}
}
/* Mark all tags as invalid. */
/* Mark all tags as invalid. */
new_tsb
=
page_address
(
page
);
memset
(
new_tsb
,
0x40
,
new_size
);
memset
(
new_tsb
,
0x40
,
size
);
/* Ok, we are about to commit the changes. If we are
/* Ok, we are about to commit the changes. If we are
* growing an existing TSB the locking is very tricky,
* growing an existing TSB the locking is very tricky,
...
@@ -304,8 +340,10 @@ retry_page_alloc:
...
@@ -304,8 +340,10 @@ retry_page_alloc:
spin_lock_irqsave
(
&
mm
->
context
.
lock
,
flags
);
spin_lock_irqsave
(
&
mm
->
context
.
lock
,
flags
);
old_tsb
=
mm
->
context
.
tsb
;
old_tsb
=
mm
->
context
.
tsb
;
old_cache_index
=
(
mm
->
context
.
tsb_reg_val
&
0x7UL
);
old_size
=
mm
->
context
.
tsb_nentries
*
sizeof
(
struct
tsb
);
old_size
=
mm
->
context
.
tsb_nentries
*
sizeof
(
struct
tsb
);
/* Handle multiple threads trying to grow the TSB at the same time.
/* Handle multiple threads trying to grow the TSB at the same time.
* One will get in here first, and bump the size and the RSS limit.
* One will get in here first, and bump the size and the RSS limit.
* The others will get in here next and hit this check.
* The others will get in here next and hit this check.
...
@@ -313,7 +351,7 @@ retry_page_alloc:
...
@@ -313,7 +351,7 @@ retry_page_alloc:
if
(
unlikely
(
old_tsb
&&
(
rss
<
mm
->
context
.
tsb_rss_limit
)))
{
if
(
unlikely
(
old_tsb
&&
(
rss
<
mm
->
context
.
tsb_rss_limit
)))
{
spin_unlock_irqrestore
(
&
mm
->
context
.
lock
,
flags
);
spin_unlock_irqrestore
(
&
mm
->
context
.
lock
,
flags
);
free_pages
((
unsigned
long
)
new_tsb
,
get_order
(
size
)
);
kmem_cache_free
(
tsb_caches
[
new_cache_index
],
new_tsb
);
return
;
return
;
}
}
...
@@ -331,11 +369,11 @@ retry_page_alloc:
...
@@ -331,11 +369,11 @@ retry_page_alloc:
old_tsb_base
=
__pa
(
old_tsb_base
);
old_tsb_base
=
__pa
(
old_tsb_base
);
new_tsb_base
=
__pa
(
new_tsb_base
);
new_tsb_base
=
__pa
(
new_tsb_base
);
}
}
copy_tsb
(
old_tsb_base
,
old_size
,
new_tsb_base
,
size
);
copy_tsb
(
old_tsb_base
,
old_size
,
new_tsb_base
,
new_
size
);
}
}
mm
->
context
.
tsb
=
new_tsb
;
mm
->
context
.
tsb
=
new_tsb
;
setup_tsb_params
(
mm
,
size
);
setup_tsb_params
(
mm
,
new_
size
);
spin_unlock_irqrestore
(
&
mm
->
context
.
lock
,
flags
);
spin_unlock_irqrestore
(
&
mm
->
context
.
lock
,
flags
);
...
@@ -350,7 +388,7 @@ retry_page_alloc:
...
@@ -350,7 +388,7 @@ retry_page_alloc:
smp_tsb_sync
(
mm
);
smp_tsb_sync
(
mm
);
/* Now it is safe to free the old tsb. */
/* Now it is safe to free the old tsb. */
free_pages
((
unsigned
long
)
old_tsb
,
get_order
(
old_size
)
);
kmem_cache_free
(
tsb_caches
[
old_cache_index
],
old_tsb
);
}
}
}
}
...
@@ -379,10 +417,10 @@ int init_new_context(struct task_struct *tsk, struct mm_struct *mm)
...
@@ -379,10 +417,10 @@ int init_new_context(struct task_struct *tsk, struct mm_struct *mm)
void
destroy_context
(
struct
mm_struct
*
mm
)
void
destroy_context
(
struct
mm_struct
*
mm
)
{
{
unsigned
long
size
=
mm
->
context
.
tsb_nentries
*
sizeof
(
struct
tsb
);
unsigned
long
flags
,
cache_index
;
unsigned
long
flags
;
free_pages
((
unsigned
long
)
mm
->
context
.
tsb
,
get_order
(
size
));
cache_index
=
(
mm
->
context
.
tsb_reg_val
&
0x7UL
);
kmem_cache_free
(
tsb_caches
[
cache_index
],
mm
->
context
.
tsb
);
/* We can remove these later, but for now it's useful
/* We can remove these later, but for now it's useful
* to catch any bogus post-destroy_context() references
* to catch any bogus post-destroy_context() references
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment