Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
L
linux-davinci
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Redmine
Redmine
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Operations
Operations
Metrics
Environments
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
linux
linux-davinci
Commits
09be207d
Commit
09be207d
authored
Nov 07, 2007
by
Chris Mason
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
Btrfs: Fix failure cleanups when allocating extent buffers fail
Signed-off-by:
Chris Mason
<
chris.mason@oracle.com
>
parent
0da5468f
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
30 additions
and
26 deletions
+30
-26
fs/btrfs/extent_map.c
fs/btrfs/extent_map.c
+30
-26
No files found.
fs/btrfs/extent_map.c
View file @
09be207d
...
@@ -2106,25 +2106,17 @@ static struct extent_buffer *__alloc_extent_buffer(struct extent_map_tree *tree,
...
@@ -2106,25 +2106,17 @@ static struct extent_buffer *__alloc_extent_buffer(struct extent_map_tree *tree,
spin_lock
(
&
tree
->
lru_lock
);
spin_lock
(
&
tree
->
lru_lock
);
eb
=
find_lru
(
tree
,
start
,
len
);
eb
=
find_lru
(
tree
,
start
,
len
);
if
(
eb
)
{
goto
lru_add
;
}
spin_unlock
(
&
tree
->
lru_lock
);
spin_unlock
(
&
tree
->
lru_lock
);
if
(
eb
)
{
if
(
eb
)
{
memset
(
eb
,
0
,
sizeof
(
*
eb
));
return
eb
;
}
else
{
eb
=
kmem_cache_zalloc
(
extent_buffer_cache
,
mask
);
}
}
eb
=
kmem_cache_zalloc
(
extent_buffer_cache
,
mask
);
INIT_LIST_HEAD
(
&
eb
->
lru
);
INIT_LIST_HEAD
(
&
eb
->
lru
);
eb
->
start
=
start
;
eb
->
start
=
start
;
eb
->
len
=
len
;
eb
->
len
=
len
;
atomic_set
(
&
eb
->
refs
,
1
);
atomic_set
(
&
eb
->
refs
,
1
);
spin_lock
(
&
tree
->
lru_lock
);
lru_add:
add_lru
(
tree
,
eb
);
spin_unlock
(
&
tree
->
lru_lock
);
return
eb
;
return
eb
;
}
}
...
@@ -2151,7 +2143,7 @@ struct extent_buffer *alloc_extent_buffer(struct extent_map_tree *tree,
...
@@ -2151,7 +2143,7 @@ struct extent_buffer *alloc_extent_buffer(struct extent_map_tree *tree,
return
NULL
;
return
NULL
;
if
(
eb
->
flags
&
EXTENT_BUFFER_FILLED
)
if
(
eb
->
flags
&
EXTENT_BUFFER_FILLED
)
return
eb
;
goto
lru_add
;
if
(
page0
)
{
if
(
page0
)
{
eb
->
first_page
=
page0
;
eb
->
first_page
=
page0
;
...
@@ -2169,11 +2161,6 @@ struct extent_buffer *alloc_extent_buffer(struct extent_map_tree *tree,
...
@@ -2169,11 +2161,6 @@ struct extent_buffer *alloc_extent_buffer(struct extent_map_tree *tree,
p
=
find_or_create_page
(
mapping
,
index
,
mask
|
__GFP_HIGHMEM
);
p
=
find_or_create_page
(
mapping
,
index
,
mask
|
__GFP_HIGHMEM
);
if
(
!
p
)
{
if
(
!
p
)
{
WARN_ON
(
1
);
WARN_ON
(
1
);
/* make sure the free only frees the pages we've
* grabbed a reference on
*/
eb
->
len
=
i
<<
PAGE_CACHE_SHIFT
;
eb
->
start
&=
~
((
u64
)
PAGE_CACHE_SIZE
-
1
);
goto
fail
;
goto
fail
;
}
}
set_page_extent_mapped
(
p
);
set_page_extent_mapped
(
p
);
...
@@ -2192,9 +2179,20 @@ struct extent_buffer *alloc_extent_buffer(struct extent_map_tree *tree,
...
@@ -2192,9 +2179,20 @@ struct extent_buffer *alloc_extent_buffer(struct extent_map_tree *tree,
if
(
uptodate
)
if
(
uptodate
)
eb
->
flags
|=
EXTENT_UPTODATE
;
eb
->
flags
|=
EXTENT_UPTODATE
;
eb
->
flags
|=
EXTENT_BUFFER_FILLED
;
eb
->
flags
|=
EXTENT_BUFFER_FILLED
;
lru_add:
spin_lock
(
&
tree
->
lru_lock
);
add_lru
(
tree
,
eb
);
spin_unlock
(
&
tree
->
lru_lock
);
return
eb
;
return
eb
;
fail:
fail:
free_extent_buffer
(
eb
);
if
(
!
atomic_dec_and_test
(
&
eb
->
refs
))
return
NULL
;
for
(
index
=
0
;
index
<
i
;
index
++
)
{
page_cache_release
(
extent_buffer_page
(
eb
,
index
));
}
__free_extent_buffer
(
eb
);
return
NULL
;
return
NULL
;
}
}
EXPORT_SYMBOL
(
alloc_extent_buffer
);
EXPORT_SYMBOL
(
alloc_extent_buffer
);
...
@@ -2204,7 +2202,8 @@ struct extent_buffer *find_extent_buffer(struct extent_map_tree *tree,
...
@@ -2204,7 +2202,8 @@ struct extent_buffer *find_extent_buffer(struct extent_map_tree *tree,
gfp_t
mask
)
gfp_t
mask
)
{
{
unsigned
long
num_pages
=
num_extent_pages
(
start
,
len
);
unsigned
long
num_pages
=
num_extent_pages
(
start
,
len
);
unsigned
long
i
;
unsigned
long
index
=
start
>>
PAGE_CACHE_SHIFT
;
unsigned
long
i
;
unsigned
long
index
=
start
>>
PAGE_CACHE_SHIFT
;
struct
extent_buffer
*
eb
;
struct
extent_buffer
*
eb
;
struct
page
*
p
;
struct
page
*
p
;
struct
address_space
*
mapping
=
tree
->
mapping
;
struct
address_space
*
mapping
=
tree
->
mapping
;
...
@@ -2215,16 +2214,11 @@ struct extent_buffer *find_extent_buffer(struct extent_map_tree *tree,
...
@@ -2215,16 +2214,11 @@ struct extent_buffer *find_extent_buffer(struct extent_map_tree *tree,
return
NULL
;
return
NULL
;
if
(
eb
->
flags
&
EXTENT_BUFFER_FILLED
)
if
(
eb
->
flags
&
EXTENT_BUFFER_FILLED
)
return
eb
;
goto
lru_add
;
for
(
i
=
0
;
i
<
num_pages
;
i
++
,
index
++
)
{
for
(
i
=
0
;
i
<
num_pages
;
i
++
,
index
++
)
{
p
=
find_lock_page
(
mapping
,
index
);
p
=
find_lock_page
(
mapping
,
index
);
if
(
!
p
)
{
if
(
!
p
)
{
/* make sure the free only frees the pages we've
* grabbed a reference on
*/
eb
->
len
=
i
<<
PAGE_CACHE_SHIFT
;
eb
->
start
&=
~
((
u64
)
PAGE_CACHE_SIZE
-
1
);
goto
fail
;
goto
fail
;
}
}
set_page_extent_mapped
(
p
);
set_page_extent_mapped
(
p
);
...
@@ -2245,9 +2239,19 @@ struct extent_buffer *find_extent_buffer(struct extent_map_tree *tree,
...
@@ -2245,9 +2239,19 @@ struct extent_buffer *find_extent_buffer(struct extent_map_tree *tree,
if
(
uptodate
)
if
(
uptodate
)
eb
->
flags
|=
EXTENT_UPTODATE
;
eb
->
flags
|=
EXTENT_UPTODATE
;
eb
->
flags
|=
EXTENT_BUFFER_FILLED
;
eb
->
flags
|=
EXTENT_BUFFER_FILLED
;
lru_add:
spin_lock
(
&
tree
->
lru_lock
);
add_lru
(
tree
,
eb
);
spin_unlock
(
&
tree
->
lru_lock
);
return
eb
;
return
eb
;
fail:
fail:
free_extent_buffer
(
eb
);
if
(
!
atomic_dec_and_test
(
&
eb
->
refs
))
return
NULL
;
for
(
index
=
0
;
index
<
i
;
index
++
)
{
page_cache_release
(
extent_buffer_page
(
eb
,
index
));
}
__free_extent_buffer
(
eb
);
return
NULL
;
return
NULL
;
}
}
EXPORT_SYMBOL
(
find_extent_buffer
);
EXPORT_SYMBOL
(
find_extent_buffer
);
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment