Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
L
linux-davinci
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Redmine
Redmine
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Operations
Operations
Metrics
Environments
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
linux
linux-davinci
Commits
5b0830cb
Commit
5b0830cb
authored
Sep 23, 2009
by
Jens Axboe
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
writeback: get rid to incorrect references to pdflush in comments
Signed-off-by:
Jens Axboe
<
jens.axboe@oracle.com
>
parent
71fd05a8
Changes
5
Hide whitespace changes
Inline
Side-by-side
Showing
5 changed files
with
17 additions
and
19 deletions
+17
-19
fs/buffer.c
fs/buffer.c
+5
-5
fs/fs-writeback.c
fs/fs-writeback.c
+1
-4
mm/page-writeback.c
mm/page-writeback.c
+4
-4
mm/shmem.c
mm/shmem.c
+3
-2
mm/vmscan.c
mm/vmscan.c
+4
-4
No files found.
fs/buffer.c
View file @
5b0830cb
...
...
@@ -274,7 +274,7 @@ void invalidate_bdev(struct block_device *bdev)
}
/*
* Kick
pdflush
then try to free up some ZONE_NORMAL memory.
* Kick
the writeback threads
then try to free up some ZONE_NORMAL memory.
*/
static
void
free_more_memory
(
void
)
{
...
...
@@ -1699,9 +1699,9 @@ static int __block_write_full_page(struct inode *inode, struct page *page,
/*
* If it's a fully non-blocking write attempt and we cannot
* lock the buffer then redirty the page. Note that this can
* potentially cause a busy-wait loop from
pdflush and kswapd
* a
ctivity, but those code paths have their own higher-level
* throttling.
* potentially cause a busy-wait loop from
writeback threads
* a
nd kswapd activity, but those code paths have their own
*
higher-level
throttling.
*/
if
(
wbc
->
sync_mode
!=
WB_SYNC_NONE
||
!
wbc
->
nonblocking
)
{
lock_buffer
(
bh
);
...
...
@@ -3191,7 +3191,7 @@ void block_sync_page(struct page *page)
* still running obsolete flush daemons, so we terminate them here.
*
* Use of bdflush() is deprecated and will be removed in a future kernel.
* The `
pdflush
' kernel threads fully replace bdflush daemons and this call.
* The `
flush-X
' kernel threads fully replace bdflush daemons and this call.
*/
SYSCALL_DEFINE2
(
bdflush
,
int
,
func
,
long
,
data
)
{
...
...
fs/fs-writeback.c
View file @
5b0830cb
...
...
@@ -320,7 +320,7 @@ static bool inode_dirtied_after(struct inode *inode, unsigned long t)
* For inodes being constantly redirtied, dirtied_when can get stuck.
* It _appears_ to be in the future, but is actually in distant past.
* This test is necessary to prevent such wrapped-around relative times
* from permanently stopping the whole
pdflush
writeback.
* from permanently stopping the whole
bdi
writeback.
*/
ret
=
ret
&&
time_before_eq
(
inode
->
dirtied_when
,
jiffies
);
#endif
...
...
@@ -1085,9 +1085,6 @@ EXPORT_SYMBOL(__mark_inode_dirty);
* If older_than_this is non-NULL, then only write out inodes which
* had their first dirtying at a time earlier than *older_than_this.
*
* If we're a pdlfush thread, then implement pdflush collision avoidance
* against the entire list.
*
* If `bdi' is non-zero then we're being asked to writeback a specific queue.
* This function assumes that the blockdev superblock's inodes are backed by
* a variety of queues, so all inodes are searched. For other superblocks,
...
...
mm/page-writeback.c
View file @
5b0830cb
...
...
@@ -58,7 +58,7 @@ static inline long sync_writeback_pages(unsigned long dirtied)
/* The following parameters are exported via /proc/sys/vm */
/*
* Start background writeback (via
pdflush
) at this percentage
* Start background writeback (via
writeback threads
) at this percentage
*/
int
dirty_background_ratio
=
10
;
...
...
@@ -477,8 +477,8 @@ get_dirty_limits(unsigned long *pbackground, unsigned long *pdirty,
* balance_dirty_pages() must be called by processes which are generating dirty
* data. It looks at the number of dirty pages in the machine and will force
* the caller to perform writeback if the system is over `vm_dirty_ratio'.
* If we're over `background_thresh' then
pdflush is woken to perform some
* writeout.
* If we're over `background_thresh' then
the writeback threads are woken to
*
perform some
writeout.
*/
static
void
balance_dirty_pages
(
struct
address_space
*
mapping
,
unsigned
long
write_chunk
)
...
...
@@ -582,7 +582,7 @@ static void balance_dirty_pages(struct address_space *mapping,
bdi
->
dirty_exceeded
=
0
;
if
(
writeback_in_progress
(
bdi
))
return
;
/* pdflush is already working this queue */
return
;
/*
* In laptop mode, we wait until hitting the higher threshold before
...
...
mm/shmem.c
View file @
5b0830cb
...
...
@@ -1046,8 +1046,9 @@ static int shmem_writepage(struct page *page, struct writeback_control *wbc)
* sync from ever calling shmem_writepage; but a stacking filesystem
* may use the ->writepage of its underlying filesystem, in which case
* tmpfs should write out to swap only in response to memory pressure,
* and not for pdflush or sync. However, in those cases, we do still
* want to check if there's a redundant swappage to be discarded.
* and not for the writeback threads or sync. However, in those cases,
* we do still want to check if there's a redundant swappage to be
* discarded.
*/
if
(
wbc
->
for_reclaim
)
swap
=
get_swap_page
();
...
...
mm/vmscan.c
View file @
5b0830cb
...
...
@@ -1709,10 +1709,10 @@ static void shrink_zones(int priority, struct zonelist *zonelist,
*
* If the caller is !__GFP_FS then the probability of a failure is reasonably
* high - the zone may be full of dirty or under-writeback pages, which this
* caller can't do much about. We kick
pdflush and take explicit naps in the
*
hope that some of these pages can be written. But if the allocating task
*
holds filesystem locks which prevent writeout this might not work, and the
* allocation attempt will fail.
* caller can't do much about. We kick
the writeback threads and take explicit
*
naps in the hope that some of these pages can be written. But if the
*
allocating task holds filesystem locks which prevent writeout this might not
*
work, and the
allocation attempt will fail.
*
* returns: 0, if no pages reclaimed
* else, the number of pages reclaimed
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment