|
@@ -2414,8 +2414,7 @@ int __set_page_dirty_no_writeback(struct page *page)
|
|
*
|
|
*
|
|
* NOTE: This relies on being atomic wrt interrupts.
|
|
* NOTE: This relies on being atomic wrt interrupts.
|
|
*/
|
|
*/
|
|
-void account_page_dirtied(struct page *page, struct address_space *mapping,
|
|
|
|
- struct mem_cgroup *memcg)
|
|
|
|
|
|
+void account_page_dirtied(struct page *page, struct address_space *mapping)
|
|
{
|
|
{
|
|
struct inode *inode = mapping->host;
|
|
struct inode *inode = mapping->host;
|
|
|
|
|
|
@@ -2427,7 +2426,7 @@ void account_page_dirtied(struct page *page, struct address_space *mapping,
|
|
inode_attach_wb(inode, page);
|
|
inode_attach_wb(inode, page);
|
|
wb = inode_to_wb(inode);
|
|
wb = inode_to_wb(inode);
|
|
|
|
|
|
- mem_cgroup_inc_page_stat(memcg, MEM_CGROUP_STAT_DIRTY);
|
|
|
|
|
|
+ mem_cgroup_inc_page_stat(page, MEM_CGROUP_STAT_DIRTY);
|
|
__inc_zone_page_state(page, NR_FILE_DIRTY);
|
|
__inc_zone_page_state(page, NR_FILE_DIRTY);
|
|
__inc_zone_page_state(page, NR_DIRTIED);
|
|
__inc_zone_page_state(page, NR_DIRTIED);
|
|
__inc_wb_stat(wb, WB_RECLAIMABLE);
|
|
__inc_wb_stat(wb, WB_RECLAIMABLE);
|
|
@@ -2445,10 +2444,10 @@ EXPORT_SYMBOL(account_page_dirtied);
|
|
* Caller must hold lock_page_memcg().
|
|
* Caller must hold lock_page_memcg().
|
|
*/
|
|
*/
|
|
void account_page_cleaned(struct page *page, struct address_space *mapping,
|
|
void account_page_cleaned(struct page *page, struct address_space *mapping,
|
|
- struct mem_cgroup *memcg, struct bdi_writeback *wb)
|
|
|
|
|
|
+ struct bdi_writeback *wb)
|
|
{
|
|
{
|
|
if (mapping_cap_account_dirty(mapping)) {
|
|
if (mapping_cap_account_dirty(mapping)) {
|
|
- mem_cgroup_dec_page_stat(memcg, MEM_CGROUP_STAT_DIRTY);
|
|
|
|
|
|
+ mem_cgroup_dec_page_stat(page, MEM_CGROUP_STAT_DIRTY);
|
|
dec_zone_page_state(page, NR_FILE_DIRTY);
|
|
dec_zone_page_state(page, NR_FILE_DIRTY);
|
|
dec_wb_stat(wb, WB_RECLAIMABLE);
|
|
dec_wb_stat(wb, WB_RECLAIMABLE);
|
|
task_io_account_cancelled_write(PAGE_CACHE_SIZE);
|
|
task_io_account_cancelled_write(PAGE_CACHE_SIZE);
|
|
@@ -2469,26 +2468,24 @@ void account_page_cleaned(struct page *page, struct address_space *mapping,
|
|
*/
|
|
*/
|
|
int __set_page_dirty_nobuffers(struct page *page)
|
|
int __set_page_dirty_nobuffers(struct page *page)
|
|
{
|
|
{
|
|
- struct mem_cgroup *memcg;
|
|
|
|
-
|
|
|
|
- memcg = lock_page_memcg(page);
|
|
|
|
|
|
+ lock_page_memcg(page);
|
|
if (!TestSetPageDirty(page)) {
|
|
if (!TestSetPageDirty(page)) {
|
|
struct address_space *mapping = page_mapping(page);
|
|
struct address_space *mapping = page_mapping(page);
|
|
unsigned long flags;
|
|
unsigned long flags;
|
|
|
|
|
|
if (!mapping) {
|
|
if (!mapping) {
|
|
- unlock_page_memcg(memcg);
|
|
|
|
|
|
+ unlock_page_memcg(page);
|
|
return 1;
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
|
|
spin_lock_irqsave(&mapping->tree_lock, flags);
|
|
spin_lock_irqsave(&mapping->tree_lock, flags);
|
|
BUG_ON(page_mapping(page) != mapping);
|
|
BUG_ON(page_mapping(page) != mapping);
|
|
WARN_ON_ONCE(!PagePrivate(page) && !PageUptodate(page));
|
|
WARN_ON_ONCE(!PagePrivate(page) && !PageUptodate(page));
|
|
- account_page_dirtied(page, mapping, memcg);
|
|
|
|
|
|
+ account_page_dirtied(page, mapping);
|
|
radix_tree_tag_set(&mapping->page_tree, page_index(page),
|
|
radix_tree_tag_set(&mapping->page_tree, page_index(page),
|
|
PAGECACHE_TAG_DIRTY);
|
|
PAGECACHE_TAG_DIRTY);
|
|
spin_unlock_irqrestore(&mapping->tree_lock, flags);
|
|
spin_unlock_irqrestore(&mapping->tree_lock, flags);
|
|
- unlock_page_memcg(memcg);
|
|
|
|
|
|
+ unlock_page_memcg(page);
|
|
|
|
|
|
if (mapping->host) {
|
|
if (mapping->host) {
|
|
/* !PageAnon && !swapper_space */
|
|
/* !PageAnon && !swapper_space */
|
|
@@ -2496,7 +2493,7 @@ int __set_page_dirty_nobuffers(struct page *page)
|
|
}
|
|
}
|
|
return 1;
|
|
return 1;
|
|
}
|
|
}
|
|
- unlock_page_memcg(memcg);
|
|
|
|
|
|
+ unlock_page_memcg(page);
|
|
return 0;
|
|
return 0;
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(__set_page_dirty_nobuffers);
|
|
EXPORT_SYMBOL(__set_page_dirty_nobuffers);
|
|
@@ -2626,17 +2623,16 @@ void cancel_dirty_page(struct page *page)
|
|
if (mapping_cap_account_dirty(mapping)) {
|
|
if (mapping_cap_account_dirty(mapping)) {
|
|
struct inode *inode = mapping->host;
|
|
struct inode *inode = mapping->host;
|
|
struct bdi_writeback *wb;
|
|
struct bdi_writeback *wb;
|
|
- struct mem_cgroup *memcg;
|
|
|
|
bool locked;
|
|
bool locked;
|
|
|
|
|
|
- memcg = lock_page_memcg(page);
|
|
|
|
|
|
+ lock_page_memcg(page);
|
|
wb = unlocked_inode_to_wb_begin(inode, &locked);
|
|
wb = unlocked_inode_to_wb_begin(inode, &locked);
|
|
|
|
|
|
if (TestClearPageDirty(page))
|
|
if (TestClearPageDirty(page))
|
|
- account_page_cleaned(page, mapping, memcg, wb);
|
|
|
|
|
|
+ account_page_cleaned(page, mapping, wb);
|
|
|
|
|
|
unlocked_inode_to_wb_end(inode, locked);
|
|
unlocked_inode_to_wb_end(inode, locked);
|
|
- unlock_page_memcg(memcg);
|
|
|
|
|
|
+ unlock_page_memcg(page);
|
|
} else {
|
|
} else {
|
|
ClearPageDirty(page);
|
|
ClearPageDirty(page);
|
|
}
|
|
}
|
|
@@ -2667,7 +2663,6 @@ int clear_page_dirty_for_io(struct page *page)
|
|
if (mapping && mapping_cap_account_dirty(mapping)) {
|
|
if (mapping && mapping_cap_account_dirty(mapping)) {
|
|
struct inode *inode = mapping->host;
|
|
struct inode *inode = mapping->host;
|
|
struct bdi_writeback *wb;
|
|
struct bdi_writeback *wb;
|
|
- struct mem_cgroup *memcg;
|
|
|
|
bool locked;
|
|
bool locked;
|
|
|
|
|
|
/*
|
|
/*
|
|
@@ -2705,16 +2700,16 @@ int clear_page_dirty_for_io(struct page *page)
|
|
* always locked coming in here, so we get the desired
|
|
* always locked coming in here, so we get the desired
|
|
* exclusion.
|
|
* exclusion.
|
|
*/
|
|
*/
|
|
- memcg = lock_page_memcg(page);
|
|
|
|
|
|
+ lock_page_memcg(page);
|
|
wb = unlocked_inode_to_wb_begin(inode, &locked);
|
|
wb = unlocked_inode_to_wb_begin(inode, &locked);
|
|
if (TestClearPageDirty(page)) {
|
|
if (TestClearPageDirty(page)) {
|
|
- mem_cgroup_dec_page_stat(memcg, MEM_CGROUP_STAT_DIRTY);
|
|
|
|
|
|
+ mem_cgroup_dec_page_stat(page, MEM_CGROUP_STAT_DIRTY);
|
|
dec_zone_page_state(page, NR_FILE_DIRTY);
|
|
dec_zone_page_state(page, NR_FILE_DIRTY);
|
|
dec_wb_stat(wb, WB_RECLAIMABLE);
|
|
dec_wb_stat(wb, WB_RECLAIMABLE);
|
|
ret = 1;
|
|
ret = 1;
|
|
}
|
|
}
|
|
unlocked_inode_to_wb_end(inode, locked);
|
|
unlocked_inode_to_wb_end(inode, locked);
|
|
- unlock_page_memcg(memcg);
|
|
|
|
|
|
+ unlock_page_memcg(page);
|
|
return ret;
|
|
return ret;
|
|
}
|
|
}
|
|
return TestClearPageDirty(page);
|
|
return TestClearPageDirty(page);
|
|
@@ -2724,10 +2719,9 @@ EXPORT_SYMBOL(clear_page_dirty_for_io);
|
|
int test_clear_page_writeback(struct page *page)
|
|
int test_clear_page_writeback(struct page *page)
|
|
{
|
|
{
|
|
struct address_space *mapping = page_mapping(page);
|
|
struct address_space *mapping = page_mapping(page);
|
|
- struct mem_cgroup *memcg;
|
|
|
|
int ret;
|
|
int ret;
|
|
|
|
|
|
- memcg = lock_page_memcg(page);
|
|
|
|
|
|
+ lock_page_memcg(page);
|
|
if (mapping) {
|
|
if (mapping) {
|
|
struct inode *inode = mapping->host;
|
|
struct inode *inode = mapping->host;
|
|
struct backing_dev_info *bdi = inode_to_bdi(inode);
|
|
struct backing_dev_info *bdi = inode_to_bdi(inode);
|
|
@@ -2751,21 +2745,20 @@ int test_clear_page_writeback(struct page *page)
|
|
ret = TestClearPageWriteback(page);
|
|
ret = TestClearPageWriteback(page);
|
|
}
|
|
}
|
|
if (ret) {
|
|
if (ret) {
|
|
- mem_cgroup_dec_page_stat(memcg, MEM_CGROUP_STAT_WRITEBACK);
|
|
|
|
|
|
+ mem_cgroup_dec_page_stat(page, MEM_CGROUP_STAT_WRITEBACK);
|
|
dec_zone_page_state(page, NR_WRITEBACK);
|
|
dec_zone_page_state(page, NR_WRITEBACK);
|
|
inc_zone_page_state(page, NR_WRITTEN);
|
|
inc_zone_page_state(page, NR_WRITTEN);
|
|
}
|
|
}
|
|
- unlock_page_memcg(memcg);
|
|
|
|
|
|
+ unlock_page_memcg(page);
|
|
return ret;
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
|
|
int __test_set_page_writeback(struct page *page, bool keep_write)
|
|
int __test_set_page_writeback(struct page *page, bool keep_write)
|
|
{
|
|
{
|
|
struct address_space *mapping = page_mapping(page);
|
|
struct address_space *mapping = page_mapping(page);
|
|
- struct mem_cgroup *memcg;
|
|
|
|
int ret;
|
|
int ret;
|
|
|
|
|
|
- memcg = lock_page_memcg(page);
|
|
|
|
|
|
+ lock_page_memcg(page);
|
|
if (mapping) {
|
|
if (mapping) {
|
|
struct inode *inode = mapping->host;
|
|
struct inode *inode = mapping->host;
|
|
struct backing_dev_info *bdi = inode_to_bdi(inode);
|
|
struct backing_dev_info *bdi = inode_to_bdi(inode);
|
|
@@ -2793,10 +2786,10 @@ int __test_set_page_writeback(struct page *page, bool keep_write)
|
|
ret = TestSetPageWriteback(page);
|
|
ret = TestSetPageWriteback(page);
|
|
}
|
|
}
|
|
if (!ret) {
|
|
if (!ret) {
|
|
- mem_cgroup_inc_page_stat(memcg, MEM_CGROUP_STAT_WRITEBACK);
|
|
|
|
|
|
+ mem_cgroup_inc_page_stat(page, MEM_CGROUP_STAT_WRITEBACK);
|
|
inc_zone_page_state(page, NR_WRITEBACK);
|
|
inc_zone_page_state(page, NR_WRITEBACK);
|
|
}
|
|
}
|
|
- unlock_page_memcg(memcg);
|
|
|
|
|
|
+ unlock_page_memcg(page);
|
|
return ret;
|
|
return ret;
|
|
|
|
|
|
}
|
|
}
|