mirror of
https://github.com/torvalds/linux.git
synced 2024-12-28 13:51:44 +00:00
memcg: fix deadlock between lock_page_cgroup and mapping tree_lock
mapping->tree_lock can be acquired from interrupt context. Then, following dead lock can occur. Assume "A" as a page. CPU0: lock_page_cgroup(A) interrupted -> take mapping->tree_lock. CPU1: take mapping->tree_lock -> lock_page_cgroup(A) This patch tries to fix above deadlock by moving memcg's hook to out of mapping->tree_lock. charge/uncharge of pagecache/swapcache is protected by page lock, not tree_lock. After this patch, lock_page_cgroup() is not called under mapping->tree_lock. Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com> Signed-off-by: Daisuke Nishimura <nishimura@mxp.nes.nec.co.jp> Cc: Balbir Singh <balbir@in.ibm.com> Cc: Daisuke Nishimura <nishimura@mxp.nes.nec.co.jp> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
bd6daba909
commit
e767e0561d
@ -437,6 +437,11 @@ static inline int mem_cgroup_cache_charge_swapin(struct page *page,
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline void
|
||||||
|
mem_cgroup_uncharge_swapcache(struct page *page, swp_entry_t ent)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
#endif /* CONFIG_SWAP */
|
#endif /* CONFIG_SWAP */
|
||||||
#endif /* __KERNEL__*/
|
#endif /* __KERNEL__*/
|
||||||
#endif /* _LINUX_SWAP_H */
|
#endif /* _LINUX_SWAP_H */
|
||||||
|
@ -121,7 +121,6 @@ void __remove_from_page_cache(struct page *page)
|
|||||||
mapping->nrpages--;
|
mapping->nrpages--;
|
||||||
__dec_zone_page_state(page, NR_FILE_PAGES);
|
__dec_zone_page_state(page, NR_FILE_PAGES);
|
||||||
BUG_ON(page_mapped(page));
|
BUG_ON(page_mapped(page));
|
||||||
mem_cgroup_uncharge_cache_page(page);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Some filesystems seem to re-dirty the page even after
|
* Some filesystems seem to re-dirty the page even after
|
||||||
@ -145,6 +144,7 @@ void remove_from_page_cache(struct page *page)
|
|||||||
spin_lock_irq(&mapping->tree_lock);
|
spin_lock_irq(&mapping->tree_lock);
|
||||||
__remove_from_page_cache(page);
|
__remove_from_page_cache(page);
|
||||||
spin_unlock_irq(&mapping->tree_lock);
|
spin_unlock_irq(&mapping->tree_lock);
|
||||||
|
mem_cgroup_uncharge_cache_page(page);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int sync_page(void *word)
|
static int sync_page(void *word)
|
||||||
@ -476,13 +476,13 @@ int add_to_page_cache_locked(struct page *page, struct address_space *mapping,
|
|||||||
if (likely(!error)) {
|
if (likely(!error)) {
|
||||||
mapping->nrpages++;
|
mapping->nrpages++;
|
||||||
__inc_zone_page_state(page, NR_FILE_PAGES);
|
__inc_zone_page_state(page, NR_FILE_PAGES);
|
||||||
|
spin_unlock_irq(&mapping->tree_lock);
|
||||||
} else {
|
} else {
|
||||||
page->mapping = NULL;
|
page->mapping = NULL;
|
||||||
|
spin_unlock_irq(&mapping->tree_lock);
|
||||||
mem_cgroup_uncharge_cache_page(page);
|
mem_cgroup_uncharge_cache_page(page);
|
||||||
page_cache_release(page);
|
page_cache_release(page);
|
||||||
}
|
}
|
||||||
|
|
||||||
spin_unlock_irq(&mapping->tree_lock);
|
|
||||||
radix_tree_preload_end();
|
radix_tree_preload_end();
|
||||||
} else
|
} else
|
||||||
mem_cgroup_uncharge_cache_page(page);
|
mem_cgroup_uncharge_cache_page(page);
|
||||||
|
@ -1488,8 +1488,9 @@ void mem_cgroup_uncharge_cache_page(struct page *page)
|
|||||||
__mem_cgroup_uncharge_common(page, MEM_CGROUP_CHARGE_TYPE_CACHE);
|
__mem_cgroup_uncharge_common(page, MEM_CGROUP_CHARGE_TYPE_CACHE);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#ifdef CONFIG_SWAP
|
||||||
/*
|
/*
|
||||||
* called from __delete_from_swap_cache() and drop "page" account.
|
* called after __delete_from_swap_cache() and drop "page" account.
|
||||||
* memcg information is recorded to swap_cgroup of "ent"
|
* memcg information is recorded to swap_cgroup of "ent"
|
||||||
*/
|
*/
|
||||||
void mem_cgroup_uncharge_swapcache(struct page *page, swp_entry_t ent)
|
void mem_cgroup_uncharge_swapcache(struct page *page, swp_entry_t ent)
|
||||||
@ -1506,6 +1507,7 @@ void mem_cgroup_uncharge_swapcache(struct page *page, swp_entry_t ent)
|
|||||||
if (memcg)
|
if (memcg)
|
||||||
css_put(&memcg->css);
|
css_put(&memcg->css);
|
||||||
}
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
#ifdef CONFIG_CGROUP_MEM_RES_CTLR_SWAP
|
#ifdef CONFIG_CGROUP_MEM_RES_CTLR_SWAP
|
||||||
/*
|
/*
|
||||||
|
@ -109,8 +109,6 @@ int add_to_swap_cache(struct page *page, swp_entry_t entry, gfp_t gfp_mask)
|
|||||||
*/
|
*/
|
||||||
void __delete_from_swap_cache(struct page *page)
|
void __delete_from_swap_cache(struct page *page)
|
||||||
{
|
{
|
||||||
swp_entry_t ent = {.val = page_private(page)};
|
|
||||||
|
|
||||||
VM_BUG_ON(!PageLocked(page));
|
VM_BUG_ON(!PageLocked(page));
|
||||||
VM_BUG_ON(!PageSwapCache(page));
|
VM_BUG_ON(!PageSwapCache(page));
|
||||||
VM_BUG_ON(PageWriteback(page));
|
VM_BUG_ON(PageWriteback(page));
|
||||||
@ -121,7 +119,6 @@ void __delete_from_swap_cache(struct page *page)
|
|||||||
total_swapcache_pages--;
|
total_swapcache_pages--;
|
||||||
__dec_zone_page_state(page, NR_FILE_PAGES);
|
__dec_zone_page_state(page, NR_FILE_PAGES);
|
||||||
INC_CACHE_INFO(del_total);
|
INC_CACHE_INFO(del_total);
|
||||||
mem_cgroup_uncharge_swapcache(page, ent);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -191,6 +188,7 @@ void delete_from_swap_cache(struct page *page)
|
|||||||
__delete_from_swap_cache(page);
|
__delete_from_swap_cache(page);
|
||||||
spin_unlock_irq(&swapper_space.tree_lock);
|
spin_unlock_irq(&swapper_space.tree_lock);
|
||||||
|
|
||||||
|
mem_cgroup_uncharge_swapcache(page, entry);
|
||||||
swap_free(entry);
|
swap_free(entry);
|
||||||
page_cache_release(page);
|
page_cache_release(page);
|
||||||
}
|
}
|
||||||
|
@ -359,6 +359,7 @@ invalidate_complete_page2(struct address_space *mapping, struct page *page)
|
|||||||
BUG_ON(page_has_private(page));
|
BUG_ON(page_has_private(page));
|
||||||
__remove_from_page_cache(page);
|
__remove_from_page_cache(page);
|
||||||
spin_unlock_irq(&mapping->tree_lock);
|
spin_unlock_irq(&mapping->tree_lock);
|
||||||
|
mem_cgroup_uncharge_cache_page(page);
|
||||||
page_cache_release(page); /* pagecache ref */
|
page_cache_release(page); /* pagecache ref */
|
||||||
return 1;
|
return 1;
|
||||||
failed:
|
failed:
|
||||||
|
@ -470,10 +470,12 @@ static int __remove_mapping(struct address_space *mapping, struct page *page)
|
|||||||
swp_entry_t swap = { .val = page_private(page) };
|
swp_entry_t swap = { .val = page_private(page) };
|
||||||
__delete_from_swap_cache(page);
|
__delete_from_swap_cache(page);
|
||||||
spin_unlock_irq(&mapping->tree_lock);
|
spin_unlock_irq(&mapping->tree_lock);
|
||||||
|
mem_cgroup_uncharge_swapcache(page, swap);
|
||||||
swap_free(swap);
|
swap_free(swap);
|
||||||
} else {
|
} else {
|
||||||
__remove_from_page_cache(page);
|
__remove_from_page_cache(page);
|
||||||
spin_unlock_irq(&mapping->tree_lock);
|
spin_unlock_irq(&mapping->tree_lock);
|
||||||
|
mem_cgroup_uncharge_cache_page(page);
|
||||||
}
|
}
|
||||||
|
|
||||||
return 1;
|
return 1;
|
||||||
|
Loading…
Reference in New Issue
Block a user