From: "Huang, Ying" <ying.huang@intel.com> To: Andrew Morton <akpm@linux-foundation.org> Cc: tim.c.chen@intel.com, dave.hansen@intel.com, andi.kleen@intel.com, aaron.lu@intel.com, linux-mm@kvack.org, linux-kernel@vger.kernel.org, Huang Ying <ying.huang@intel.com>, Andrea Arcangeli <aarcange@redhat.com>, "Kirill A . Shutemov" <kirill.shutemov@linux.intel.com>, Vladimir Davydov <vdavydov@virtuozzo.com>, Johannes Weiner <hannes@cmpxchg.org>, Michal Hocko <mhocko@kernel.org>, Tejun Heo <tj@kernel.org>, cgroups@vger.kernel.org Subject: [PATCH -v4 RESEND 2/9] mm, memcg: Support to charge/uncharge multiple swap entries Date: Fri, 28 Oct 2016 13:56:01 +0800 [thread overview] Message-ID: <20161028055608.1736-3-ying.huang@intel.com> (raw) In-Reply-To: <20161028055608.1736-1-ying.huang@intel.com> From: Huang Ying <ying.huang@intel.com> This patch make it possible to charge or uncharge a set of continuous swap entries in the swap cgroup. The number of swap entries is specified via an added parameter. This will be used for the THP (Transparent Huge Page) swap support. Where a swap cluster backing a THP may be allocated and freed as a whole. So a set of (HPAGE_PMD_NR) continuous swap entries backing one THP need to be charged or uncharged together. This will batch the cgroup operations for the THP swap too. Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: Kirill A. Shutemov <kirill.shutemov@linux.intel.com> Cc: Vladimir Davydov <vdavydov@virtuozzo.com> Cc: Johannes Weiner <hannes@cmpxchg.org> Cc: Michal Hocko <mhocko@kernel.org> Cc: Tejun Heo <tj@kernel.org> Cc: cgroups@vger.kernel.org Signed-off-by: "Huang, Ying" <ying.huang@intel.com> --- include/linux/swap.h | 12 ++++++---- include/linux/swap_cgroup.h | 6 +++-- mm/memcontrol.c | 55 +++++++++++++++++++++++++-------------------- mm/shmem.c | 2 +- mm/swap_cgroup.c | 40 ++++++++++++++++++++++++--------- mm/swap_state.c | 2 +- mm/swapfile.c | 2 +- 7 files changed, 76 insertions(+), 43 deletions(-) diff --git a/include/linux/swap.h b/include/linux/swap.h index a56523c..001b506 100644 --- a/include/linux/swap.h +++ b/include/linux/swap.h @@ -552,8 +552,10 @@ static inline int mem_cgroup_swappiness(struct mem_cgroup *mem) #ifdef CONFIG_MEMCG_SWAP extern void mem_cgroup_swapout(struct page *page, swp_entry_t entry); -extern int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry); -extern void mem_cgroup_uncharge_swap(swp_entry_t entry); +extern int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry, + unsigned int nr_entries); +extern void mem_cgroup_uncharge_swap(swp_entry_t entry, + unsigned int nr_entries); extern long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg); extern bool mem_cgroup_swap_full(struct page *page); #else @@ -562,12 +564,14 @@ static inline void mem_cgroup_swapout(struct page *page, swp_entry_t entry) } static inline int mem_cgroup_try_charge_swap(struct page *page, - swp_entry_t entry) + swp_entry_t entry, + unsigned int nr_entries) { return 0; } -static inline void mem_cgroup_uncharge_swap(swp_entry_t entry) +static inline void mem_cgroup_uncharge_swap(swp_entry_t entry, + unsigned int nr_entries) { } diff --git a/include/linux/swap_cgroup.h b/include/linux/swap_cgroup.h index 145306b..b2b8ec7 100644 --- a/include/linux/swap_cgroup.h +++ b/include/linux/swap_cgroup.h @@ -7,7 +7,8 @@ extern unsigned short swap_cgroup_cmpxchg(swp_entry_t ent, unsigned short old, unsigned short new); -extern unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id); +extern unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id, + unsigned int nr_ents); extern unsigned short lookup_swap_cgroup_id(swp_entry_t ent); extern int swap_cgroup_swapon(int type, unsigned long max_pages); extern void swap_cgroup_swapoff(int type); @@ -15,7 +16,8 @@ extern void swap_cgroup_swapoff(int type); #else static inline -unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id) +unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id, + unsigned int nr_ents) { return 0; } diff --git a/mm/memcontrol.c b/mm/memcontrol.c index ae052b5..581f705 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -2383,10 +2383,9 @@ void mem_cgroup_split_huge_fixup(struct page *head) #ifdef CONFIG_MEMCG_SWAP static void mem_cgroup_swap_statistics(struct mem_cgroup *memcg, - bool charge) + int nr_entries) { - int val = (charge) ? 1 : -1; - this_cpu_add(memcg->stat->count[MEM_CGROUP_STAT_SWAP], val); + this_cpu_add(memcg->stat->count[MEM_CGROUP_STAT_SWAP], nr_entries); } /** @@ -2412,8 +2411,8 @@ static int mem_cgroup_move_swap_account(swp_entry_t entry, new_id = mem_cgroup_id(to); if (swap_cgroup_cmpxchg(entry, old_id, new_id) == old_id) { - mem_cgroup_swap_statistics(from, false); - mem_cgroup_swap_statistics(to, true); + mem_cgroup_swap_statistics(from, -1); + mem_cgroup_swap_statistics(to, 1); return 0; } return -EINVAL; @@ -5433,7 +5432,7 @@ void mem_cgroup_commit_charge(struct page *page, struct mem_cgroup *memcg, * let's not wait for it. The page already received a * memory+swap charge, drop the swap entry duplicate. */ - mem_cgroup_uncharge_swap(entry); + mem_cgroup_uncharge_swap(entry, nr_pages); } } @@ -5844,9 +5843,9 @@ void mem_cgroup_swapout(struct page *page, swp_entry_t entry) * ancestor for the swap instead and transfer the memory+swap charge. */ swap_memcg = mem_cgroup_id_get_online(memcg); - oldid = swap_cgroup_record(entry, mem_cgroup_id(swap_memcg)); + oldid = swap_cgroup_record(entry, mem_cgroup_id(swap_memcg), 1); VM_BUG_ON_PAGE(oldid, page); - mem_cgroup_swap_statistics(swap_memcg, true); + mem_cgroup_swap_statistics(swap_memcg, 1); page->mem_cgroup = NULL; @@ -5873,16 +5872,19 @@ void mem_cgroup_swapout(struct page *page, swp_entry_t entry) css_put(&memcg->css); } -/* - * mem_cgroup_try_charge_swap - try charging a swap entry +/** + * mem_cgroup_try_charge_swap - try charging a set of swap entries * @page: page being added to swap - * @entry: swap entry to charge + * @entry: the first swap entry to charge + * @nr_entries: the number of swap entries to charge * - * Try to charge @entry to the memcg that @page belongs to. + * Try to charge @nr_entries swap entries starting from @entry to the + * memcg that @page belongs to. * * Returns 0 on success, -ENOMEM on failure. */ -int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry) +int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry, + unsigned int nr_entries) { struct mem_cgroup *memcg; struct page_counter *counter; @@ -5900,25 +5902,29 @@ int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry) memcg = mem_cgroup_id_get_online(memcg); if (!mem_cgroup_is_root(memcg) && - !page_counter_try_charge(&memcg->swap, 1, &counter)) { + !page_counter_try_charge(&memcg->swap, nr_entries, &counter)) { mem_cgroup_id_put(memcg); return -ENOMEM; } - oldid = swap_cgroup_record(entry, mem_cgroup_id(memcg)); + if (nr_entries > 1) + mem_cgroup_id_get_many(memcg, nr_entries - 1); + oldid = swap_cgroup_record(entry, mem_cgroup_id(memcg), nr_entries); VM_BUG_ON_PAGE(oldid, page); - mem_cgroup_swap_statistics(memcg, true); + mem_cgroup_swap_statistics(memcg, nr_entries); return 0; } /** - * mem_cgroup_uncharge_swap - uncharge a swap entry - * @entry: swap entry to uncharge + * mem_cgroup_uncharge_swap - uncharge a set of swap entries + * @entry: the first swap entry to uncharge + * @nr_entries: the number of swap entries to uncharge * - * Drop the swap charge associated with @entry. + * Drop the swap charge associated with @nr_entries swap entries + * starting from @entry. */ -void mem_cgroup_uncharge_swap(swp_entry_t entry) +void mem_cgroup_uncharge_swap(swp_entry_t entry, unsigned int nr_entries) { struct mem_cgroup *memcg; unsigned short id; @@ -5926,17 +5932,18 @@ void mem_cgroup_uncharge_swap(swp_entry_t entry) if (!do_swap_account) return; - id = swap_cgroup_record(entry, 0); + id = swap_cgroup_record(entry, 0, nr_entries); rcu_read_lock(); memcg = mem_cgroup_from_id(id); if (memcg) { if (!mem_cgroup_is_root(memcg)) { if (cgroup_subsys_on_dfl(memory_cgrp_subsys)) - page_counter_uncharge(&memcg->swap, 1); + page_counter_uncharge(&memcg->swap, nr_entries); else - page_counter_uncharge(&memcg->memsw, 1); + page_counter_uncharge(&memcg->memsw, + nr_entries); } - mem_cgroup_swap_statistics(memcg, false); + mem_cgroup_swap_statistics(memcg, -nr_entries); mem_cgroup_id_put(memcg); } rcu_read_unlock(); diff --git a/mm/shmem.c b/mm/shmem.c index ad7813d..51a0fa9 100644 --- a/mm/shmem.c +++ b/mm/shmem.c @@ -1249,7 +1249,7 @@ static int shmem_writepage(struct page *page, struct writeback_control *wbc) if (!swap.val) goto redirty; - if (mem_cgroup_try_charge_swap(page, swap)) + if (mem_cgroup_try_charge_swap(page, swap, 1)) goto free_swap; /* diff --git a/mm/swap_cgroup.c b/mm/swap_cgroup.c index 310ac0b..8cee2d1 100644 --- a/mm/swap_cgroup.c +++ b/mm/swap_cgroup.c @@ -58,21 +58,27 @@ static int swap_cgroup_prepare(int type) return -ENOMEM; } +static struct swap_cgroup *__lookup_swap_cgroup(struct swap_cgroup_ctrl *ctrl, + pgoff_t offset) +{ + struct page *mappage; + struct swap_cgroup *sc; + + mappage = ctrl->map[offset / SC_PER_PAGE]; + sc = page_address(mappage); + return sc + offset % SC_PER_PAGE; +} + static struct swap_cgroup *lookup_swap_cgroup(swp_entry_t ent, struct swap_cgroup_ctrl **ctrlp) { pgoff_t offset = swp_offset(ent); struct swap_cgroup_ctrl *ctrl; - struct page *mappage; - struct swap_cgroup *sc; ctrl = &swap_cgroup_ctrl[swp_type(ent)]; if (ctrlp) *ctrlp = ctrl; - - mappage = ctrl->map[offset / SC_PER_PAGE]; - sc = page_address(mappage); - return sc + offset % SC_PER_PAGE; + return __lookup_swap_cgroup(ctrl, offset); } /** @@ -105,25 +111,39 @@ unsigned short swap_cgroup_cmpxchg(swp_entry_t ent, } /** - * swap_cgroup_record - record mem_cgroup for this swp_entry. - * @ent: swap entry to be recorded into + * swap_cgroup_record - record mem_cgroup for a set of swap entries + * @ent: the first swap entry to be recorded into * @id: mem_cgroup to be recorded + * @nr_ents: number of swap entries to be recorded * * Returns old value at success, 0 at failure. * (Of course, old value can be 0.) */ -unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id) +unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id, + unsigned int nr_ents) { struct swap_cgroup_ctrl *ctrl; struct swap_cgroup *sc; unsigned short old; unsigned long flags; + pgoff_t offset = swp_offset(ent); + pgoff_t end = offset + nr_ents; sc = lookup_swap_cgroup(ent, &ctrl); spin_lock_irqsave(&ctrl->lock, flags); old = sc->id; - sc->id = id; + for (;;) { + VM_BUG_ON(sc->id != old); + sc->id = id; + offset++; + if (offset == end) + break; + if (offset % SC_PER_PAGE) + sc++; + else + sc = __lookup_swap_cgroup(ctrl, offset); + } spin_unlock_irqrestore(&ctrl->lock, flags); return old; diff --git a/mm/swap_state.c b/mm/swap_state.c index 35d7e0e..d3f047b 100644 --- a/mm/swap_state.c +++ b/mm/swap_state.c @@ -172,7 +172,7 @@ int add_to_swap(struct page *page, struct list_head *list) if (!entry.val) return 0; - if (mem_cgroup_try_charge_swap(page, entry)) { + if (mem_cgroup_try_charge_swap(page, entry, 1)) { swapcache_free(entry); return 0; } diff --git a/mm/swapfile.c b/mm/swapfile.c index 18e247b..f3fc83f 100644 --- a/mm/swapfile.c +++ b/mm/swapfile.c @@ -806,7 +806,7 @@ static unsigned char swap_entry_free(struct swap_info_struct *p, /* free if no reference */ if (!usage) { - mem_cgroup_uncharge_swap(entry); + mem_cgroup_uncharge_swap(entry, 1); dec_cluster_info_page(p, p->cluster_info, offset); if (offset < p->lowest_bit) p->lowest_bit = offset; -- 2.9.3
WARNING: multiple messages have this Message-ID (diff)
From: "Huang, Ying" <ying.huang@intel.com> To: Andrew Morton <akpm@linux-foundation.org> Cc: tim.c.chen@intel.com, dave.hansen@intel.com, andi.kleen@intel.com, aaron.lu@intel.com, linux-mm@kvack.org, linux-kernel@vger.kernel.org, Huang Ying <ying.huang@intel.com>, Andrea Arcangeli <aarcange@redhat.com>, "Kirill A . Shutemov" <kirill.shutemov@linux.intel.com>, Vladimir Davydov <vdavydov@virtuozzo.com>, Johannes Weiner <hannes@cmpxchg.org>, Michal Hocko <mhocko@kernel.org>, Tejun Heo <tj@kernel.org>, cgroups@vger.kernel.org Subject: [PATCH -v4 RESEND 2/9] mm, memcg: Support to charge/uncharge multiple swap entries Date: Fri, 28 Oct 2016 13:56:01 +0800 [thread overview] Message-ID: <20161028055608.1736-3-ying.huang@intel.com> (raw) In-Reply-To: <20161028055608.1736-1-ying.huang@intel.com> From: Huang Ying <ying.huang@intel.com> This patch make it possible to charge or uncharge a set of continuous swap entries in the swap cgroup. The number of swap entries is specified via an added parameter. This will be used for the THP (Transparent Huge Page) swap support. Where a swap cluster backing a THP may be allocated and freed as a whole. So a set of (HPAGE_PMD_NR) continuous swap entries backing one THP need to be charged or uncharged together. This will batch the cgroup operations for the THP swap too. Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: Kirill A. Shutemov <kirill.shutemov@linux.intel.com> Cc: Vladimir Davydov <vdavydov@virtuozzo.com> Cc: Johannes Weiner <hannes@cmpxchg.org> Cc: Michal Hocko <mhocko@kernel.org> Cc: Tejun Heo <tj@kernel.org> Cc: cgroups@vger.kernel.org Signed-off-by: "Huang, Ying" <ying.huang@intel.com> --- include/linux/swap.h | 12 ++++++---- include/linux/swap_cgroup.h | 6 +++-- mm/memcontrol.c | 55 +++++++++++++++++++++++++-------------------- mm/shmem.c | 2 +- mm/swap_cgroup.c | 40 ++++++++++++++++++++++++--------- mm/swap_state.c | 2 +- mm/swapfile.c | 2 +- 7 files changed, 76 insertions(+), 43 deletions(-) diff --git a/include/linux/swap.h b/include/linux/swap.h index a56523c..001b506 100644 --- a/include/linux/swap.h +++ b/include/linux/swap.h @@ -552,8 +552,10 @@ static inline int mem_cgroup_swappiness(struct mem_cgroup *mem) #ifdef CONFIG_MEMCG_SWAP extern void mem_cgroup_swapout(struct page *page, swp_entry_t entry); -extern int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry); -extern void mem_cgroup_uncharge_swap(swp_entry_t entry); +extern int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry, + unsigned int nr_entries); +extern void mem_cgroup_uncharge_swap(swp_entry_t entry, + unsigned int nr_entries); extern long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg); extern bool mem_cgroup_swap_full(struct page *page); #else @@ -562,12 +564,14 @@ static inline void mem_cgroup_swapout(struct page *page, swp_entry_t entry) } static inline int mem_cgroup_try_charge_swap(struct page *page, - swp_entry_t entry) + swp_entry_t entry, + unsigned int nr_entries) { return 0; } -static inline void mem_cgroup_uncharge_swap(swp_entry_t entry) +static inline void mem_cgroup_uncharge_swap(swp_entry_t entry, + unsigned int nr_entries) { } diff --git a/include/linux/swap_cgroup.h b/include/linux/swap_cgroup.h index 145306b..b2b8ec7 100644 --- a/include/linux/swap_cgroup.h +++ b/include/linux/swap_cgroup.h @@ -7,7 +7,8 @@ extern unsigned short swap_cgroup_cmpxchg(swp_entry_t ent, unsigned short old, unsigned short new); -extern unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id); +extern unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id, + unsigned int nr_ents); extern unsigned short lookup_swap_cgroup_id(swp_entry_t ent); extern int swap_cgroup_swapon(int type, unsigned long max_pages); extern void swap_cgroup_swapoff(int type); @@ -15,7 +16,8 @@ extern void swap_cgroup_swapoff(int type); #else static inline -unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id) +unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id, + unsigned int nr_ents) { return 0; } diff --git a/mm/memcontrol.c b/mm/memcontrol.c index ae052b5..581f705 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -2383,10 +2383,9 @@ void mem_cgroup_split_huge_fixup(struct page *head) #ifdef CONFIG_MEMCG_SWAP static void mem_cgroup_swap_statistics(struct mem_cgroup *memcg, - bool charge) + int nr_entries) { - int val = (charge) ? 1 : -1; - this_cpu_add(memcg->stat->count[MEM_CGROUP_STAT_SWAP], val); + this_cpu_add(memcg->stat->count[MEM_CGROUP_STAT_SWAP], nr_entries); } /** @@ -2412,8 +2411,8 @@ static int mem_cgroup_move_swap_account(swp_entry_t entry, new_id = mem_cgroup_id(to); if (swap_cgroup_cmpxchg(entry, old_id, new_id) == old_id) { - mem_cgroup_swap_statistics(from, false); - mem_cgroup_swap_statistics(to, true); + mem_cgroup_swap_statistics(from, -1); + mem_cgroup_swap_statistics(to, 1); return 0; } return -EINVAL; @@ -5433,7 +5432,7 @@ void mem_cgroup_commit_charge(struct page *page, struct mem_cgroup *memcg, * let's not wait for it. The page already received a * memory+swap charge, drop the swap entry duplicate. */ - mem_cgroup_uncharge_swap(entry); + mem_cgroup_uncharge_swap(entry, nr_pages); } } @@ -5844,9 +5843,9 @@ void mem_cgroup_swapout(struct page *page, swp_entry_t entry) * ancestor for the swap instead and transfer the memory+swap charge. */ swap_memcg = mem_cgroup_id_get_online(memcg); - oldid = swap_cgroup_record(entry, mem_cgroup_id(swap_memcg)); + oldid = swap_cgroup_record(entry, mem_cgroup_id(swap_memcg), 1); VM_BUG_ON_PAGE(oldid, page); - mem_cgroup_swap_statistics(swap_memcg, true); + mem_cgroup_swap_statistics(swap_memcg, 1); page->mem_cgroup = NULL; @@ -5873,16 +5872,19 @@ void mem_cgroup_swapout(struct page *page, swp_entry_t entry) css_put(&memcg->css); } -/* - * mem_cgroup_try_charge_swap - try charging a swap entry +/** + * mem_cgroup_try_charge_swap - try charging a set of swap entries * @page: page being added to swap - * @entry: swap entry to charge + * @entry: the first swap entry to charge + * @nr_entries: the number of swap entries to charge * - * Try to charge @entry to the memcg that @page belongs to. + * Try to charge @nr_entries swap entries starting from @entry to the + * memcg that @page belongs to. * * Returns 0 on success, -ENOMEM on failure. */ -int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry) +int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry, + unsigned int nr_entries) { struct mem_cgroup *memcg; struct page_counter *counter; @@ -5900,25 +5902,29 @@ int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry) memcg = mem_cgroup_id_get_online(memcg); if (!mem_cgroup_is_root(memcg) && - !page_counter_try_charge(&memcg->swap, 1, &counter)) { + !page_counter_try_charge(&memcg->swap, nr_entries, &counter)) { mem_cgroup_id_put(memcg); return -ENOMEM; } - oldid = swap_cgroup_record(entry, mem_cgroup_id(memcg)); + if (nr_entries > 1) + mem_cgroup_id_get_many(memcg, nr_entries - 1); + oldid = swap_cgroup_record(entry, mem_cgroup_id(memcg), nr_entries); VM_BUG_ON_PAGE(oldid, page); - mem_cgroup_swap_statistics(memcg, true); + mem_cgroup_swap_statistics(memcg, nr_entries); return 0; } /** - * mem_cgroup_uncharge_swap - uncharge a swap entry - * @entry: swap entry to uncharge + * mem_cgroup_uncharge_swap - uncharge a set of swap entries + * @entry: the first swap entry to uncharge + * @nr_entries: the number of swap entries to uncharge * - * Drop the swap charge associated with @entry. + * Drop the swap charge associated with @nr_entries swap entries + * starting from @entry. */ -void mem_cgroup_uncharge_swap(swp_entry_t entry) +void mem_cgroup_uncharge_swap(swp_entry_t entry, unsigned int nr_entries) { struct mem_cgroup *memcg; unsigned short id; @@ -5926,17 +5932,18 @@ void mem_cgroup_uncharge_swap(swp_entry_t entry) if (!do_swap_account) return; - id = swap_cgroup_record(entry, 0); + id = swap_cgroup_record(entry, 0, nr_entries); rcu_read_lock(); memcg = mem_cgroup_from_id(id); if (memcg) { if (!mem_cgroup_is_root(memcg)) { if (cgroup_subsys_on_dfl(memory_cgrp_subsys)) - page_counter_uncharge(&memcg->swap, 1); + page_counter_uncharge(&memcg->swap, nr_entries); else - page_counter_uncharge(&memcg->memsw, 1); + page_counter_uncharge(&memcg->memsw, + nr_entries); } - mem_cgroup_swap_statistics(memcg, false); + mem_cgroup_swap_statistics(memcg, -nr_entries); mem_cgroup_id_put(memcg); } rcu_read_unlock(); diff --git a/mm/shmem.c b/mm/shmem.c index ad7813d..51a0fa9 100644 --- a/mm/shmem.c +++ b/mm/shmem.c @@ -1249,7 +1249,7 @@ static int shmem_writepage(struct page *page, struct writeback_control *wbc) if (!swap.val) goto redirty; - if (mem_cgroup_try_charge_swap(page, swap)) + if (mem_cgroup_try_charge_swap(page, swap, 1)) goto free_swap; /* diff --git a/mm/swap_cgroup.c b/mm/swap_cgroup.c index 310ac0b..8cee2d1 100644 --- a/mm/swap_cgroup.c +++ b/mm/swap_cgroup.c @@ -58,21 +58,27 @@ static int swap_cgroup_prepare(int type) return -ENOMEM; } +static struct swap_cgroup *__lookup_swap_cgroup(struct swap_cgroup_ctrl *ctrl, + pgoff_t offset) +{ + struct page *mappage; + struct swap_cgroup *sc; + + mappage = ctrl->map[offset / SC_PER_PAGE]; + sc = page_address(mappage); + return sc + offset % SC_PER_PAGE; +} + static struct swap_cgroup *lookup_swap_cgroup(swp_entry_t ent, struct swap_cgroup_ctrl **ctrlp) { pgoff_t offset = swp_offset(ent); struct swap_cgroup_ctrl *ctrl; - struct page *mappage; - struct swap_cgroup *sc; ctrl = &swap_cgroup_ctrl[swp_type(ent)]; if (ctrlp) *ctrlp = ctrl; - - mappage = ctrl->map[offset / SC_PER_PAGE]; - sc = page_address(mappage); - return sc + offset % SC_PER_PAGE; + return __lookup_swap_cgroup(ctrl, offset); } /** @@ -105,25 +111,39 @@ unsigned short swap_cgroup_cmpxchg(swp_entry_t ent, } /** - * swap_cgroup_record - record mem_cgroup for this swp_entry. - * @ent: swap entry to be recorded into + * swap_cgroup_record - record mem_cgroup for a set of swap entries + * @ent: the first swap entry to be recorded into * @id: mem_cgroup to be recorded + * @nr_ents: number of swap entries to be recorded * * Returns old value at success, 0 at failure. * (Of course, old value can be 0.) */ -unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id) +unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id, + unsigned int nr_ents) { struct swap_cgroup_ctrl *ctrl; struct swap_cgroup *sc; unsigned short old; unsigned long flags; + pgoff_t offset = swp_offset(ent); + pgoff_t end = offset + nr_ents; sc = lookup_swap_cgroup(ent, &ctrl); spin_lock_irqsave(&ctrl->lock, flags); old = sc->id; - sc->id = id; + for (;;) { + VM_BUG_ON(sc->id != old); + sc->id = id; + offset++; + if (offset == end) + break; + if (offset % SC_PER_PAGE) + sc++; + else + sc = __lookup_swap_cgroup(ctrl, offset); + } spin_unlock_irqrestore(&ctrl->lock, flags); return old; diff --git a/mm/swap_state.c b/mm/swap_state.c index 35d7e0e..d3f047b 100644 --- a/mm/swap_state.c +++ b/mm/swap_state.c @@ -172,7 +172,7 @@ int add_to_swap(struct page *page, struct list_head *list) if (!entry.val) return 0; - if (mem_cgroup_try_charge_swap(page, entry)) { + if (mem_cgroup_try_charge_swap(page, entry, 1)) { swapcache_free(entry); return 0; } diff --git a/mm/swapfile.c b/mm/swapfile.c index 18e247b..f3fc83f 100644 --- a/mm/swapfile.c +++ b/mm/swapfile.c @@ -806,7 +806,7 @@ static unsigned char swap_entry_free(struct swap_info_struct *p, /* free if no reference */ if (!usage) { - mem_cgroup_uncharge_swap(entry); + mem_cgroup_uncharge_swap(entry, 1); dec_cluster_info_page(p, p->cluster_info, offset); if (offset < p->lowest_bit) p->lowest_bit = offset; -- 2.9.3 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
next prev parent reply other threads:[~2016-10-28 6:01 UTC|newest] Thread overview: 28+ messages / expand[flat|nested] mbox.gz Atom feed top 2016-10-28 5:55 [PATCH -v4 RESEND 0/9] THP swap: Delay splitting THP during swapping out Huang, Ying 2016-10-28 5:55 ` Huang, Ying 2016-10-28 5:56 ` [PATCH -v4 RESEND 1/9] mm, swap: Make swap cluster size same of THP size on x86_64 Huang, Ying 2016-10-28 5:56 ` Huang, Ying 2016-10-28 5:56 ` Huang, Ying [this message] 2016-10-28 5:56 ` [PATCH -v4 RESEND 2/9] mm, memcg: Support to charge/uncharge multiple swap entries Huang, Ying 2016-10-28 5:56 ` [PATCH -v4 RESEND 3/9] mm, THP, swap: Add swap cluster allocate/free functions Huang, Ying 2016-10-28 5:56 ` Huang, Ying 2016-10-28 5:56 ` [PATCH -v4 RESEND 4/9] mm, THP, swap: Add get_huge_swap_page() Huang, Ying 2016-10-28 5:56 ` Huang, Ying 2016-10-28 5:56 ` [PATCH -v4 RESEND 5/9] mm, THP, swap: Support to clear SWAP_HAS_CACHE for huge page Huang, Ying 2016-10-28 5:56 ` Huang, Ying 2016-10-28 5:56 ` [PATCH -v4 RESEND 6/9] mm, THP, swap: Support to add/delete THP to/from swap cache Huang, Ying 2016-10-28 5:56 ` Huang, Ying 2016-10-28 8:08 ` Hillf Danton 2016-10-28 8:08 ` Hillf Danton 2016-10-28 8:30 ` Huang, Ying 2016-10-28 8:30 ` Huang, Ying 2016-10-28 5:56 ` [PATCH -v4 RESEND 7/9] mm, THP: Add can_split_huge_page() Huang, Ying 2016-10-28 5:56 ` Huang, Ying 2016-10-28 5:56 ` [PATCH -v4 RESEND 8/9] mm, THP, swap: Support to split THP in swap cache Huang, Ying 2016-10-28 5:56 ` Huang, Ying 2016-10-28 10:18 ` Hillf Danton 2016-10-28 10:18 ` Hillf Danton 2016-10-31 1:26 ` Huang, Ying 2016-10-31 1:26 ` Huang, Ying 2016-10-28 5:56 ` [PATCH -v4 RESEND 9/9] mm, THP, swap: Delay splitting THP during swap out Huang, Ying 2016-10-28 5:56 ` Huang, Ying
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=20161028055608.1736-3-ying.huang@intel.com \ --to=ying.huang@intel.com \ --cc=aarcange@redhat.com \ --cc=aaron.lu@intel.com \ --cc=akpm@linux-foundation.org \ --cc=andi.kleen@intel.com \ --cc=cgroups@vger.kernel.org \ --cc=dave.hansen@intel.com \ --cc=hannes@cmpxchg.org \ --cc=kirill.shutemov@linux.intel.com \ --cc=linux-kernel@vger.kernel.org \ --cc=linux-mm@kvack.org \ --cc=mhocko@kernel.org \ --cc=tim.c.chen@intel.com \ --cc=tj@kernel.org \ --cc=vdavydov@virtuozzo.com \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.