From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1755410AbcKPDMl (ORCPT ); Tue, 15 Nov 2016 22:12:41 -0500 Received: from mga14.intel.com ([192.55.52.115]:25366 "EHLO mga14.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S932373AbcKPDMe (ORCPT ); Tue, 15 Nov 2016 22:12:34 -0500 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.31,497,1473145200"; d="scan'208";a="31391296" From: "Huang, Ying" To: Andrew Morton Cc: tim.c.chen@intel.com, dave.hansen@intel.com, andi.kleen@intel.com, aaron.lu@intel.com, linux-mm@kvack.org, linux-kernel@vger.kernel.org, Huang Ying , Andrea Arcangeli , "Kirill A . Shutemov" , Hugh Dickins , Shaohua Li , Minchan Kim , Rik van Riel Subject: [PATCH -v5 5/9] mm, THP, swap: Support to clear SWAP_HAS_CACHE for huge page Date: Wed, 16 Nov 2016 11:10:53 +0800 Message-Id: <20161116031057.12977-6-ying.huang@intel.com> X-Mailer: git-send-email 2.10.2 In-Reply-To: <20161116031057.12977-1-ying.huang@intel.com> References: <20161116031057.12977-1-ying.huang@intel.com> Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org From: Huang Ying __swapcache_free() is added to support to clear the SWAP_HAS_CACHE flag for the huge page. This will free the specified swap cluster now. Because now this function will be called only in the error path to free the swap cluster just allocated. So the corresponding swap_map[i] == SWAP_HAS_CACHE, that is, the swap count is 0. This makes the implementation simpler than that of the ordinary swap entry. This will be used for delaying splitting THP (Transparent Huge Page) during swapping out. Where for one THP to swap out, we will allocate a swap cluster, add the THP into the swap cache, then split the THP. If anything fails after allocating the swap cluster and before splitting the THP successfully, the swapcache_free_trans_huge() will be used to free the swap space allocated. Cc: Andrea Arcangeli Cc: Kirill A. Shutemov Cc: Hugh Dickins Cc: Shaohua Li Cc: Minchan Kim Cc: Rik van Riel Signed-off-by: "Huang, Ying" --- include/linux/swap.h | 9 +++++++-- mm/swapfile.c | 33 +++++++++++++++++++++++++++++++-- 2 files changed, 38 insertions(+), 4 deletions(-) diff --git a/include/linux/swap.h b/include/linux/swap.h index 1df1e23..cd1dc5c 100644 --- a/include/linux/swap.h +++ b/include/linux/swap.h @@ -411,7 +411,7 @@ extern void swap_shmem_alloc(swp_entry_t); extern int swap_duplicate(swp_entry_t); extern int swapcache_prepare(swp_entry_t); extern void swap_free(swp_entry_t); -extern void swapcache_free(swp_entry_t); +extern void __swapcache_free(swp_entry_t, bool); extern int free_swap_and_cache(swp_entry_t); extern int swap_type_of(dev_t, sector_t, struct block_device **); extern unsigned int count_swap_pages(int, int); @@ -483,7 +483,7 @@ static inline void swap_free(swp_entry_t swp) { } -static inline void swapcache_free(swp_entry_t swp) +static inline void __swapcache_free(swp_entry_t swp, bool huge) { } @@ -554,6 +554,11 @@ static inline swp_entry_t get_huge_swap_page(void) #endif /* CONFIG_SWAP */ +static inline void swapcache_free(swp_entry_t entry) +{ + __swapcache_free(entry, false); +} + #ifdef CONFIG_MEMCG static inline int mem_cgroup_swappiness(struct mem_cgroup *memcg) { diff --git a/mm/swapfile.c b/mm/swapfile.c index 6d9dffb..e8d64ef 100644 --- a/mm/swapfile.c +++ b/mm/swapfile.c @@ -732,6 +732,27 @@ static void swap_free_huge_cluster(struct swap_info_struct *si, __swap_entry_free(si, offset, true); } +/* + * Caller should hold si->lock. + */ +static void swapcache_free_trans_huge(struct swap_info_struct *si, + swp_entry_t entry) +{ + unsigned long offset = swp_offset(entry); + unsigned long idx = offset / SWAPFILE_CLUSTER; + unsigned char *map; + unsigned int i; + + map = si->swap_map + offset; + for (i = 0; i < SWAPFILE_CLUSTER; i++) { + VM_BUG_ON(map[i] != SWAP_HAS_CACHE); + map[i] &= ~SWAP_HAS_CACHE; + } + /* Cluster size is same as huge page size */ + mem_cgroup_uncharge_swap(entry, HPAGE_PMD_NR); + swap_free_huge_cluster(si, idx); +} + static unsigned long swap_alloc_huge_cluster(struct swap_info_struct *si) { unsigned long idx; @@ -758,6 +779,11 @@ static inline unsigned long swap_alloc_huge_cluster(struct swap_info_struct *si) { return 0; } + +static inline void swapcache_free_trans_huge(struct swap_info_struct *si, + swp_entry_t entry) +{ +} #endif swp_entry_t __get_swap_page(bool huge) @@ -949,13 +975,16 @@ void swap_free(swp_entry_t entry) /* * Called after dropping swapcache to decrease refcnt to swap entries. */ -void swapcache_free(swp_entry_t entry) +void __swapcache_free(swp_entry_t entry, bool huge) { struct swap_info_struct *p; p = swap_info_get(entry); if (p) { - swap_entry_free(p, entry, SWAP_HAS_CACHE); + if (unlikely(huge)) + swapcache_free_trans_huge(p, entry); + else + swap_entry_free(p, entry, SWAP_HAS_CACHE); spin_unlock(&p->lock); } } -- 2.10.2