From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-2.8 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,SPF_PASS,URIBL_BLOCKED,USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 0EE4BC3279B for ; Sat, 7 Jul 2018 00:36:00 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id BE80D224A1 for ; Sat, 7 Jul 2018 00:35:59 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org BE80D224A1 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=intel.com Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S932867AbeGGAfy (ORCPT ); Fri, 6 Jul 2018 20:35:54 -0400 Received: from mga07.intel.com ([134.134.136.100]:20600 "EHLO mga07.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1753854AbeGGAf2 (ORCPT ); Fri, 6 Jul 2018 20:35:28 -0400 X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from orsmga005.jf.intel.com ([10.7.209.41]) by orsmga105.jf.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 06 Jul 2018 17:35:27 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.51,318,1526367600"; d="scan'208";a="238445481" Received: from linksys13920.jf.intel.com (HELO rpedgeco-HP-Z240-Tower-Workstation.jf.intel.com) ([10.7.197.56]) by orsmga005.jf.intel.com with ESMTP; 06 Jul 2018 17:35:27 -0700 From: Rick Edgecombe To: tglx@linutronix.de, mingo@redhat.com, hpa@zytor.com, x86@kernel.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, kernel-hardening@lists.openwall.com Cc: kristen@linux.intel.com, dave.hansen@intel.com, arjan@linux.intel.com, Rick Edgecombe Subject: [PATCH 1/3] vmalloc: Add __vmalloc_node_try_addr function Date: Fri, 6 Jul 2018 17:35:42 -0700 Message-Id: <1530923744-25687-2-git-send-email-rick.p.edgecombe@intel.com> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1530923744-25687-1-git-send-email-rick.p.edgecombe@intel.com> References: <1530923744-25687-1-git-send-email-rick.p.edgecombe@intel.com> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Create __vmalloc_node_try_addr function that tries to allocate at a specific address and supports caller specified behavior for whether any lazy purging happens if there is a collision. Signed-off-by: Rick Edgecombe --- include/linux/vmalloc.h | 3 + mm/vmalloc.c | 174 ++++++++++++++++++++++++++++++++++++++++++++++++ 2 files changed, 177 insertions(+) diff --git a/include/linux/vmalloc.h b/include/linux/vmalloc.h index 398e9c9..c7712c8 100644 --- a/include/linux/vmalloc.h +++ b/include/linux/vmalloc.h @@ -82,6 +82,9 @@ extern void *__vmalloc_node_range(unsigned long size, unsigned long align, unsigned long start, unsigned long end, gfp_t gfp_mask, pgprot_t prot, unsigned long vm_flags, int node, const void *caller); +extern void *__vmalloc_node_try_addr(unsigned long addr, unsigned long size, + gfp_t gfp_mask, pgprot_t prot, unsigned long vm_flags, + int node, int try_purge, const void *caller); #ifndef CONFIG_MMU extern void *__vmalloc_node_flags(unsigned long size, int node, gfp_t flags); static inline void *__vmalloc_node_flags_caller(unsigned long size, int node, diff --git a/mm/vmalloc.c b/mm/vmalloc.c index cfea25b..b6f2449 100644 --- a/mm/vmalloc.c +++ b/mm/vmalloc.c @@ -1710,6 +1710,180 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask, } /** + * __vmalloc_try_addr - try to alloc at a specific address + * @addr: address to try + * @size: size to try + * @gfp_mask: flags for the page level allocator + * @prot: protection mask for the allocated pages + * @vm_flags: additional vm area flags (e.g. %VM_NO_GUARD) + * @node: node to use for allocation or NUMA_NO_NODE + * @try_purge: try to purge if needed to fulfill and allocation + * @caller: caller's return address + * + * Try to allocate at the specific address. If it succeeds the address is + * returned. If it fails NULL is returned. If try_purge is zero, it will + * return an EBUSY ERR_PTR if it could have allocated if it was allowed to + * purge. It may trigger TLB flushes if a purge is needed, and try_purge is + * set. + */ +void *__vmalloc_node_try_addr(unsigned long addr, unsigned long size, + gfp_t gfp_mask, pgprot_t prot, unsigned long vm_flags, + int node, int try_purge, const void *caller) +{ + struct vmap_area *va; + struct vm_struct *area; + struct rb_node *n; + struct vmap_area *cur_va = NULL; + struct vmap_area *first_before = NULL; + + int not_at_end = 0; + int need_purge = 0; + int blocked = 0; + int purged = 0; + + unsigned long real_size = size; + unsigned long addr_end; + + size = PAGE_ALIGN(size); + if (!size || (size >> PAGE_SHIFT) > totalram_pages) + return NULL; + + WARN_ON(in_interrupt()); + + va = kmalloc_node(sizeof(struct vmap_area), + gfp_mask & GFP_RECLAIM_MASK, node); + if (unlikely(!va)) { + warn_alloc(gfp_mask, NULL, + "kmalloc: allocation failure"); + return NULL; + } + + area = kzalloc_node(sizeof(*area), gfp_mask & GFP_RECLAIM_MASK, node); + if (unlikely(!area)) { + warn_alloc(gfp_mask, NULL, + "kmalloc: allocation failure"); + goto failva; + } + /* + * Only scan the relevant parts containing pointers to other objects + * to avoid false negatives. + */ + kmemleak_scan_area(&va->rb_node, SIZE_MAX, gfp_mask & GFP_RECLAIM_MASK); + + if (!(vm_flags & VM_NO_GUARD)) + size += PAGE_SIZE; + + addr_end = addr + size; + if (addr > addr_end) + return NULL; + +retry: + spin_lock(&vmap_area_lock); + + n = vmap_area_root.rb_node; + while (n) { + cur_va = rb_entry(n, struct vmap_area, rb_node); + if (addr < cur_va->va_end) { + not_at_end = 1; + if (cur_va->va_start == addr) { + first_before = cur_va; + break; + } + n = n->rb_left; + } else { + first_before = cur_va; + n = n->rb_right; + } + } + + /* + * Linearly search through to make sure there is a hole, unless we are + * at the end of the VA list. + */ + if (not_at_end) { + /* + * If there is no VA that starts before the + * target address, start the check from the closest VA. + */ + if (first_before) + cur_va = first_before; + + while (cur_va->va_start < addr_end) { + if (cur_va->va_end > addr) { + if (cur_va->flags & VM_LAZY_FREE) { + need_purge = 1; + } else { + blocked = 1; + break; + } + } + + if (list_is_last(&cur_va->list, &vmap_area_list)) + break; + + cur_va = list_next_entry(cur_va, list); + } + + if (blocked || (!try_purge && need_purge)) { + /* + * If a non-lazy free va blocks the allocation, or + * we are not supposed to purge, but we need to the + * allocation fails. + */ + goto fail; + } + if (try_purge && need_purge) { + if (purged) { + /* if purged once before, give up */ + goto fail; + } else { + /* + * If the va blocking the allocation is set to + * be purged then purge all vmap_areas that are + * set to purged since this will flush the TLBs + * anyway. + */ + spin_unlock(&vmap_area_lock); + purge_vmap_area_lazy(); + need_purge = 0; + purged = 1; + goto retry; + } + } + } + + va->va_start = addr; + va->va_end = addr_end; + va->flags = 0; + __insert_vmap_area(va); + + spin_unlock(&vmap_area_lock); + + setup_vmalloc_vm(area, va, vm_flags, caller); + + addr = (unsigned long)__vmalloc_area_node(area, gfp_mask, prot, node); + if (!addr) { + warn_alloc(gfp_mask, NULL, + "vmalloc: allocation failure: %lu bytes", real_size); + return NULL; + } + + clear_vm_uninitialized_flag(area); + + kmemleak_vmalloc(area, size, gfp_mask); + + return (void *)addr; +fail: + kfree(area); + spin_unlock(&vmap_area_lock); +failva: + kfree(va); + if (need_purge && !blocked) + return ERR_PTR(-EBUSY); + return NULL; +} + +/** * __vmalloc_node_range - allocate virtually contiguous memory * @size: allocation size * @align: desired alignment -- 2.7.4