LKML Archive on lore.kernel.org
 help / color / Atom feed
From: Pankaj Gupta <pagupta@redhat.com>
To: Barret Rhoden <brho@google.com>
Cc: "Dan Williams" <dan.j.williams@intel.com>,
	"David Hildenbrand" <david@redhat.com>,
	"Dave Jiang" <dave.jiang@intel.com>,
	"Ross Zwisler" <zwisler@kernel.org>,
	"Vishal Verma" <vishal.l.verma@intel.com>,
	"Paolo Bonzini" <pbonzini@redhat.com>,
	"Radim Krčmář" <rkrcmar@redhat.com>,
	"Thomas Gleixner" <tglx@linutronix.de>,
	"Ingo Molnar" <mingo@redhat.com>,
	"Borislav Petkov" <bp@alien8.de>,
	linux-nvdimm@lists.01.org, linux-kernel@vger.kernel.org,
	"H. Peter Anvin" <hpa@zytor.com>,
	x86@kernel.org, kvm@vger.kernel.org,
	"yu c zhang" <yu.c.zhang@intel.com>,
	"yi z zhang" <yi.z.zhang@intel.com>
Subject: Re: [PATCH 2/2] kvm: Use huge pages for DAX-backed files
Date: Tue, 13 Nov 2018 05:02:33 -0500 (EST)
Message-ID: <286665658.33247363.1542103353780.JavaMail.zimbra@redhat.com> (raw)
In-Reply-To: <043a592d-6592-3053-15a0-68cc54a26deb@redhat.com>


> 
> On 09.11.18 21:39, Barret Rhoden wrote:
> > This change allows KVM to map DAX-backed files made of huge pages with
> > huge mappings in the EPT/TDP.
> > 
> > DAX pages are not PageTransCompound.  The existing check is trying to
> > determine if the mapping for the pfn is a huge mapping or not.  For
> > non-DAX maps, e.g. hugetlbfs, that means checking PageTransCompound.
> > For DAX, we can check the page table itself.
> > 
> > Note that KVM already faulted in the page (or huge page) in the host's
> > page table, and we hold the KVM mmu spinlock (grabbed before checking
> > the mmu seq).
> 
> I wonder if the KVM mmu spinlock is enough for walking (not KVM
> exclusive) host page tables. Can you elaborate?

As this patch is dependent on PageReserved patch(which is in progress), just 
wondering if we are able to test the code path for hugepage with DAX.

Thanks,
Pankaj 
 
> 
> > 
> > Signed-off-by: Barret Rhoden <brho@google.com>
> > ---
> >  arch/x86/kvm/mmu.c | 34 ++++++++++++++++++++++++++++++++--
> >  1 file changed, 32 insertions(+), 2 deletions(-)
> > 
> > diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c
> > index cf5f572f2305..2df8c459dc6a 100644
> > --- a/arch/x86/kvm/mmu.c
> > +++ b/arch/x86/kvm/mmu.c
> > @@ -3152,6 +3152,36 @@ static int kvm_handle_bad_page(struct kvm_vcpu
> > *vcpu, gfn_t gfn, kvm_pfn_t pfn)
> >  	return -EFAULT;
> >  }
> >  
> > +static bool pfn_is_huge_mapped(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn)
> > +{
> > +	struct page *page = pfn_to_page(pfn);
> > +	unsigned long hva, map_shift;
> > +
> > +	if (!is_zone_device_page(page))
> > +		return PageTransCompoundMap(page);
> > +
> > +	/*
> > +	 * DAX pages do not use compound pages.  The page should have already
> > +	 * been mapped into the host-side page table during try_async_pf(), so
> > +	 * we can check the page tables directly.
> > +	 */
> > +	hva = gfn_to_hva(kvm, gfn);
> > +	if (kvm_is_error_hva(hva))
> > +		return false;
> > +
> > +	/*
> > +	 * Our caller grabbed the KVM mmu_lock with a successful
> > +	 * mmu_notifier_retry, so we're safe to walk the page table.
> > +	 */
> > +	map_shift = dev_pagemap_mapping_shift(hva, current->mm);
> 
> You could get rid of that local variable map_shift.
> 
> > +	switch (map_shift) {
> > +	case PMD_SHIFT:
> > +	case PUD_SIZE:
> > +		return true;
> > +	}
> > +	return false;
> > +}
> > +
> >  static void transparent_hugepage_adjust(struct kvm_vcpu *vcpu,
> >  					gfn_t *gfnp, kvm_pfn_t *pfnp,
> >  					int *levelp)
> > @@ -3168,7 +3198,7 @@ static void transparent_hugepage_adjust(struct
> > kvm_vcpu *vcpu,
> >  	 */
> >  	if (!is_error_noslot_pfn(pfn) && !kvm_is_reserved_pfn(pfn) &&
> >  	    level == PT_PAGE_TABLE_LEVEL &&
> > -	    PageTransCompoundMap(pfn_to_page(pfn)) &&
> > +	    pfn_is_huge_mapped(vcpu->kvm, gfn, pfn) &&
> >  	    !mmu_gfn_lpage_is_disallowed(vcpu, gfn, PT_DIRECTORY_LEVEL)) {
> >  		unsigned long mask;
> >  		/*
> > @@ -5678,7 +5708,7 @@ static bool kvm_mmu_zap_collapsible_spte(struct kvm
> > *kvm,
> >  		 */
> >  		if (sp->role.direct &&
> >  			!kvm_is_reserved_pfn(pfn) &&
> > -			PageTransCompoundMap(pfn_to_page(pfn))) {
> > +			pfn_is_huge_mapped(kvm, sp->gfn, pfn)) {
> >  			pte_list_remove(rmap_head, sptep);
> >  			need_tlb_flush = 1;
> >  			goto restart;
> > 
> 
> This looks surprisingly simple to me :)
> 
> --
> 
> Thanks,
> 
> David / dhildenb
> 

  reply index

Thread overview: 23+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2018-11-09 20:39 [PATCH 0/2] " Barret Rhoden
2018-11-09 20:39 ` [PATCH 1/2] mm: make dev_pagemap_mapping_shift() externally visible Barret Rhoden
2018-11-13  9:28   ` David Hildenbrand
2018-11-09 20:39 ` [PATCH 2/2] kvm: Use huge pages for DAX-backed files Barret Rhoden
2018-11-12 19:31   ` Paolo Bonzini
2018-11-13 16:21     ` Barret Rhoden
2018-11-13 18:22       ` Paolo Bonzini
2018-11-13  9:36   ` David Hildenbrand
2018-11-13 10:02     ` Pankaj Gupta [this message]
2018-11-13 12:41       ` Paolo Bonzini
2018-11-13 15:56       ` Barret Rhoden
2018-11-14  9:09         ` Pankaj Gupta
2018-11-13 15:50     ` Barret Rhoden
2018-11-14 21:51 ` [PATCH v2 0/3] " Barret Rhoden
2018-11-14 21:51   ` [PATCH v2 1/3] mm: make dev_pagemap_mapping_shift() externally visible Barret Rhoden
2018-11-26 16:50     ` Paolo Bonzini
2018-11-26 18:32     ` Dan Williams
2018-11-14 21:51   ` [PATCH v2 2/3] kvm: Use huge pages for DAX-backed files Barret Rhoden
2018-11-14 21:51   ` [PATCH v2 3/3] kvm: remove redundant PageReserved() check Barret Rhoden
2018-11-27 13:31     ` David Hildenbrand
2018-11-15  0:55   ` [PATCH v2 0/3] kvm: Use huge pages for DAX-backed files Dan Williams
2018-12-03 17:40     ` Barret Rhoden
2018-12-03 18:32       ` Alexander Duyck

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=286665658.33247363.1542103353780.JavaMail.zimbra@redhat.com \
    --to=pagupta@redhat.com \
    --cc=bp@alien8.de \
    --cc=brho@google.com \
    --cc=dan.j.williams@intel.com \
    --cc=dave.jiang@intel.com \
    --cc=david@redhat.com \
    --cc=hpa@zytor.com \
    --cc=kvm@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-nvdimm@lists.01.org \
    --cc=mingo@redhat.com \
    --cc=pbonzini@redhat.com \
    --cc=rkrcmar@redhat.com \
    --cc=tglx@linutronix.de \
    --cc=vishal.l.verma@intel.com \
    --cc=x86@kernel.org \
    --cc=yi.z.zhang@intel.com \
    --cc=yu.c.zhang@intel.com \
    --cc=zwisler@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link

LKML Archive on lore.kernel.org

Archives are clonable:
	git clone --mirror https://lore.kernel.org/lkml/0 lkml/git/0.git
	git clone --mirror https://lore.kernel.org/lkml/1 lkml/git/1.git
	git clone --mirror https://lore.kernel.org/lkml/2 lkml/git/2.git
	git clone --mirror https://lore.kernel.org/lkml/3 lkml/git/3.git
	git clone --mirror https://lore.kernel.org/lkml/4 lkml/git/4.git
	git clone --mirror https://lore.kernel.org/lkml/5 lkml/git/5.git
	git clone --mirror https://lore.kernel.org/lkml/6 lkml/git/6.git
	git clone --mirror https://lore.kernel.org/lkml/7 lkml/git/7.git
	git clone --mirror https://lore.kernel.org/lkml/8 lkml/git/8.git

	# If you have public-inbox 1.1+ installed, you may
	# initialize and index your mirror using the following commands:
	public-inbox-init -V2 lkml lkml/ https://lore.kernel.org/lkml \
		linux-kernel@vger.kernel.org
	public-inbox-index lkml

Example config snippet for mirrors

Newsgroup available over NNTP:
	nntp://nntp.lore.kernel.org/org.kernel.vger.linux-kernel


AGPL code for this site: git clone https://public-inbox.org/public-inbox.git