nvdimm.lists.linux.dev archive mirror
 help / color / mirror / Atom feed
From: Alexander Duyck <alexander.h.duyck@linux.intel.com>
To: Dan Williams <dan.j.williams@intel.com>
Cc: Linux Kernel Mailing List <linux-kernel@vger.kernel.org>,
	linux-nvdimm <linux-nvdimm@lists.01.org>
Subject: Re: [PATCH] libnvdimm, dimm: Maximize label transfer size
Date: Mon, 1 Oct 2018 08:41:08 -0700	[thread overview]
Message-ID: <9f24ed83-5fea-da70-5470-36992b90c521@linux.intel.com> (raw)
In-Reply-To: <CAPcyv4hi7XLAswcoWEKOcQq_0FCTybxTUd_4H+mJj0S7BkTCjg@mail.gmail.com>

On 9/28/2018 4:03 PM, Dan Williams wrote:
> On Fri, Sep 28, 2018 at 3:44 PM Alexander Duyck
> <alexander.h.duyck@linux.intel.com> wrote:
>>
>> On 9/28/2018 11:56 AM, Dan Williams wrote:
>>> Use kvzalloc() to bypass the arbitrary PAGE_SIZE limit of label transfer
>>> operations. Given the expense of calling into firmware, maximize the
>>> amount of label data we transfer per call to be up to the total label
>>> space if allowed by the firmware.
>>>
>>> Cc: Alexander Duyck <alexander.h.duyck@linux.intel.com>
>>> Signed-off-by: Dan Williams <dan.j.williams@intel.com>
>>
>> This looks mostly fine to me.
>>
>> The only thing I am not certain about is if we want to use the raw
>> max_cmd_size value or not since there is no guarantee it is aligned to
>> any sort of boundary. It might be beneficial to either cache line align
>> it or align it to 256B if it is larger than that since that way we at
>> least would be getting full labels instead of partial ones.
> 
> If we leave any portion of the max transfer size unused it just means
> we're that much more likely to need an additional call into the ASL
> interpreter. That said It might be worthwhile capping the max transfer
> size in case the BIOS specifies something extreme (> 1MB) in the
> future.

Actually it all depends on the size of the max transfer versus the size 
of the label area. So for example if the label area is 128K and the max 
transfer size is 64K minus 32 then it doesn't really make much 
difference until we get below 44K for our max read size.

My main concern is that I don't want to waste resources on allocating 
memory that isn't needed. For smaller allocations I agree we need to use 
everything we have to minimize calls to the ASL interpreter, but when 
max transfer sizes start getting large we are just allocating multiple 
pages of memory that do not need to be used.

If nothing else I can look at doing the logic to reduce our memory 
allocations as a set of follow-up patches.

- Alex
_______________________________________________
Linux-nvdimm mailing list
Linux-nvdimm@lists.01.org
https://lists.01.org/mailman/listinfo/linux-nvdimm

  reply	other threads:[~2018-10-01 15:41 UTC|newest]

Thread overview: 5+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2018-09-28 18:56 Dan Williams
2018-09-28 22:44 ` Alexander Duyck
2018-09-28 23:03   ` Dan Williams
2018-10-01 15:41     ` Alexander Duyck [this message]
2018-10-01 16:55       ` Dan Williams

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=9f24ed83-5fea-da70-5470-36992b90c521@linux.intel.com \
    --to=alexander.h.duyck@linux.intel.com \
    --cc=dan.j.williams@intel.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-nvdimm@lists.01.org \
    --subject='Re: [PATCH] libnvdimm, dimm: Maximize label transfer size' \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).