linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: David Hildenbrand <david@redhat.com>
To: linux-kernel@vger.kernel.org, Michal Hocko <mhocko@kernel.org>
Cc: linux-mm@kvack.org, virtio-dev@lists.oasis-open.org,
	virtualization@lists.linux-foundation.org, kvm@vger.kernel.org,
	Andrew Morton <akpm@linux-foundation.org>,
	"Michael S . Tsirkin" <mst@redhat.com>,
	Sebastien Boeuf <sebastien.boeuf@intel.com>,
	Samuel Ortiz <samuel.ortiz@intel.com>,
	Robert Bradford <robert.bradford@intel.com>,
	Luiz Capitulino <lcapitulino@redhat.com>,
	Alexander Duyck <alexander.h.duyck@linux.intel.com>,
	Alexander Potapenko <glider@google.com>,
	Alexander Viro <viro@zeniv.linux.org.uk>,
	Anshuman Khandual <anshuman.khandual@arm.com>,
	Anthony Yznaga <anthony.yznaga@oracle.com>,
	Dan Williams <dan.j.williams@intel.com>,
	Dave Young <dyoung@redhat.com>,
	Igor Mammedov <imammedo@redhat.com>,
	Jason Wang <jasowang@redhat.com>,
	Johannes Weiner <hannes@cmpxchg.org>,
	Juergen Gross <jgross@suse.com>,
	Konrad Rzeszutek Wilk <konrad.wilk@oracle.com>,
	Len Brown <lenb@kernel.org>,
	Mel Gorman <mgorman@techsingularity.net>,
	Michal Hocko <mhocko@suse.com>,
	Mike Rapoport <rppt@linux.ibm.com>,
	Oscar Salvador <osalvador@suse.com>,
	Oscar Salvador <osalvador@suse.de>,
	Pavel Tatashin <pasha.tatashin@soleen.com>,
	Pavel Tatashin <pavel.tatashin@microsoft.com>,
	Pingfan Liu <kernelfans@gmail.com>, Qian Cai <cai@lca.pw>,
	"Rafael J. Wysocki" <rjw@rjwysocki.net>,
	Stefan Hajnoczi <stefanha@redhat.com>,
	Vlastimil Babka <vbabka@suse.cz>,
	Wei Yang <richard.weiyang@gmail.com>
Subject: Re: [PATCH RFC v4 00/13] virtio-mem: paravirtualized memory
Date: Tue, 25 Feb 2020 10:58:36 +0100	[thread overview]
Message-ID: <3468efe4-0878-3d22-4f74-3092acfd0c47@redhat.com> (raw)
In-Reply-To: <d07c6b4a-46e3-3f81-04db-ea1475fe48c2@redhat.com>

On 29.01.20 10:41, David Hildenbrand wrote:
> On 09.01.20 14:48, David Hildenbrand wrote:
>> On 12.12.19 18:11, David Hildenbrand wrote:
>>> This series is based on latest linux-next. The patches are located at:
>>>     https://github.com/davidhildenbrand/linux.git virtio-mem-rfc-v4
>>>
>>> The basic idea of virtio-mem is to provide a flexible,
>>> cross-architecture memory hot(un)plug solution that avoids many limitations
>>> imposed by existing technologies, architectures, and interfaces. More
>>> details can be found below and in linked material.
>>>
>>> This RFC is limited to x86-64, however, should theoretically work on any
>>> architecture that supports virtio and implements memory hot(un)plug under
>>> Linux - like s390x, powerpc64 and arm64. On x86-64, it is currently
>>> possible to add/remove memory to the system in >= 4MB granularity.
>>> Memory hotplug works very reliably. For memory unplug, there are no
>>> guarantees how much memory can actually get unplugged, it depends on the
>>> setup (especially: fragmentation of (unmovable) memory). I have plans to
>>> improve that in the future.
>>>
>>> --------------------------------------------------------------------------
>>> 1. virtio-mem
>>> --------------------------------------------------------------------------
>>>
>>> The basic idea behind virtio-mem was presented at KVM Forum 2018. The
>>> slides can be found at [1]. The previous RFC can be found at [2]. The
>>> first RFC can be found at [3]. However, the concept evolved over time. The
>>> KVM Forum slides roughly match the current design.
>>>
>>> Patch #2 ("virtio-mem: Paravirtualized memory hotplug") contains quite some
>>> information, especially in "include/uapi/linux/virtio_mem.h":
>>>
>>>   Each virtio-mem device manages a dedicated region in physical address
>>>   space. Each device can belong to a single NUMA node, multiple devices
>>>   for a single NUMA node are possible. A virtio-mem device is like a
>>>   "resizable DIMM" consisting of small memory blocks that can be plugged
>>>   or unplugged. The device driver is responsible for (un)plugging memory
>>>   blocks on demand.
>>>
>>>   Virtio-mem devices can only operate on their assigned memory region in
>>>   order to (un)plug memory. A device cannot (un)plug memory belonging to
>>>   other devices.
>>>
>>>   The "region_size" corresponds to the maximum amount of memory that can
>>>   be provided by a device. The "size" corresponds to the amount of memory
>>>   that is currently plugged. "requested_size" corresponds to a request
>>>   from the device to the device driver to (un)plug blocks. The
>>>   device driver should try to (un)plug blocks in order to reach the
>>>   "requested_size". It is impossible to plug more memory than requested.
>>>
>>>   The "usable_region_size" represents the memory region that can actually
>>>   be used to (un)plug memory. It is always at least as big as the
>>>   "requested_size" and will grow dynamically. It will only shrink when
>>>   explicitly triggered (VIRTIO_MEM_REQ_UNPLUG).
>>>
>>>   Memory in the usable region can usually be read, however, there are no
>>>   guarantees. It can happen that the device cannot process a request,
>>>   because it is busy. The device driver has to retry later.
>>>
>>>   Usually, during system resets all memory will get unplugged, so the
>>>   device driver can start with a clean state. However, in specific
>>>   scenarios (if the device is busy) it can happen that the device still
>>>   has memory plugged. The device driver can request to unplug all memory
>>>   (VIRTIO_MEM_REQ_UNPLUG) - which might take a while to succeed if the
>>>   device is busy.
>>>
>>> --------------------------------------------------------------------------
>>> 2. Linux Implementation
>>> --------------------------------------------------------------------------
>>>
>>> This RFC reuses quite some existing MM infrastructure, however, has to
>>> expose some additional functionality.
>>>
>>> Memory blocks (e.g., 128MB) are added/removed on demand. Within these
>>> memory blocks, subblocks (e.g., 4MB) are plugged/unplugged. The sizes
>>> depend on the target architecture, MAX_ORDER + pageblock_order, and
>>> the block size of a virtio-mem device.
>>>
>>> add_memory()/try_remove_memory() is used to add/remove memory blocks.
>>> virtio-mem will not online memory blocks itself. This has to be done by
>>> user space, or configured into the kernel
>>> (CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE). virtio-mem will only unplug memory
>>> that was online to the ZONE_NORMAL. Memory is suggested to be onlined to
>>> the ZONE_NORMAL for now.
>>>
>>> The memory hotplug notifier is used to properly synchronize against
>>> onlining/offlining of memory blocks and to track the states of memory
>>> blocks (including the zone memory blocks are onlined to).
>>>
>>> The set_online_page() callback is used to keep unplugged subblocks
>>> of a memory block fake-offline when onlining the memory block.
>>> generic_online_page() is used to fake-online plugged subblocks. This
>>> handling is similar to the Hyper-V balloon driver.
>>>
>>> PG_offline is used to mark unplugged subblocks as offline, so e.g.,
>>> dumping tools (makedumpfile) will skip these pages. This is similar to
>>> other balloon drivers like virtio-balloon and Hyper-V.
>>>
>>> Memory offlining code is extended to allow drivers to drop their reference
>>> to PG_offline pages when MEM_GOING_OFFLINE, so these pages can be skipped
>>> when offlining memory blocks. This allows to offline memory blocks that
>>> have partially unplugged (allocated e.g., via alloc_contig_range())
>>> subblocks - or are completely unplugged.
>>>
>>> alloc_contig_range()/free_contig_range() [now exposed] is used to
>>> unplug/plug subblocks of memory blocks the are already exposed to Linux.
>>>
>>> offline_and_remove_memory() [new] is used to offline a fully unplugged
>>> memory block and remove it from Linux.
>>>
>>>
>>> A lot of additional information can be found in the separate patches and
>>> as comments in the code itself.
>>>
>>> --------------------------------------------------------------------------
>>> 3. Changes RFC v2 -> v3
>>> --------------------------------------------------------------------------
>>>
>>> A lot of things changed, especially also on the QEMU + virtio side. The
>>> biggest changes on the Linux driver side are:
>>> - Onlining/offlining of subblocks is now emulated on top of memory blocks.
>>>   set_online_page()+alloc_contig_range()+free_contig_range() is now used
>>>   for that. Core MM does not have to be modified and will continue to
>>>   online/offline full memory blocks.
>>> - Onlining/offlining of memory blocks is no longer performed by virtio-mem.
>>> - Pg_offline is upstream and can be used. It is also used to allow
>>>   offlining of partially unplugged memory blocks.
>>> - Memory block states + subblocks are now tracked more space-efficient.
>>> - Proper kexec(), kdump(), driver unload, driver reload, ZONE_MOVABLE, ...
>>>   handling.
>>>
>>> --------------------------------------------------------------------------
>>> 4. Changes RFC v3 -> RFC v4
>>> --------------------------------------------------------------------------
>>>
>>> Only minor things changed, especially nothing on the QEMU + virtio side.
>>> Interresting changes on the Linux driver side are:
>>> - "mm: Allow to offline unmovable PageOffline() pages via
>>>    MEM_GOING_OFFLINE"
>>> -- Rework to Michals suggestion (allow to isolate all PageOffline() pages
>>>    by skipping all PageOffline() pages in has_unmovable_pages(). Fail
>>>    offlining later if the pages cannot be offlined/migrated).
>>> - "virtio-mem: Allow to offline partially unplugged memory blocks"
>>> -- Adapt to Michals suggestion on core-mm part.
>>> - "virtio-mem: Better retry handling"
>>> -- Optimize retry intervals
>>> - "virtio-mem: Drop slab objects when unplug continues to fail"
>>> -- Call drop_slab()/drop_slab_node() when unplug keeps failing for a longer
>>>    time.
>>> - Multiple cleanups and fixes.
>>>
>>> --------------------------------------------------------------------------
>>> 5. Future work
>>> --------------------------------------------------------------------------
>>>
>>> The separate patches contain a lot of future work items. One of the next
>>> steps is to make memory unplug more likely to succeed - currently, there
>>> are no guarantees on how much memory can get unplugged again. I have
>>> various ideas on how to limit fragmentation of all memory blocks that
>>> virtio-mem added.
>>>
>>> Memory hotplug:
>>> - Reduce the amount of memory resources if that turnes out to be an
>>>   issue. Or try to speed up relevant code paths to deal with many
>>>   resources.
>>> - Allocate the vmemmap from the added memory. Makes hotplug more likely
>>>   to succeed, the vmemmap is stored on the same NUMA node and that
>>>   unmovable memory will later not hinder unplug.
>>>
>>> Memory hotunplug:
>>> - Performance improvements:
>>> -- Sense (lockless) if it make sense to try alloc_contig_range() at all
>>>    before directly trying to isolate and taking locks.
>>> -- Try to unplug bigger chunks if possible first.
>>> -- Identify free areas first, that don't have to be evacuated.
>>> - Make unplug more likely to succeed:
>>> -- There are various idea to limit fragmentation on memory block
>>>    granularity. (e.g., ZONE_PREFER_MOVABLE and smart balancing)
>>> -- Allocate memmap from added memory. This way, less unmovable data can
>>>    end up on the memory blocks.
>>> - OOM handling, e.g., via an OOM handler.
>>> - Defragmentation
>>> -- Will require a new virtio-mem CMD to exchange plugged<->unplugged blocks
>>>
>>> --------------------------------------------------------------------------
>>> 6. Example Usage
>>> --------------------------------------------------------------------------
>>>
>>> A very basic QEMU prototype (kept updated) is available at:
>>>     https://github.com/davidhildenbrand/qemu.git virtio-mem
>>>
>>> It lacks various features, however, works to test the guest driver side:
>>> - No support for resizable memory regions / memory backends yet
>>> - No protection of unplugged memory (esp., userfaultfd-wp) yet
>>> - No dump/migration/XXX optimizations to skip unplugged memory (and avoid
>>>   touching it)
>>>
>>> Start QEMU with two virtio-mem devices (one per NUMA node):
>>>  $ qemu-system-x86_64 -m 4G,maxmem=20G \
>>>   -smp sockets=2,cores=2 \
>>>   -numa node,nodeid=0,cpus=0-1 -numa node,nodeid=1,cpus=2-3 \
>>>   [...]
>>>   -object memory-backend-ram,id=mem0,size=8G \
>>>   -device virtio-mem-pci,id=vm0,memdev=mem0,node=0,requested-size=128M \
>>>   -object memory-backend-ram,id=mem1,size=8G \
>>>   -device virtio-mem-pci,id=vm1,memdev=mem1,node=1,requested-size=80M
>>>
>>> Query the configuration:
>>>  QEMU 4.1.95 monitor - type 'help' for more information
>>>  (qemu) info memory-devices
>>>  Memory device [virtio-mem]: "vm0"
>>>    memaddr: 0x140000000
>>>    node: 0
>>>    requested-size: 134217728
>>>    size: 134217728
>>>    max-size: 8589934592
>>>    block-size: 2097152
>>>    memdev: /objects/mem0
>>>  Memory device [virtio-mem]: "vm1"
>>>    memaddr: 0x340000000
>>>    node: 1
>>>    requested-size: 83886080
>>>    size: 83886080
>>>    max-size: 8589934592
>>>    block-size: 2097152
>>>    memdev: /objects/mem1
>>>
>>> Add some memory to node 1:
>>>  QEMU 4.1.95 monitor - type 'help' for more information
>>>  (qemu) qom-set vm1 requested-size 1G
>>>
>>> Remove some memory from node 0:
>>>  QEMU 4.1.95 monitor - type 'help' for more information
>>>  (qemu) qom-set vm0 requested-size 64M
>>>
>>> Query the configuration again:
>>>  (qemu) info memory-devices
>>>  Memory device [virtio-mem]: "vm0"
>>>    memaddr: 0x140000000
>>>    node: 0
>>>    requested-size: 67108864
>>>    size: 67108864
>>>    max-size: 8589934592
>>>    block-size: 2097152
>>>    memdev: /objects/mem0
>>>  Memory device [virtio-mem]: "vm1"
>>>    memaddr: 0x340000000
>>>    node: 1
>>>    requested-size: 1073741824
>>>    size: 1073741824
>>>    max-size: 8589934592
>>>    block-size: 2097152
>>>    memdev: /objects/mem1
>>>
>>> --------------------------------------------------------------------------
>>> 7. Q/A
>>> --------------------------------------------------------------------------
>>>
>>> Q: Why add/remove parts ("subblocks") of memory blocks/sections?
>>> A: Flexibility (section size depends on the architecture) - e.g., some
>>>    architectures have a section size of 2GB. Also, the memory block size
>>>    is variable (e.g., on x86-64). I want to avoid any such restrictions.
>>>    Some use cases want to add/remove memory in smaller granularities to a
>>>    VM (e.g., the Hyper-V balloon also implements this) - especially smaller
>>>    VMs like used for kata containers. Also, on memory unplug, it is more
>>>    reliable to free-up and unplug multiple small chunks instead
>>>    of one big chunk. E.g., if one page of a DIMM is either unmovable or
>>>    pinned, the DIMM can't get unplugged. This approach is basically a
>>>    compromise between DIMM-based memory hot(un)plug and balloon
>>>    inflation/deflation, which works mostly on page granularity.
>>>
>>> Q: Why care about memory blocks?
>>> A: They are the way to tell user space about new memory. This way,
>>>    memory can get onlined/offlined by user space. Also, e.g., kdump
>>>    relies on udev events to reload kexec when memory blocks are
>>>    onlined/offlined. Memory blocks are the "real" memory hot(un)plug
>>>    granularity. Everything that's smaller has to be emulated "on top".
>>>
>>> Q: Won't memory unplug of subblocks fragment memory?
>>> A: Yes and no. Unplugging e.g., >=4MB subblocks on x86-64 will not really
>>>    fragment memory like unplugging random pages like a balloon driver does.
>>>    Buddy merging will not be limited. However, any allocation that requires
>>>    bigger consecutive memory chunks (e.g., gigantic pages) might observe
>>>    the fragmentation. Possible solutions: Allocate gigantic huge pages
>>>    before unplugging memory, don't unplug memory, combine virtio-mem with
>>>    DIMM based memory or bigger initial memory. Remember, a virtio-mem
>>>    device will only unplug on the memory range it manages, not on other
>>>    DIMMs. Unplug of single memory blocks will result in similar
>>>    fragmentation in respect to gigantic huge pages. I ahve plans for a
>>>    virtio-mem defragmentation feature in the future.
>>>
>>> Q: How reliable is memory unplug?
>>> A: There are no guarantees on how much memory can get unplugged
>>>    again. However, it is more likely to find 4MB chunks to unplug than
>>>    e.g., 128MB chunks. If memory is terribly fragmented, there is nothing
>>>    we can do - for now. I consider memory hotplug the first primary use
>>>    of virtio-mem. Memory unplug might usually work, but we want to improve
>>>    the performance and the amount of memory we can actually unplug later.
>>>
>>> Q: Why not unplug from the ZONE_MOVABLE?
>>> A: Unplugged memory chunks are unmovable. Unmovable data must not end up
>>>    on the ZONE_MOVABLE - similar to gigantic pages - they will never be
>>>    allocated from ZONE_MOVABLE. virtio-mem added memory can be onlined
>>>    to the ZONE_MOVABLE, but subblocks will not get unplugged from it.
>>>
>>> Q: How big should the initial (!virtio-mem) memory of a VM be?
>>> A: virtio-mem memory will not go to the DMA zones. So to avoid running out
>>>    of DMA memory, I suggest something like 2-3GB on x86-64. But many
>>>    VMs can most probably deal with less DMA memory - depends on the use
>>>    case.
>>>
>>> [1] https://events.linuxfoundation.org/wp-content/uploads/2017/12/virtio-mem-Paravirtualized-Memory-David-Hildenbrand-Red-Hat-1.pdf
>>> [2] https://lkml.kernel.org/r/20190919142228.5483-1-david@redhat.com
>>> [3] https://lkml.kernel.org/r/547865a9-d6c2-7140-47e2-5af01e7d761d@redhat.com
>>>
>>> Cc: Sebastien Boeuf  <sebastien.boeuf@intel.com>
>>> Cc: Samuel Ortiz <samuel.ortiz@intel.com>
>>> Cc: Robert Bradford <robert.bradford@intel.com>
>>> Cc: Luiz Capitulino <lcapitulino@redhat.com>
>>>
>>> David Hildenbrand (13):
>>>   ACPI: NUMA: export pxm_to_node
>>>   virtio-mem: Paravirtualized memory hotplug
>>>   virtio-mem: Paravirtualized memory hotunplug part 1
>>>   mm: Export alloc_contig_range() / free_contig_range()
>>>   virtio-mem: Paravirtualized memory hotunplug part 2
>>>   mm: Allow to offline unmovable PageOffline() pages via
>>>     MEM_GOING_OFFLINE
>>>   virtio-mem: Allow to offline partially unplugged memory blocks
>>>   mm/memory_hotplug: Introduce offline_and_remove_memory()
>>>   virtio-mem: Offline and remove completely unplugged memory blocks
>>>   virtio-mem: Better retry handling
>>>   mm/vmscan: Move count_vm_event(DROP_SLAB) into drop_slab()
>>>   mm/vmscan: Export drop_slab() and drop_slab_node()
>>>   virtio-mem: Drop slab objects when unplug continues to fail
>>
>> Ping,
>>
>> I'd love to get some feedback on
>>
>> a) The remaining MM bits from MM folks (especially, patch #6 and #8).
> 
> Friendly ping again:
> 
> Can I get some feedback on the two important MM changes in this series
> 
> "[PATCH RFC v4 06/13] mm: Allow to offline unmovable PageOffline() pages
> via MEM_GOING_OFFLINE"
> 
> and
> 
> "[PATCH RFC v4 08/13] mm/memory_hotplug: Introduce
> offline_and_remove_memory()"
> 

Yet another ping.

-- 
Thanks,

David / dhildenb



  reply	other threads:[~2020-02-25  9:59 UTC|newest]

Thread overview: 47+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2019-12-12 17:11 [PATCH RFC v4 00/13] virtio-mem: paravirtualized memory David Hildenbrand
2019-12-12 17:11 ` [PATCH RFC v4 01/13] ACPI: NUMA: export pxm_to_node David Hildenbrand
2019-12-12 21:43   ` Rafael J. Wysocki
2019-12-13  9:41     ` David Hildenbrand
2019-12-13  9:47       ` Rafael J. Wysocki
2019-12-12 17:11 ` [PATCH RFC v4 02/13] virtio-mem: Paravirtualized memory hotplug David Hildenbrand
2019-12-12 17:11 ` [PATCH RFC v4 03/13] virtio-mem: Paravirtualized memory hotunplug part 1 David Hildenbrand
2019-12-12 17:11 ` [PATCH RFC v4 04/13] mm: Export alloc_contig_range() / free_contig_range() David Hildenbrand
2019-12-12 17:11 ` [PATCH RFC v4 05/13] virtio-mem: Paravirtualized memory hotunplug part 2 David Hildenbrand
2019-12-12 17:11 ` [PATCH RFC v4 06/13] mm: Allow to offline unmovable PageOffline() pages via MEM_GOING_OFFLINE David Hildenbrand
2020-02-25 18:26   ` Alexander Duyck
2020-02-25 18:49     ` David Hildenbrand
2020-02-25 21:46       ` Alexander Duyck
2020-02-25 22:19         ` David Hildenbrand
2020-02-26 16:27           ` Alexander Duyck
2019-12-12 17:11 ` [PATCH RFC v4 07/13] virtio-mem: Allow to offline partially unplugged memory blocks David Hildenbrand
2019-12-12 17:11 ` [PATCH RFC v4 08/13] mm/memory_hotplug: Introduce offline_and_remove_memory() David Hildenbrand
2020-02-25 14:11   ` Michal Hocko
2020-02-25 14:27     ` David Hildenbrand
2020-03-02 12:48       ` Michal Hocko
2020-03-02 12:53         ` David Hildenbrand
2019-12-12 17:11 ` [PATCH RFC v4 09/13] virtio-mem: Offline and remove completely unplugged memory blocks David Hildenbrand
2019-12-12 17:11 ` [PATCH RFC v4 10/13] virtio-mem: Better retry handling David Hildenbrand
2019-12-12 17:11 ` [PATCH RFC v4 11/13] mm/vmscan: Move count_vm_event(DROP_SLAB) into drop_slab() David Hildenbrand
2020-02-25 14:13   ` Michal Hocko
2019-12-12 17:11 ` [PATCH RFC v4 12/13] mm/vmscan: Export drop_slab() and drop_slab_node() David Hildenbrand
2020-02-25 14:58   ` Michal Hocko
2020-02-25 15:09     ` David Hildenbrand
2020-02-25 17:06       ` Michal Hocko
2020-02-25 17:23         ` David Hildenbrand
2019-12-12 17:11 ` [PATCH RFC v4 13/13] virtio-mem: Drop slab objects when unplug continues to fail David Hildenbrand
2019-12-13 20:15 ` [PATCH RFC v4 00/13] virtio-mem: paravirtualized memory Konrad Rzeszutek Wilk
2019-12-16 11:03   ` David Hildenbrand
2019-12-24  6:58 ` teawater
2019-12-24  9:28   ` David Hildenbrand
2020-01-09 13:48 ` David Hildenbrand
2020-01-29  9:41   ` David Hildenbrand
2020-02-25  9:58     ` David Hildenbrand [this message]
2020-06-05  8:55   ` Alex Shi
2020-06-05  9:08     ` David Hildenbrand
2020-06-05  9:36       ` David Hildenbrand
2020-06-05 10:05         ` David Hildenbrand
2020-06-05 10:46           ` Alex Shi
2020-06-05 12:18             ` David Hildenbrand
2020-06-09  3:05               ` Alex Shi
2020-06-05 10:08         ` Alex Shi
2020-06-05 10:06       ` Alex Shi

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=3468efe4-0878-3d22-4f74-3092acfd0c47@redhat.com \
    --to=david@redhat.com \
    --cc=akpm@linux-foundation.org \
    --cc=alexander.h.duyck@linux.intel.com \
    --cc=anshuman.khandual@arm.com \
    --cc=anthony.yznaga@oracle.com \
    --cc=cai@lca.pw \
    --cc=dan.j.williams@intel.com \
    --cc=dyoung@redhat.com \
    --cc=glider@google.com \
    --cc=hannes@cmpxchg.org \
    --cc=imammedo@redhat.com \
    --cc=jasowang@redhat.com \
    --cc=jgross@suse.com \
    --cc=kernelfans@gmail.com \
    --cc=konrad.wilk@oracle.com \
    --cc=kvm@vger.kernel.org \
    --cc=lcapitulino@redhat.com \
    --cc=lenb@kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=mgorman@techsingularity.net \
    --cc=mhocko@kernel.org \
    --cc=mhocko@suse.com \
    --cc=mst@redhat.com \
    --cc=osalvador@suse.com \
    --cc=osalvador@suse.de \
    --cc=pasha.tatashin@soleen.com \
    --cc=pavel.tatashin@microsoft.com \
    --cc=richard.weiyang@gmail.com \
    --cc=rjw@rjwysocki.net \
    --cc=robert.bradford@intel.com \
    --cc=rppt@linux.ibm.com \
    --cc=samuel.ortiz@intel.com \
    --cc=sebastien.boeuf@intel.com \
    --cc=stefanha@redhat.com \
    --cc=vbabka@suse.cz \
    --cc=viro@zeniv.linux.org.uk \
    --cc=virtio-dev@lists.oasis-open.org \
    --cc=virtualization@lists.linux-foundation.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).