All of lore.kernel.org
 help / color / mirror / Atom feed
From: Minchan Kim <minchan@kernel.org>
To: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Andrew Morton <akpm@linux-foundation.org>,
	"Aneesh Kumar K.V" <aneesh.kumar@linux.vnet.ibm.com>,
	Marek Szyprowski <m.szyprowski@samsung.com>,
	Michal Nazarewicz <mina86@mina86.com>,
	Russell King - ARM Linux <linux@arm.linux.org.uk>,
	kvm@vger.kernel.org, linux-mm@kvack.org,
	Gleb Natapov <gleb@kernel.org>,
	Greg Kroah-Hartman <gregkh@linuxfoundation.org>,
	Alexander Graf <agraf@suse.de>,
	kvm-ppc@vger.kernel.org, linux-kernel@vger.kernel.org,
	Paul Mackerras <paulus@samba.org>,
	Benjamin Herrenschmidt <benh@kernel.crashing.org>,
	Paolo Bonzini <pbonzini@redhat.com>,
	linuxppc-dev@lists.ozlabs.org,
	linux-arm-kernel@lists.infradead.org
Subject: Re: [PATCH v2 04/10] DMA, CMA: support alignment constraint on cma region
Date: Thu, 12 Jun 2014 14:52:20 +0900	[thread overview]
Message-ID: <20140612055219.GG12415@bbox> (raw)
In-Reply-To: <1402543307-29800-5-git-send-email-iamjoonsoo.kim@lge.com>

On Thu, Jun 12, 2014 at 12:21:41PM +0900, Joonsoo Kim wrote:
> ppc kvm's cma area management needs alignment constraint on
> cma region. So support it to prepare generalization of cma area
> management functionality.
> 
> Additionally, add some comments which tell us why alignment
> constraint is needed on cma region.
> 
> Signed-off-by: Joonsoo Kim <iamjoonsoo.kim@lge.com>
> 
> diff --git a/drivers/base/dma-contiguous.c b/drivers/base/dma-contiguous.c
> index 8a44c82..bc4c171 100644
> --- a/drivers/base/dma-contiguous.c
> +++ b/drivers/base/dma-contiguous.c
> @@ -32,6 +32,7 @@
>  #include <linux/swap.h>
>  #include <linux/mm_types.h>
>  #include <linux/dma-contiguous.h>
> +#include <linux/log2.h>
>  
>  struct cma {
>  	unsigned long	base_pfn;
> @@ -219,6 +220,7 @@ core_initcall(cma_init_reserved_areas);
>   * @size: Size of the reserved area (in bytes),
>   * @base: Base address of the reserved area optional, use 0 for any
>   * @limit: End address of the reserved memory (optional, 0 for any).
> + * @alignment: Alignment for the contiguous memory area, should be power of 2
>   * @res_cma: Pointer to store the created cma region.
>   * @fixed: hint about where to place the reserved area
>   *

Pz, move the all description to new API function rather than internal one.

> @@ -233,15 +235,15 @@ core_initcall(cma_init_reserved_areas);
>   */
>  static int __init __dma_contiguous_reserve_area(phys_addr_t size,
>  				phys_addr_t base, phys_addr_t limit,
> +				phys_addr_t alignment,
>  				struct cma **res_cma, bool fixed)
>  {
>  	struct cma *cma = &cma_areas[cma_area_count];
> -	phys_addr_t alignment;
>  	int ret = 0;
>  
> -	pr_debug("%s(size %lx, base %08lx, limit %08lx)\n", __func__,
> -		 (unsigned long)size, (unsigned long)base,
> -		 (unsigned long)limit);
> +	pr_debug("%s(size %lx, base %08lx, limit %08lx align_order %08lx)\n",

Why is it called by "align_order"?

> +		__func__, (unsigned long)size, (unsigned long)base,
> +		(unsigned long)limit, (unsigned long)alignment);
>  
>  	/* Sanity checks */
>  	if (cma_area_count == ARRAY_SIZE(cma_areas)) {
> @@ -253,8 +255,17 @@ static int __init __dma_contiguous_reserve_area(phys_addr_t size,
>  	if (!size)
>  		return -EINVAL;
>  
> -	/* Sanitise input arguments */
> -	alignment = PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order);
> +	if (alignment && !is_power_of_2(alignment))
> +		return -EINVAL;
> +
> +	/*
> +	 * Sanitise input arguments.
> +	 * CMA area should be at least MAX_ORDER - 1 aligned. Otherwise,
> +	 * CMA area could be merged into other MIGRATE_TYPE by buddy mechanism

I'm not a native but try for clear documenation.

         Pages both ends in CMA area could be merged into adjacent unmovable
         migratetype page by page allocator's buddy algorithm. In the case,
         you couldn't get a contiguous memory, which is not what we want.

> +	 * and CMA property will be broken.
> +	 */
> +	alignment = max(alignment,
> +		(phys_addr_t)PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order));
>  	base = ALIGN(base, alignment);
>  	size = ALIGN(size, alignment);
>  	limit &= ~(alignment - 1);
> @@ -302,7 +313,8 @@ int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base,
>  {
>  	int ret;
>  
> -	ret = __dma_contiguous_reserve_area(size, base, limit, res_cma, fixed);
> +	ret = __dma_contiguous_reserve_area(size, base, limit, 0,
> +						res_cma, fixed);
>  	if (ret)
>  		return ret;
>  
> -- 
> 1.7.9.5

-- 
Kind regards,
Minchan Kim

WARNING: multiple messages have this Message-ID (diff)
From: Minchan Kim <minchan@kernel.org>
To: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Andrew Morton <akpm@linux-foundation.org>,
	"Aneesh Kumar K.V" <aneesh.kumar@linux.vnet.ibm.com>,
	Marek Szyprowski <m.szyprowski@samsung.com>,
	Michal Nazarewicz <mina86@mina86.com>,
	Russell King - ARM Linux <linux@arm.linux.org.uk>,
	kvm@vger.kernel.org, linux-mm@kvack.org,
	Gleb Natapov <gleb@kernel.org>,
	Greg Kroah-Hartman <gregkh@linuxfoundation.org>,
	Alexander Graf <agraf@suse.de>,
	kvm-ppc@vger.kernel.org, linux-kernel@vger.kernel.org,
	Paul Mackerras <paulus@samba.org>,
	Benjamin Herrenschmidt <benh@kernel.crashing.org>,
	Paolo Bonzini <pbonzini@redhat.com>,
	linuxppc-dev@lists.ozlabs.org,
	linux-arm-kernel@lists.infradead.org
Subject: Re: [PATCH v2 04/10] DMA, CMA: support alignment constraint on cma region
Date: Thu, 12 Jun 2014 14:52:20 +0900	[thread overview]
Message-ID: <20140612055219.GG12415@bbox> (raw)
In-Reply-To: <1402543307-29800-5-git-send-email-iamjoonsoo.kim@lge.com>

On Thu, Jun 12, 2014 at 12:21:41PM +0900, Joonsoo Kim wrote:
> ppc kvm's cma area management needs alignment constraint on
> cma region. So support it to prepare generalization of cma area
> management functionality.
> 
> Additionally, add some comments which tell us why alignment
> constraint is needed on cma region.
> 
> Signed-off-by: Joonsoo Kim <iamjoonsoo.kim@lge.com>
> 
> diff --git a/drivers/base/dma-contiguous.c b/drivers/base/dma-contiguous.c
> index 8a44c82..bc4c171 100644
> --- a/drivers/base/dma-contiguous.c
> +++ b/drivers/base/dma-contiguous.c
> @@ -32,6 +32,7 @@
>  #include <linux/swap.h>
>  #include <linux/mm_types.h>
>  #include <linux/dma-contiguous.h>
> +#include <linux/log2.h>
>  
>  struct cma {
>  	unsigned long	base_pfn;
> @@ -219,6 +220,7 @@ core_initcall(cma_init_reserved_areas);
>   * @size: Size of the reserved area (in bytes),
>   * @base: Base address of the reserved area optional, use 0 for any
>   * @limit: End address of the reserved memory (optional, 0 for any).
> + * @alignment: Alignment for the contiguous memory area, should be power of 2
>   * @res_cma: Pointer to store the created cma region.
>   * @fixed: hint about where to place the reserved area
>   *

Pz, move the all description to new API function rather than internal one.

> @@ -233,15 +235,15 @@ core_initcall(cma_init_reserved_areas);
>   */
>  static int __init __dma_contiguous_reserve_area(phys_addr_t size,
>  				phys_addr_t base, phys_addr_t limit,
> +				phys_addr_t alignment,
>  				struct cma **res_cma, bool fixed)
>  {
>  	struct cma *cma = &cma_areas[cma_area_count];
> -	phys_addr_t alignment;
>  	int ret = 0;
>  
> -	pr_debug("%s(size %lx, base %08lx, limit %08lx)\n", __func__,
> -		 (unsigned long)size, (unsigned long)base,
> -		 (unsigned long)limit);
> +	pr_debug("%s(size %lx, base %08lx, limit %08lx align_order %08lx)\n",

Why is it called by "align_order"?

> +		__func__, (unsigned long)size, (unsigned long)base,
> +		(unsigned long)limit, (unsigned long)alignment);
>  
>  	/* Sanity checks */
>  	if (cma_area_count == ARRAY_SIZE(cma_areas)) {
> @@ -253,8 +255,17 @@ static int __init __dma_contiguous_reserve_area(phys_addr_t size,
>  	if (!size)
>  		return -EINVAL;
>  
> -	/* Sanitise input arguments */
> -	alignment = PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order);
> +	if (alignment && !is_power_of_2(alignment))
> +		return -EINVAL;
> +
> +	/*
> +	 * Sanitise input arguments.
> +	 * CMA area should be at least MAX_ORDER - 1 aligned. Otherwise,
> +	 * CMA area could be merged into other MIGRATE_TYPE by buddy mechanism

I'm not a native but try for clear documenation.

         Pages both ends in CMA area could be merged into adjacent unmovable
         migratetype page by page allocator's buddy algorithm. In the case,
         you couldn't get a contiguous memory, which is not what we want.

> +	 * and CMA property will be broken.
> +	 */
> +	alignment = max(alignment,
> +		(phys_addr_t)PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order));
>  	base = ALIGN(base, alignment);
>  	size = ALIGN(size, alignment);
>  	limit &= ~(alignment - 1);
> @@ -302,7 +313,8 @@ int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base,
>  {
>  	int ret;
>  
> -	ret = __dma_contiguous_reserve_area(size, base, limit, res_cma, fixed);
> +	ret = __dma_contiguous_reserve_area(size, base, limit, 0,
> +						res_cma, fixed);
>  	if (ret)
>  		return ret;
>  
> -- 
> 1.7.9.5

-- 
Kind regards,
Minchan Kim

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

WARNING: multiple messages have this Message-ID (diff)
From: Minchan Kim <minchan@kernel.org>
To: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: kvm-ppc@vger.kernel.org,
	Russell King - ARM Linux <linux@arm.linux.org.uk>,
	kvm@vger.kernel.org, Gleb Natapov <gleb@kernel.org>,
	Greg Kroah-Hartman <gregkh@linuxfoundation.org>,
	Alexander Graf <agraf@suse.de>,
	Michal Nazarewicz <mina86@mina86.com>,
	linux-kernel@vger.kernel.org, linux-mm@kvack.org,
	Paul Mackerras <paulus@samba.org>,
	"Aneesh Kumar K.V" <aneesh.kumar@linux.vnet.ibm.com>,
	Paolo Bonzini <pbonzini@redhat.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	linuxppc-dev@lists.ozlabs.org,
	linux-arm-kernel@lists.infradead.org,
	Marek Szyprowski <m.szyprowski@samsung.com>
Subject: Re: [PATCH v2 04/10] DMA, CMA: support alignment constraint on cma region
Date: Thu, 12 Jun 2014 14:52:20 +0900	[thread overview]
Message-ID: <20140612055219.GG12415@bbox> (raw)
In-Reply-To: <1402543307-29800-5-git-send-email-iamjoonsoo.kim@lge.com>

On Thu, Jun 12, 2014 at 12:21:41PM +0900, Joonsoo Kim wrote:
> ppc kvm's cma area management needs alignment constraint on
> cma region. So support it to prepare generalization of cma area
> management functionality.
> 
> Additionally, add some comments which tell us why alignment
> constraint is needed on cma region.
> 
> Signed-off-by: Joonsoo Kim <iamjoonsoo.kim@lge.com>
> 
> diff --git a/drivers/base/dma-contiguous.c b/drivers/base/dma-contiguous.c
> index 8a44c82..bc4c171 100644
> --- a/drivers/base/dma-contiguous.c
> +++ b/drivers/base/dma-contiguous.c
> @@ -32,6 +32,7 @@
>  #include <linux/swap.h>
>  #include <linux/mm_types.h>
>  #include <linux/dma-contiguous.h>
> +#include <linux/log2.h>
>  
>  struct cma {
>  	unsigned long	base_pfn;
> @@ -219,6 +220,7 @@ core_initcall(cma_init_reserved_areas);
>   * @size: Size of the reserved area (in bytes),
>   * @base: Base address of the reserved area optional, use 0 for any
>   * @limit: End address of the reserved memory (optional, 0 for any).
> + * @alignment: Alignment for the contiguous memory area, should be power of 2
>   * @res_cma: Pointer to store the created cma region.
>   * @fixed: hint about where to place the reserved area
>   *

Pz, move the all description to new API function rather than internal one.

> @@ -233,15 +235,15 @@ core_initcall(cma_init_reserved_areas);
>   */
>  static int __init __dma_contiguous_reserve_area(phys_addr_t size,
>  				phys_addr_t base, phys_addr_t limit,
> +				phys_addr_t alignment,
>  				struct cma **res_cma, bool fixed)
>  {
>  	struct cma *cma = &cma_areas[cma_area_count];
> -	phys_addr_t alignment;
>  	int ret = 0;
>  
> -	pr_debug("%s(size %lx, base %08lx, limit %08lx)\n", __func__,
> -		 (unsigned long)size, (unsigned long)base,
> -		 (unsigned long)limit);
> +	pr_debug("%s(size %lx, base %08lx, limit %08lx align_order %08lx)\n",

Why is it called by "align_order"?

> +		__func__, (unsigned long)size, (unsigned long)base,
> +		(unsigned long)limit, (unsigned long)alignment);
>  
>  	/* Sanity checks */
>  	if (cma_area_count == ARRAY_SIZE(cma_areas)) {
> @@ -253,8 +255,17 @@ static int __init __dma_contiguous_reserve_area(phys_addr_t size,
>  	if (!size)
>  		return -EINVAL;
>  
> -	/* Sanitise input arguments */
> -	alignment = PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order);
> +	if (alignment && !is_power_of_2(alignment))
> +		return -EINVAL;
> +
> +	/*
> +	 * Sanitise input arguments.
> +	 * CMA area should be at least MAX_ORDER - 1 aligned. Otherwise,
> +	 * CMA area could be merged into other MIGRATE_TYPE by buddy mechanism

I'm not a native but try for clear documenation.

         Pages both ends in CMA area could be merged into adjacent unmovable
         migratetype page by page allocator's buddy algorithm. In the case,
         you couldn't get a contiguous memory, which is not what we want.

> +	 * and CMA property will be broken.
> +	 */
> +	alignment = max(alignment,
> +		(phys_addr_t)PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order));
>  	base = ALIGN(base, alignment);
>  	size = ALIGN(size, alignment);
>  	limit &= ~(alignment - 1);
> @@ -302,7 +313,8 @@ int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base,
>  {
>  	int ret;
>  
> -	ret = __dma_contiguous_reserve_area(size, base, limit, res_cma, fixed);
> +	ret = __dma_contiguous_reserve_area(size, base, limit, 0,
> +						res_cma, fixed);
>  	if (ret)
>  		return ret;
>  
> -- 
> 1.7.9.5

-- 
Kind regards,
Minchan Kim

WARNING: multiple messages have this Message-ID (diff)
From: minchan@kernel.org (Minchan Kim)
To: linux-arm-kernel@lists.infradead.org
Subject: [PATCH v2 04/10] DMA, CMA: support alignment constraint on cma region
Date: Thu, 12 Jun 2014 14:52:20 +0900	[thread overview]
Message-ID: <20140612055219.GG12415@bbox> (raw)
In-Reply-To: <1402543307-29800-5-git-send-email-iamjoonsoo.kim@lge.com>

On Thu, Jun 12, 2014 at 12:21:41PM +0900, Joonsoo Kim wrote:
> ppc kvm's cma area management needs alignment constraint on
> cma region. So support it to prepare generalization of cma area
> management functionality.
> 
> Additionally, add some comments which tell us why alignment
> constraint is needed on cma region.
> 
> Signed-off-by: Joonsoo Kim <iamjoonsoo.kim@lge.com>
> 
> diff --git a/drivers/base/dma-contiguous.c b/drivers/base/dma-contiguous.c
> index 8a44c82..bc4c171 100644
> --- a/drivers/base/dma-contiguous.c
> +++ b/drivers/base/dma-contiguous.c
> @@ -32,6 +32,7 @@
>  #include <linux/swap.h>
>  #include <linux/mm_types.h>
>  #include <linux/dma-contiguous.h>
> +#include <linux/log2.h>
>  
>  struct cma {
>  	unsigned long	base_pfn;
> @@ -219,6 +220,7 @@ core_initcall(cma_init_reserved_areas);
>   * @size: Size of the reserved area (in bytes),
>   * @base: Base address of the reserved area optional, use 0 for any
>   * @limit: End address of the reserved memory (optional, 0 for any).
> + * @alignment: Alignment for the contiguous memory area, should be power of 2
>   * @res_cma: Pointer to store the created cma region.
>   * @fixed: hint about where to place the reserved area
>   *

Pz, move the all description to new API function rather than internal one.

> @@ -233,15 +235,15 @@ core_initcall(cma_init_reserved_areas);
>   */
>  static int __init __dma_contiguous_reserve_area(phys_addr_t size,
>  				phys_addr_t base, phys_addr_t limit,
> +				phys_addr_t alignment,
>  				struct cma **res_cma, bool fixed)
>  {
>  	struct cma *cma = &cma_areas[cma_area_count];
> -	phys_addr_t alignment;
>  	int ret = 0;
>  
> -	pr_debug("%s(size %lx, base %08lx, limit %08lx)\n", __func__,
> -		 (unsigned long)size, (unsigned long)base,
> -		 (unsigned long)limit);
> +	pr_debug("%s(size %lx, base %08lx, limit %08lx align_order %08lx)\n",

Why is it called by "align_order"?

> +		__func__, (unsigned long)size, (unsigned long)base,
> +		(unsigned long)limit, (unsigned long)alignment);
>  
>  	/* Sanity checks */
>  	if (cma_area_count == ARRAY_SIZE(cma_areas)) {
> @@ -253,8 +255,17 @@ static int __init __dma_contiguous_reserve_area(phys_addr_t size,
>  	if (!size)
>  		return -EINVAL;
>  
> -	/* Sanitise input arguments */
> -	alignment = PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order);
> +	if (alignment && !is_power_of_2(alignment))
> +		return -EINVAL;
> +
> +	/*
> +	 * Sanitise input arguments.
> +	 * CMA area should be at least MAX_ORDER - 1 aligned. Otherwise,
> +	 * CMA area could be merged into other MIGRATE_TYPE by buddy mechanism

I'm not a native but try for clear documenation.

         Pages both ends in CMA area could be merged into adjacent unmovable
         migratetype page by page allocator's buddy algorithm. In the case,
         you couldn't get a contiguous memory, which is not what we want.

> +	 * and CMA property will be broken.
> +	 */
> +	alignment = max(alignment,
> +		(phys_addr_t)PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order));
>  	base = ALIGN(base, alignment);
>  	size = ALIGN(size, alignment);
>  	limit &= ~(alignment - 1);
> @@ -302,7 +313,8 @@ int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base,
>  {
>  	int ret;
>  
> -	ret = __dma_contiguous_reserve_area(size, base, limit, res_cma, fixed);
> +	ret = __dma_contiguous_reserve_area(size, base, limit, 0,
> +						res_cma, fixed);
>  	if (ret)
>  		return ret;
>  
> -- 
> 1.7.9.5

-- 
Kind regards,
Minchan Kim

WARNING: multiple messages have this Message-ID (diff)
From: Minchan Kim <minchan@kernel.org>
To: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Andrew Morton <akpm@linux-foundation.org>,
	"Aneesh Kumar K.V" <aneesh.kumar@linux.vnet.ibm.com>,
	Marek Szyprowski <m.szyprowski@samsung.com>,
	Michal Nazarewicz <mina86@mina86.com>,
	Russell King - ARM Linux <linux@arm.linux.org.uk>,
	kvm@vger.kernel.org, linux-mm@kvack.org,
	Gleb Natapov <gleb@kernel.org>,
	Greg Kroah-Hartman <gregkh@linuxfoundation.org>,
	Alexander Graf <agraf@suse.de>,
	kvm-ppc@vger.kernel.org, linux-kernel@vger.kernel.org,
	Paul Mackerras <paulus@samba.org>,
	Benjamin Herrenschmidt <benh@kernel.crashing.org>,
	Paolo Bonzini <pbonzini@redhat.com>,
	linuxppc-dev@lists.ozlabs.org,
	linux-arm-kernel@lists.infradead.org
Subject: Re: [PATCH v2 04/10] DMA, CMA: support alignment constraint on cma region
Date: Thu, 12 Jun 2014 05:52:20 +0000	[thread overview]
Message-ID: <20140612055219.GG12415@bbox> (raw)
In-Reply-To: <1402543307-29800-5-git-send-email-iamjoonsoo.kim@lge.com>

On Thu, Jun 12, 2014 at 12:21:41PM +0900, Joonsoo Kim wrote:
> ppc kvm's cma area management needs alignment constraint on
> cma region. So support it to prepare generalization of cma area
> management functionality.
> 
> Additionally, add some comments which tell us why alignment
> constraint is needed on cma region.
> 
> Signed-off-by: Joonsoo Kim <iamjoonsoo.kim@lge.com>
> 
> diff --git a/drivers/base/dma-contiguous.c b/drivers/base/dma-contiguous.c
> index 8a44c82..bc4c171 100644
> --- a/drivers/base/dma-contiguous.c
> +++ b/drivers/base/dma-contiguous.c
> @@ -32,6 +32,7 @@
>  #include <linux/swap.h>
>  #include <linux/mm_types.h>
>  #include <linux/dma-contiguous.h>
> +#include <linux/log2.h>
>  
>  struct cma {
>  	unsigned long	base_pfn;
> @@ -219,6 +220,7 @@ core_initcall(cma_init_reserved_areas);
>   * @size: Size of the reserved area (in bytes),
>   * @base: Base address of the reserved area optional, use 0 for any
>   * @limit: End address of the reserved memory (optional, 0 for any).
> + * @alignment: Alignment for the contiguous memory area, should be power of 2
>   * @res_cma: Pointer to store the created cma region.
>   * @fixed: hint about where to place the reserved area
>   *

Pz, move the all description to new API function rather than internal one.

> @@ -233,15 +235,15 @@ core_initcall(cma_init_reserved_areas);
>   */
>  static int __init __dma_contiguous_reserve_area(phys_addr_t size,
>  				phys_addr_t base, phys_addr_t limit,
> +				phys_addr_t alignment,
>  				struct cma **res_cma, bool fixed)
>  {
>  	struct cma *cma = &cma_areas[cma_area_count];
> -	phys_addr_t alignment;
>  	int ret = 0;
>  
> -	pr_debug("%s(size %lx, base %08lx, limit %08lx)\n", __func__,
> -		 (unsigned long)size, (unsigned long)base,
> -		 (unsigned long)limit);
> +	pr_debug("%s(size %lx, base %08lx, limit %08lx align_order %08lx)\n",

Why is it called by "align_order"?

> +		__func__, (unsigned long)size, (unsigned long)base,
> +		(unsigned long)limit, (unsigned long)alignment);
>  
>  	/* Sanity checks */
>  	if (cma_area_count = ARRAY_SIZE(cma_areas)) {
> @@ -253,8 +255,17 @@ static int __init __dma_contiguous_reserve_area(phys_addr_t size,
>  	if (!size)
>  		return -EINVAL;
>  
> -	/* Sanitise input arguments */
> -	alignment = PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order);
> +	if (alignment && !is_power_of_2(alignment))
> +		return -EINVAL;
> +
> +	/*
> +	 * Sanitise input arguments.
> +	 * CMA area should be at least MAX_ORDER - 1 aligned. Otherwise,
> +	 * CMA area could be merged into other MIGRATE_TYPE by buddy mechanism

I'm not a native but try for clear documenation.

         Pages both ends in CMA area could be merged into adjacent unmovable
         migratetype page by page allocator's buddy algorithm. In the case,
         you couldn't get a contiguous memory, which is not what we want.

> +	 * and CMA property will be broken.
> +	 */
> +	alignment = max(alignment,
> +		(phys_addr_t)PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order));
>  	base = ALIGN(base, alignment);
>  	size = ALIGN(size, alignment);
>  	limit &= ~(alignment - 1);
> @@ -302,7 +313,8 @@ int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base,
>  {
>  	int ret;
>  
> -	ret = __dma_contiguous_reserve_area(size, base, limit, res_cma, fixed);
> +	ret = __dma_contiguous_reserve_area(size, base, limit, 0,
> +						res_cma, fixed);
>  	if (ret)
>  		return ret;
>  
> -- 
> 1.7.9.5

-- 
Kind regards,
Minchan Kim

  parent reply	other threads:[~2014-06-12  5:52 UTC|newest]

Thread overview: 372+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2014-06-12  3:21 [PATCH v2 00/10] CMA: generalize CMA reserved area management code Joonsoo Kim
2014-06-12  3:21 ` Joonsoo Kim
2014-06-12  3:21 ` Joonsoo Kim
2014-06-12  3:21 ` Joonsoo Kim
2014-06-12  3:21 ` Joonsoo Kim
2014-06-12  3:21 ` [PATCH v2 01/10] DMA, CMA: clean-up log message Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  4:41   ` Aneesh Kumar K.V
2014-06-12  4:53     ` Aneesh Kumar K.V
2014-06-12  4:41     ` Aneesh Kumar K.V
2014-06-12  4:41     ` Aneesh Kumar K.V
2014-06-12  4:41     ` Aneesh Kumar K.V
2014-06-12  4:41     ` Aneesh Kumar K.V
2014-06-12  5:53     ` Joonsoo Kim
2014-06-12  5:53       ` Joonsoo Kim
2014-06-12  5:53       ` Joonsoo Kim
2014-06-12  5:53       ` Joonsoo Kim
2014-06-12  5:53       ` Joonsoo Kim
2014-06-12  8:55       ` Michal Nazarewicz
2014-06-12  8:55         ` Michal Nazarewicz
2014-06-12  8:55         ` Michal Nazarewicz
2014-06-12  8:55         ` Michal Nazarewicz
2014-06-12  8:55         ` Michal Nazarewicz
2014-06-12  9:53         ` Michal Nazarewicz
2014-06-12  9:53           ` Michal Nazarewicz
2014-06-12  9:53           ` Michal Nazarewicz
2014-06-12  9:53           ` Michal Nazarewicz
2014-06-12  9:53           ` Michal Nazarewicz
2014-06-16  5:18           ` Joonsoo Kim
2014-06-16  5:18             ` Joonsoo Kim
2014-06-16  5:18             ` Joonsoo Kim
2014-06-16  5:18             ` Joonsoo Kim
2014-06-16  5:18             ` Joonsoo Kim
2014-06-16  5:18             ` Joonsoo Kim
2014-06-12  5:18   ` Minchan Kim
2014-06-12  5:18     ` Minchan Kim
2014-06-12  5:18     ` Minchan Kim
2014-06-12  5:18     ` Minchan Kim
2014-06-12  5:18     ` Minchan Kim
2014-06-12  5:55     ` Joonsoo Kim
2014-06-12  5:55       ` Joonsoo Kim
2014-06-12  5:55       ` Joonsoo Kim
2014-06-12  5:55       ` Joonsoo Kim
2014-06-12  5:55       ` Joonsoo Kim
2014-06-12  8:15   ` Zhang Yanfei
2014-06-12  8:15     ` Zhang Yanfei
2014-06-12  8:15     ` Zhang Yanfei
2014-06-12  8:15     ` Zhang Yanfei
2014-06-12  8:15     ` Zhang Yanfei
2014-06-12  8:15     ` Zhang Yanfei
2014-06-12  8:56   ` Michal Nazarewicz
2014-06-12  8:56     ` Michal Nazarewicz
2014-06-12  8:56     ` Michal Nazarewicz
2014-06-12  8:56     ` Michal Nazarewicz
2014-06-12  8:56     ` Michal Nazarewicz
2014-06-12  8:56     ` Michal Nazarewicz
2014-06-12  3:21 ` [PATCH v2 02/10] DMA, CMA: fix possible memory leak Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  4:43   ` Aneesh Kumar K.V
2014-06-12  4:55     ` Aneesh Kumar K.V
2014-06-12  4:43     ` Aneesh Kumar K.V
2014-06-12  4:43     ` Aneesh Kumar K.V
2014-06-12  4:43     ` Aneesh Kumar K.V
2014-06-12  4:43     ` Aneesh Kumar K.V
2014-06-12  5:25   ` Minchan Kim
2014-06-12  5:25     ` Minchan Kim
2014-06-12  5:25     ` Minchan Kim
2014-06-12  5:25     ` Minchan Kim
2014-06-12  5:25     ` Minchan Kim
2014-06-12  5:58     ` Joonsoo Kim
2014-06-12  6:02       ` Joonsoo Kim
2014-06-12  6:02       ` Joonsoo Kim
2014-06-12  6:02       ` Joonsoo Kim
2014-06-12  6:02       ` Joonsoo Kim
2014-06-12  8:19       ` Zhang Yanfei
2014-06-12  8:19         ` Zhang Yanfei
2014-06-12  8:19         ` Zhang Yanfei
2014-06-12  8:19         ` Zhang Yanfei
2014-06-12  8:19         ` Zhang Yanfei
2014-06-12  8:19         ` Zhang Yanfei
2014-06-12  9:47   ` Michal Nazarewicz
2014-06-12  9:47     ` Michal Nazarewicz
2014-06-12  9:47     ` Michal Nazarewicz
2014-06-12  9:47     ` Michal Nazarewicz
2014-06-12  9:47     ` Michal Nazarewicz
2014-06-12  9:47     ` Michal Nazarewicz
2014-06-12  3:21 ` [PATCH v2 03/10] DMA, CMA: separate core cma management codes from DMA APIs Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  4:44   ` Aneesh Kumar K.V
2014-06-12  4:56     ` Aneesh Kumar K.V
2014-06-12  4:44     ` Aneesh Kumar K.V
2014-06-12  4:44     ` Aneesh Kumar K.V
2014-06-12  4:44     ` Aneesh Kumar K.V
2014-06-12  4:44     ` Aneesh Kumar K.V
2014-06-12  5:37   ` Minchan Kim
2014-06-12  5:37     ` Minchan Kim
2014-06-12  5:37     ` Minchan Kim
2014-06-12  5:37     ` Minchan Kim
2014-06-12  5:37     ` Minchan Kim
2014-06-16  5:24     ` Joonsoo Kim
2014-06-16  5:24       ` Joonsoo Kim
2014-06-16  5:24       ` Joonsoo Kim
2014-06-16  5:24       ` Joonsoo Kim
2014-06-16  5:24       ` Joonsoo Kim
2014-06-12  9:55   ` Michal Nazarewicz
2014-06-12  9:55     ` Michal Nazarewicz
2014-06-12  9:55     ` Michal Nazarewicz
2014-06-12  9:55     ` Michal Nazarewicz
2014-06-12  9:55     ` Michal Nazarewicz
2014-06-12  9:55     ` Michal Nazarewicz
2014-06-12  3:21 ` [PATCH v2 04/10] DMA, CMA: support alignment constraint on cma region Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  4:50   ` Aneesh Kumar K.V
2014-06-12  4:50     ` Aneesh Kumar K.V
2014-06-12  4:50     ` Aneesh Kumar K.V
2014-06-12  4:50     ` Aneesh Kumar K.V
2014-06-12  4:50     ` Aneesh Kumar K.V
2014-06-12  4:50     ` Aneesh Kumar K.V
2014-06-12  5:52   ` Minchan Kim [this message]
2014-06-12  5:52     ` Minchan Kim
2014-06-12  5:52     ` Minchan Kim
2014-06-12  5:52     ` Minchan Kim
2014-06-12  5:52     ` Minchan Kim
2014-06-12  6:07     ` Joonsoo Kim
2014-06-12  6:07       ` Joonsoo Kim
2014-06-12  6:07       ` Joonsoo Kim
2014-06-12  6:07       ` Joonsoo Kim
2014-06-12  6:07       ` Joonsoo Kim
2014-06-12 10:02   ` Michal Nazarewicz
2014-06-12 10:02     ` Michal Nazarewicz
2014-06-12 10:02     ` Michal Nazarewicz
2014-06-12 10:02     ` Michal Nazarewicz
2014-06-12 10:02     ` Michal Nazarewicz
2014-06-12 10:02     ` Michal Nazarewicz
2014-06-16  5:19     ` Joonsoo Kim
2014-06-16  5:19       ` Joonsoo Kim
2014-06-16  5:19       ` Joonsoo Kim
2014-06-16  5:19       ` Joonsoo Kim
2014-06-16  5:19       ` Joonsoo Kim
2014-06-16  5:19       ` Joonsoo Kim
2014-06-12  3:21 ` [PATCH v2 05/10] DMA, CMA: support arbitrary bitmap granularity Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  6:06   ` Minchan Kim
2014-06-12  6:06     ` Minchan Kim
2014-06-12  6:06     ` Minchan Kim
2014-06-12  6:06     ` Minchan Kim
2014-06-12  6:06     ` Minchan Kim
2014-06-12  6:43     ` Joonsoo Kim
2014-06-12  6:43       ` Joonsoo Kim
2014-06-12  6:43       ` Joonsoo Kim
2014-06-12  6:43       ` Joonsoo Kim
2014-06-12  6:43       ` Joonsoo Kim
2014-06-12  6:42       ` Minchan Kim
2014-06-12  6:42         ` Minchan Kim
2014-06-12  6:42         ` Minchan Kim
2014-06-12  6:42         ` Minchan Kim
2014-06-12  6:42         ` Minchan Kim
2014-06-12  7:08   ` Minchan Kim
2014-06-12  7:08     ` Minchan Kim
2014-06-12  7:08     ` Minchan Kim
2014-06-12  7:08     ` Minchan Kim
2014-06-12  7:08     ` Minchan Kim
2014-06-12  7:25     ` Zhang Yanfei
2014-06-12  7:25       ` Zhang Yanfei
2014-06-12  7:25       ` Zhang Yanfei
2014-06-12  7:25       ` Zhang Yanfei
2014-06-12  7:25       ` Zhang Yanfei
2014-06-12  7:25       ` Zhang Yanfei
2014-06-12  7:41     ` Joonsoo Kim
2014-06-12  7:41       ` Joonsoo Kim
2014-06-12  7:41       ` Joonsoo Kim
2014-06-12  7:41       ` Joonsoo Kim
2014-06-12  7:41       ` Joonsoo Kim
2014-06-12  8:28   ` Zhang Yanfei
2014-06-12  8:28     ` Zhang Yanfei
2014-06-12  8:28     ` Zhang Yanfei
2014-06-12  8:28     ` Zhang Yanfei
2014-06-12  8:28     ` Zhang Yanfei
2014-06-12  8:28     ` Zhang Yanfei
2014-06-12 10:19   ` Michal Nazarewicz
2014-06-12 10:19     ` Michal Nazarewicz
2014-06-12 10:19     ` Michal Nazarewicz
2014-06-12 10:19     ` Michal Nazarewicz
2014-06-12 10:19     ` Michal Nazarewicz
2014-06-12 10:19     ` Michal Nazarewicz
2014-06-16  5:23     ` Joonsoo Kim
2014-06-16  5:23       ` Joonsoo Kim
2014-06-16  5:23       ` Joonsoo Kim
2014-06-16  5:23       ` Joonsoo Kim
2014-06-16  5:23       ` Joonsoo Kim
2014-06-14 10:09   ` Aneesh Kumar K.V
2014-06-14 10:21     ` Aneesh Kumar K.V
2014-06-14 10:09     ` Aneesh Kumar K.V
2014-06-14 10:09     ` Aneesh Kumar K.V
2014-06-14 10:09     ` Aneesh Kumar K.V
2014-06-14 10:09     ` Aneesh Kumar K.V
2014-06-12  3:21 ` [PATCH v2 06/10] CMA: generalize CMA reserved area management functionality Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  7:13   ` Minchan Kim
2014-06-12  7:13     ` Minchan Kim
2014-06-12  7:13     ` Minchan Kim
2014-06-12  7:13     ` Minchan Kim
2014-06-12  7:13     ` Minchan Kim
2014-06-12  7:42     ` Joonsoo Kim
2014-06-12  7:42       ` Joonsoo Kim
2014-06-12  7:42       ` Joonsoo Kim
2014-06-12  7:42       ` Joonsoo Kim
2014-06-12  7:42       ` Joonsoo Kim
2014-06-12  8:29   ` Zhang Yanfei
2014-06-12  8:29     ` Zhang Yanfei
2014-06-12  8:29     ` Zhang Yanfei
2014-06-12  8:29     ` Zhang Yanfei
2014-06-12  8:29     ` Zhang Yanfei
2014-06-12  8:29     ` Zhang Yanfei
2014-06-14 10:06   ` Aneesh Kumar K.V
2014-06-14 10:18     ` Aneesh Kumar K.V
2014-06-14 10:06     ` Aneesh Kumar K.V
2014-06-14 10:06     ` Aneesh Kumar K.V
2014-06-14 10:06     ` Aneesh Kumar K.V
2014-06-14 10:06     ` Aneesh Kumar K.V
2014-06-14 10:08   ` Aneesh Kumar K.V
2014-06-14 10:20     ` Aneesh Kumar K.V
2014-06-14 10:08     ` Aneesh Kumar K.V
2014-06-14 10:08     ` Aneesh Kumar K.V
2014-06-14 10:08     ` Aneesh Kumar K.V
2014-06-14 10:08     ` Aneesh Kumar K.V
2014-06-14 10:16   ` Aneesh Kumar K.V
2014-06-14 10:28     ` Aneesh Kumar K.V
2014-06-14 10:16     ` Aneesh Kumar K.V
2014-06-14 10:16     ` Aneesh Kumar K.V
2014-06-14 10:16     ` Aneesh Kumar K.V
2014-06-14 10:16     ` Aneesh Kumar K.V
2014-06-16  5:27     ` Joonsoo Kim
2014-06-16  5:27       ` Joonsoo Kim
2014-06-16  5:27       ` Joonsoo Kim
2014-06-16  5:27       ` Joonsoo Kim
2014-06-16  5:27       ` Joonsoo Kim
2014-06-12  3:21 ` [PATCH v2 07/10] PPC, KVM, CMA: use general CMA reserved area management framework Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-14  8:53   ` Aneesh Kumar K.V
2014-06-14  8:54     ` Aneesh Kumar K.V
2014-06-14  8:53     ` Aneesh Kumar K.V
2014-06-14  8:53     ` Aneesh Kumar K.V
2014-06-14  8:53     ` Aneesh Kumar K.V
2014-06-14  8:53     ` Aneesh Kumar K.V
2014-06-16  5:34     ` Joonsoo Kim
2014-06-16  5:34       ` Joonsoo Kim
2014-06-16  5:34       ` Joonsoo Kim
2014-06-16  5:34       ` Joonsoo Kim
2014-06-16  5:34       ` Joonsoo Kim
2014-06-16  7:02       ` Aneesh Kumar K.V
2014-06-16  7:14         ` Aneesh Kumar K.V
2014-06-16  7:02         ` Aneesh Kumar K.V
2014-06-16  7:02         ` Aneesh Kumar K.V
2014-06-16  7:02         ` Aneesh Kumar K.V
2014-06-14 10:05   ` Aneesh Kumar K.V
2014-06-14 10:17     ` Aneesh Kumar K.V
2014-06-14 10:05     ` Aneesh Kumar K.V
2014-06-14 10:05     ` Aneesh Kumar K.V
2014-06-14 10:05     ` Aneesh Kumar K.V
2014-06-14 10:05     ` Aneesh Kumar K.V
2014-06-16  5:29     ` Joonsoo Kim
2014-06-16  5:29       ` Joonsoo Kim
2014-06-16  5:29       ` Joonsoo Kim
2014-06-16  5:29       ` Joonsoo Kim
2014-06-16  5:29       ` Joonsoo Kim
2014-06-12  3:21 ` [PATCH v2 08/10] mm, cma: clean-up cma allocation error path Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  7:16   ` Minchan Kim
2014-06-12  7:16     ` Minchan Kim
2014-06-12  7:16     ` Minchan Kim
2014-06-12  7:16     ` Minchan Kim
2014-06-12  7:16     ` Minchan Kim
2014-06-12  8:31   ` Zhang Yanfei
2014-06-12  8:31     ` Zhang Yanfei
2014-06-12  8:31     ` Zhang Yanfei
2014-06-12  8:31     ` Zhang Yanfei
2014-06-12  8:31     ` Zhang Yanfei
2014-06-12  8:31     ` Zhang Yanfei
2014-06-12 11:34   ` Michal Nazarewicz
2014-06-12 11:34     ` Michal Nazarewicz
2014-06-12 11:34     ` Michal Nazarewicz
2014-06-12 11:34     ` Michal Nazarewicz
2014-06-12 11:34     ` Michal Nazarewicz
2014-06-12 11:34     ` Michal Nazarewicz
2014-06-14  7:18   ` Aneesh Kumar K.V
2014-06-14  7:30     ` Aneesh Kumar K.V
2014-06-14  7:18     ` Aneesh Kumar K.V
2014-06-14  7:18     ` Aneesh Kumar K.V
2014-06-14  7:18     ` Aneesh Kumar K.V
2014-06-12  3:21 ` [PATCH v2 09/10] mm, cma: move output param to the end of param list Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  7:19   ` Minchan Kim
2014-06-12  7:19     ` Minchan Kim
2014-06-12  7:19     ` Minchan Kim
2014-06-12  7:19     ` Minchan Kim
2014-06-12  7:19     ` Minchan Kim
2014-06-12  7:43     ` Joonsoo Kim
2014-06-12  7:43       ` Joonsoo Kim
2014-06-12  7:43       ` Joonsoo Kim
2014-06-12  7:43       ` Joonsoo Kim
2014-06-12  7:43       ` Joonsoo Kim
2014-06-12 11:38   ` Michal Nazarewicz
2014-06-12 11:38     ` Michal Nazarewicz
2014-06-12 11:38     ` Michal Nazarewicz
2014-06-12 11:38     ` Michal Nazarewicz
2014-06-12 11:38     ` Michal Nazarewicz
2014-06-12 11:38     ` Michal Nazarewicz
2014-06-14  7:20   ` Aneesh Kumar K.V
2014-06-14  7:32     ` Aneesh Kumar K.V
2014-06-14  7:20     ` Aneesh Kumar K.V
2014-06-14  7:20     ` Aneesh Kumar K.V
2014-06-14  7:20     ` Aneesh Kumar K.V
2014-06-14  7:20     ` Aneesh Kumar K.V
2014-06-12  3:21 ` [PATCH v2 10/10] mm, cma: use spinlock instead of mutex Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  3:21   ` Joonsoo Kim
2014-06-12  7:40   ` Minchan Kim
2014-06-12  7:40     ` Minchan Kim
2014-06-12  7:40     ` Minchan Kim
2014-06-12  7:40     ` Minchan Kim
2014-06-12  7:40     ` Minchan Kim
2014-06-12  7:56     ` Joonsoo Kim
2014-06-12  7:56       ` Joonsoo Kim
2014-06-12  7:56       ` Joonsoo Kim
2014-06-12  7:56       ` Joonsoo Kim
2014-06-12  7:56       ` Joonsoo Kim
2014-06-14  7:25 ` [PATCH v2 00/10] CMA: generalize CMA reserved area management code Aneesh Kumar K.V
2014-06-14  7:37   ` Aneesh Kumar K.V
2014-06-14  7:25   ` Aneesh Kumar K.V
2014-06-14  7:25   ` Aneesh Kumar K.V
2014-06-14  7:25   ` Aneesh Kumar K.V
2014-06-14  7:25   ` Aneesh Kumar K.V
2014-06-16  5:32   ` Joonsoo Kim
2014-06-16  5:32     ` Joonsoo Kim
2014-06-16  5:32     ` Joonsoo Kim
2014-06-16  5:32     ` Joonsoo Kim
2014-06-16  5:32     ` Joonsoo Kim
2014-06-16  7:04     ` Aneesh Kumar K.V
2014-06-16  7:16       ` Aneesh Kumar K.V
2014-06-16  7:04       ` Aneesh Kumar K.V
2014-06-16  7:04       ` Aneesh Kumar K.V
2014-06-16  7:04       ` Aneesh Kumar K.V

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20140612055219.GG12415@bbox \
    --to=minchan@kernel.org \
    --cc=agraf@suse.de \
    --cc=akpm@linux-foundation.org \
    --cc=aneesh.kumar@linux.vnet.ibm.com \
    --cc=benh@kernel.crashing.org \
    --cc=gleb@kernel.org \
    --cc=gregkh@linuxfoundation.org \
    --cc=iamjoonsoo.kim@lge.com \
    --cc=kvm-ppc@vger.kernel.org \
    --cc=kvm@vger.kernel.org \
    --cc=linux-arm-kernel@lists.infradead.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=linux@arm.linux.org.uk \
    --cc=linuxppc-dev@lists.ozlabs.org \
    --cc=m.szyprowski@samsung.com \
    --cc=mina86@mina86.com \
    --cc=paulus@samba.org \
    --cc=pbonzini@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.