All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
@ 2021-09-07 21:23 ` Minchan Kim
  0 siblings, 0 replies; 14+ messages in thread
From: Minchan Kim @ 2021-09-07 21:23 UTC (permalink / raw)
  To: Andrew Morton
  Cc: Linus Torvalds, Laura Abbott, Oliver Sang, David Hildenbrand,
	John Dias, Matthew Wilcox, Michal Hocko, Suren Baghdasaryan,
	Vlastimil Babka, LKML, linux-mm, lkp, lkp, ying.huang, feng.tang,
	zhengjun.xing, Minchan Kim, Chris Goldsworthy

kernel test robot reported the regression of fio.write_iops[1]
with [2].

Since lru_add_drain is called frequently, invalidate bh_lrus
there could increase bh_lrus cache miss ratio, which needs
more IO in the end.

This patch moves the bh_lrus invalidation from the hot path(
e.g., zap_page_range, pagevec_release) to cold path(i.e.,
lru_add_drain_all, lru_cache_disable).

[1] https://lore.kernel.org/lkml/20210520083144.GD14190@xsang-OptiPlex-9020/
[2] 8cc621d2f45d, mm: fs: invalidate BH LRU during page migration
Reviewed-by: Chris Goldsworthy <cgoldswo@codeaurora.org>
Reported-by: kernel test robot <oliver.sang@intel.com>
Signed-off-by: Minchan Kim <minchan@kernel.org>
---
* v2: https://lore.kernel.org/lkml/20210601145425.1396981-1-minchan@kernel.org/
* v1: https://lore.kernel.org/lkml/YK0oQ76zX0uVZCwQ@google.com/
 fs/buffer.c                 |  8 ++++++--
 include/linux/buffer_head.h |  4 ++--
 mm/swap.c                   | 19 ++++++++++++++++---
 3 files changed, 24 insertions(+), 7 deletions(-)

diff --git a/fs/buffer.c b/fs/buffer.c
index ab7573d72dd7..c615387aedca 100644
--- a/fs/buffer.c
+++ b/fs/buffer.c
@@ -1425,12 +1425,16 @@ void invalidate_bh_lrus(void)
 }
 EXPORT_SYMBOL_GPL(invalidate_bh_lrus);
 
-void invalidate_bh_lrus_cpu(int cpu)
+/*
+ * It's called from workqueue context so we need a bh_lru_lock to close
+ * the race with preemption/irq.
+ */
+void invalidate_bh_lrus_cpu(void)
 {
 	struct bh_lru *b;
 
 	bh_lru_lock();
-	b = per_cpu_ptr(&bh_lrus, cpu);
+	b = this_cpu_ptr(&bh_lrus);
 	__invalidate_bh_lrus(b);
 	bh_lru_unlock();
 }
diff --git a/include/linux/buffer_head.h b/include/linux/buffer_head.h
index 6486d3c19463..36f33685c8c0 100644
--- a/include/linux/buffer_head.h
+++ b/include/linux/buffer_head.h
@@ -194,7 +194,7 @@ void __breadahead_gfp(struct block_device *, sector_t block, unsigned int size,
 struct buffer_head *__bread_gfp(struct block_device *,
 				sector_t block, unsigned size, gfp_t gfp);
 void invalidate_bh_lrus(void);
-void invalidate_bh_lrus_cpu(int cpu);
+void invalidate_bh_lrus_cpu(void);
 bool has_bh_in_lru(int cpu, void *dummy);
 struct buffer_head *alloc_buffer_head(gfp_t gfp_flags);
 void free_buffer_head(struct buffer_head * bh);
@@ -408,7 +408,7 @@ static inline int inode_has_buffers(struct inode *inode) { return 0; }
 static inline void invalidate_inode_buffers(struct inode *inode) {}
 static inline int remove_inode_buffers(struct inode *inode) { return 1; }
 static inline int sync_mapping_buffers(struct address_space *mapping) { return 0; }
-static inline void invalidate_bh_lrus_cpu(int cpu) {}
+static inline void invalidate_bh_lrus_cpu(void) {}
 static inline bool has_bh_in_lru(int cpu, void *dummy) { return false; }
 #define buffer_heads_over_limit 0
 
diff --git a/mm/swap.c b/mm/swap.c
index 897200d27dd0..af3cad4e5378 100644
--- a/mm/swap.c
+++ b/mm/swap.c
@@ -620,7 +620,6 @@ void lru_add_drain_cpu(int cpu)
 		pagevec_lru_move_fn(pvec, lru_lazyfree_fn);
 
 	activate_page_drain(cpu);
-	invalidate_bh_lrus_cpu(cpu);
 }
 
 /**
@@ -703,6 +702,20 @@ void lru_add_drain(void)
 	local_unlock(&lru_pvecs.lock);
 }
 
+/*
+ * It's called from per-cpu workqueue context in SMP case so
+ * lru_add_drain_cpu and invalidate_bh_lrus_cpu should run on
+ * the same cpu. It shouldn't be a problem in !SMP case since
+ * the core is only one and the locks will disable preemption.
+ */
+static void lru_add_and_bh_lrus_drain(void)
+{
+	local_lock(&lru_pvecs.lock);
+	lru_add_drain_cpu(smp_processor_id());
+	local_unlock(&lru_pvecs.lock);
+	invalidate_bh_lrus_cpu();
+}
+
 void lru_add_drain_cpu_zone(struct zone *zone)
 {
 	local_lock(&lru_pvecs.lock);
@@ -717,7 +730,7 @@ static DEFINE_PER_CPU(struct work_struct, lru_add_drain_work);
 
 static void lru_add_drain_per_cpu(struct work_struct *dummy)
 {
-	lru_add_drain();
+	lru_add_and_bh_lrus_drain();
 }
 
 /*
@@ -858,7 +871,7 @@ void lru_cache_disable(void)
 	 */
 	__lru_add_drain_all(true);
 #else
-	lru_add_drain();
+	lru_add_and_bh_lrus_drain();
 #endif
 }
 
-- 
2.33.0.309.g3052b89438-goog


^ permalink raw reply related	[flat|nested] 14+ messages in thread

* [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
@ 2021-09-07 21:23 ` Minchan Kim
  0 siblings, 0 replies; 14+ messages in thread
From: Minchan Kim @ 2021-09-07 21:23 UTC (permalink / raw)
  To: lkp

[-- Attachment #1: Type: text/plain, Size: 4161 bytes --]

kernel test robot reported the regression of fio.write_iops[1]
with [2].

Since lru_add_drain is called frequently, invalidate bh_lrus
there could increase bh_lrus cache miss ratio, which needs
more IO in the end.

This patch moves the bh_lrus invalidation from the hot path(
e.g., zap_page_range, pagevec_release) to cold path(i.e.,
lru_add_drain_all, lru_cache_disable).

[1] https://lore.kernel.org/lkml/20210520083144.GD14190(a)xsang-OptiPlex-9020/
[2] 8cc621d2f45d, mm: fs: invalidate BH LRU during page migration
Reviewed-by: Chris Goldsworthy <cgoldswo@codeaurora.org>
Reported-by: kernel test robot <oliver.sang@intel.com>
Signed-off-by: Minchan Kim <minchan@kernel.org>
---
* v2: https://lore.kernel.org/lkml/20210601145425.1396981-1-minchan(a)kernel.org/
* v1: https://lore.kernel.org/lkml/YK0oQ76zX0uVZCwQ(a)google.com/
 fs/buffer.c                 |  8 ++++++--
 include/linux/buffer_head.h |  4 ++--
 mm/swap.c                   | 19 ++++++++++++++++---
 3 files changed, 24 insertions(+), 7 deletions(-)

diff --git a/fs/buffer.c b/fs/buffer.c
index ab7573d72dd7..c615387aedca 100644
--- a/fs/buffer.c
+++ b/fs/buffer.c
@@ -1425,12 +1425,16 @@ void invalidate_bh_lrus(void)
 }
 EXPORT_SYMBOL_GPL(invalidate_bh_lrus);
 
-void invalidate_bh_lrus_cpu(int cpu)
+/*
+ * It's called from workqueue context so we need a bh_lru_lock to close
+ * the race with preemption/irq.
+ */
+void invalidate_bh_lrus_cpu(void)
 {
 	struct bh_lru *b;
 
 	bh_lru_lock();
-	b = per_cpu_ptr(&bh_lrus, cpu);
+	b = this_cpu_ptr(&bh_lrus);
 	__invalidate_bh_lrus(b);
 	bh_lru_unlock();
 }
diff --git a/include/linux/buffer_head.h b/include/linux/buffer_head.h
index 6486d3c19463..36f33685c8c0 100644
--- a/include/linux/buffer_head.h
+++ b/include/linux/buffer_head.h
@@ -194,7 +194,7 @@ void __breadahead_gfp(struct block_device *, sector_t block, unsigned int size,
 struct buffer_head *__bread_gfp(struct block_device *,
 				sector_t block, unsigned size, gfp_t gfp);
 void invalidate_bh_lrus(void);
-void invalidate_bh_lrus_cpu(int cpu);
+void invalidate_bh_lrus_cpu(void);
 bool has_bh_in_lru(int cpu, void *dummy);
 struct buffer_head *alloc_buffer_head(gfp_t gfp_flags);
 void free_buffer_head(struct buffer_head * bh);
@@ -408,7 +408,7 @@ static inline int inode_has_buffers(struct inode *inode) { return 0; }
 static inline void invalidate_inode_buffers(struct inode *inode) {}
 static inline int remove_inode_buffers(struct inode *inode) { return 1; }
 static inline int sync_mapping_buffers(struct address_space *mapping) { return 0; }
-static inline void invalidate_bh_lrus_cpu(int cpu) {}
+static inline void invalidate_bh_lrus_cpu(void) {}
 static inline bool has_bh_in_lru(int cpu, void *dummy) { return false; }
 #define buffer_heads_over_limit 0
 
diff --git a/mm/swap.c b/mm/swap.c
index 897200d27dd0..af3cad4e5378 100644
--- a/mm/swap.c
+++ b/mm/swap.c
@@ -620,7 +620,6 @@ void lru_add_drain_cpu(int cpu)
 		pagevec_lru_move_fn(pvec, lru_lazyfree_fn);
 
 	activate_page_drain(cpu);
-	invalidate_bh_lrus_cpu(cpu);
 }
 
 /**
@@ -703,6 +702,20 @@ void lru_add_drain(void)
 	local_unlock(&lru_pvecs.lock);
 }
 
+/*
+ * It's called from per-cpu workqueue context in SMP case so
+ * lru_add_drain_cpu and invalidate_bh_lrus_cpu should run on
+ * the same cpu. It shouldn't be a problem in !SMP case since
+ * the core is only one and the locks will disable preemption.
+ */
+static void lru_add_and_bh_lrus_drain(void)
+{
+	local_lock(&lru_pvecs.lock);
+	lru_add_drain_cpu(smp_processor_id());
+	local_unlock(&lru_pvecs.lock);
+	invalidate_bh_lrus_cpu();
+}
+
 void lru_add_drain_cpu_zone(struct zone *zone)
 {
 	local_lock(&lru_pvecs.lock);
@@ -717,7 +730,7 @@ static DEFINE_PER_CPU(struct work_struct, lru_add_drain_work);
 
 static void lru_add_drain_per_cpu(struct work_struct *dummy)
 {
-	lru_add_drain();
+	lru_add_and_bh_lrus_drain();
 }
 
 /*
@@ -858,7 +871,7 @@ void lru_cache_disable(void)
 	 */
 	__lru_add_drain_all(true);
 #else
-	lru_add_drain();
+	lru_add_and_bh_lrus_drain();
 #endif
 }
 
-- 
2.33.0.309.g3052b89438-goog

^ permalink raw reply related	[flat|nested] 14+ messages in thread

* Re: [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
  2021-09-07 21:23 ` Minchan Kim
@ 2021-09-20 22:33   ` Minchan Kim
  -1 siblings, 0 replies; 14+ messages in thread
From: Minchan Kim @ 2021-09-20 22:33 UTC (permalink / raw)
  To: Andrew Morton
  Cc: Linus Torvalds, Laura Abbott, Oliver Sang, David Hildenbrand,
	John Dias, Matthew Wilcox, Michal Hocko, Suren Baghdasaryan,
	Vlastimil Babka, LKML, linux-mm, lkp, lkp, ying.huang, feng.tang,
	zhengjun.xing, Chris Goldsworthy

Andrew, Could you take a look?

On Tue, Sep 07, 2021 at 02:23:47PM -0700, Minchan Kim wrote:
> kernel test robot reported the regression of fio.write_iops[1]
> with [2].
> 
> Since lru_add_drain is called frequently, invalidate bh_lrus
> there could increase bh_lrus cache miss ratio, which needs
> more IO in the end.
> 
> This patch moves the bh_lrus invalidation from the hot path(
> e.g., zap_page_range, pagevec_release) to cold path(i.e.,
> lru_add_drain_all, lru_cache_disable).
> 
> [1] https://lore.kernel.org/lkml/20210520083144.GD14190@xsang-OptiPlex-9020/
> [2] 8cc621d2f45d, mm: fs: invalidate BH LRU during page migration
> Reviewed-by: Chris Goldsworthy <cgoldswo@codeaurora.org>
> Reported-by: kernel test robot <oliver.sang@intel.com>
> Signed-off-by: Minchan Kim <minchan@kernel.org>
> ---
> * v2: https://lore.kernel.org/lkml/20210601145425.1396981-1-minchan@kernel.org/
> * v1: https://lore.kernel.org/lkml/YK0oQ76zX0uVZCwQ@google.com/
>  fs/buffer.c                 |  8 ++++++--
>  include/linux/buffer_head.h |  4 ++--
>  mm/swap.c                   | 19 ++++++++++++++++---
>  3 files changed, 24 insertions(+), 7 deletions(-)
> 
> diff --git a/fs/buffer.c b/fs/buffer.c
> index ab7573d72dd7..c615387aedca 100644
> --- a/fs/buffer.c
> +++ b/fs/buffer.c
> @@ -1425,12 +1425,16 @@ void invalidate_bh_lrus(void)
>  }
>  EXPORT_SYMBOL_GPL(invalidate_bh_lrus);
>  
> -void invalidate_bh_lrus_cpu(int cpu)
> +/*
> + * It's called from workqueue context so we need a bh_lru_lock to close
> + * the race with preemption/irq.
> + */
> +void invalidate_bh_lrus_cpu(void)
>  {
>  	struct bh_lru *b;
>  
>  	bh_lru_lock();
> -	b = per_cpu_ptr(&bh_lrus, cpu);
> +	b = this_cpu_ptr(&bh_lrus);
>  	__invalidate_bh_lrus(b);
>  	bh_lru_unlock();
>  }
> diff --git a/include/linux/buffer_head.h b/include/linux/buffer_head.h
> index 6486d3c19463..36f33685c8c0 100644
> --- a/include/linux/buffer_head.h
> +++ b/include/linux/buffer_head.h
> @@ -194,7 +194,7 @@ void __breadahead_gfp(struct block_device *, sector_t block, unsigned int size,
>  struct buffer_head *__bread_gfp(struct block_device *,
>  				sector_t block, unsigned size, gfp_t gfp);
>  void invalidate_bh_lrus(void);
> -void invalidate_bh_lrus_cpu(int cpu);
> +void invalidate_bh_lrus_cpu(void);
>  bool has_bh_in_lru(int cpu, void *dummy);
>  struct buffer_head *alloc_buffer_head(gfp_t gfp_flags);
>  void free_buffer_head(struct buffer_head * bh);
> @@ -408,7 +408,7 @@ static inline int inode_has_buffers(struct inode *inode) { return 0; }
>  static inline void invalidate_inode_buffers(struct inode *inode) {}
>  static inline int remove_inode_buffers(struct inode *inode) { return 1; }
>  static inline int sync_mapping_buffers(struct address_space *mapping) { return 0; }
> -static inline void invalidate_bh_lrus_cpu(int cpu) {}
> +static inline void invalidate_bh_lrus_cpu(void) {}
>  static inline bool has_bh_in_lru(int cpu, void *dummy) { return false; }
>  #define buffer_heads_over_limit 0
>  
> diff --git a/mm/swap.c b/mm/swap.c
> index 897200d27dd0..af3cad4e5378 100644
> --- a/mm/swap.c
> +++ b/mm/swap.c
> @@ -620,7 +620,6 @@ void lru_add_drain_cpu(int cpu)
>  		pagevec_lru_move_fn(pvec, lru_lazyfree_fn);
>  
>  	activate_page_drain(cpu);
> -	invalidate_bh_lrus_cpu(cpu);
>  }
>  
>  /**
> @@ -703,6 +702,20 @@ void lru_add_drain(void)
>  	local_unlock(&lru_pvecs.lock);
>  }
>  
> +/*
> + * It's called from per-cpu workqueue context in SMP case so
> + * lru_add_drain_cpu and invalidate_bh_lrus_cpu should run on
> + * the same cpu. It shouldn't be a problem in !SMP case since
> + * the core is only one and the locks will disable preemption.
> + */
> +static void lru_add_and_bh_lrus_drain(void)
> +{
> +	local_lock(&lru_pvecs.lock);
> +	lru_add_drain_cpu(smp_processor_id());
> +	local_unlock(&lru_pvecs.lock);
> +	invalidate_bh_lrus_cpu();
> +}
> +
>  void lru_add_drain_cpu_zone(struct zone *zone)
>  {
>  	local_lock(&lru_pvecs.lock);
> @@ -717,7 +730,7 @@ static DEFINE_PER_CPU(struct work_struct, lru_add_drain_work);
>  
>  static void lru_add_drain_per_cpu(struct work_struct *dummy)
>  {
> -	lru_add_drain();
> +	lru_add_and_bh_lrus_drain();
>  }
>  
>  /*
> @@ -858,7 +871,7 @@ void lru_cache_disable(void)
>  	 */
>  	__lru_add_drain_all(true);
>  #else
> -	lru_add_drain();
> +	lru_add_and_bh_lrus_drain();
>  #endif
>  }
>  
> -- 
> 2.33.0.309.g3052b89438-goog
> 

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
@ 2021-09-20 22:33   ` Minchan Kim
  0 siblings, 0 replies; 14+ messages in thread
From: Minchan Kim @ 2021-09-20 22:33 UTC (permalink / raw)
  To: lkp

[-- Attachment #1: Type: text/plain, Size: 4503 bytes --]

Andrew, Could you take a look?

On Tue, Sep 07, 2021 at 02:23:47PM -0700, Minchan Kim wrote:
> kernel test robot reported the regression of fio.write_iops[1]
> with [2].
> 
> Since lru_add_drain is called frequently, invalidate bh_lrus
> there could increase bh_lrus cache miss ratio, which needs
> more IO in the end.
> 
> This patch moves the bh_lrus invalidation from the hot path(
> e.g., zap_page_range, pagevec_release) to cold path(i.e.,
> lru_add_drain_all, lru_cache_disable).
> 
> [1] https://lore.kernel.org/lkml/20210520083144.GD14190(a)xsang-OptiPlex-9020/
> [2] 8cc621d2f45d, mm: fs: invalidate BH LRU during page migration
> Reviewed-by: Chris Goldsworthy <cgoldswo@codeaurora.org>
> Reported-by: kernel test robot <oliver.sang@intel.com>
> Signed-off-by: Minchan Kim <minchan@kernel.org>
> ---
> * v2: https://lore.kernel.org/lkml/20210601145425.1396981-1-minchan(a)kernel.org/
> * v1: https://lore.kernel.org/lkml/YK0oQ76zX0uVZCwQ(a)google.com/
>  fs/buffer.c                 |  8 ++++++--
>  include/linux/buffer_head.h |  4 ++--
>  mm/swap.c                   | 19 ++++++++++++++++---
>  3 files changed, 24 insertions(+), 7 deletions(-)
> 
> diff --git a/fs/buffer.c b/fs/buffer.c
> index ab7573d72dd7..c615387aedca 100644
> --- a/fs/buffer.c
> +++ b/fs/buffer.c
> @@ -1425,12 +1425,16 @@ void invalidate_bh_lrus(void)
>  }
>  EXPORT_SYMBOL_GPL(invalidate_bh_lrus);
>  
> -void invalidate_bh_lrus_cpu(int cpu)
> +/*
> + * It's called from workqueue context so we need a bh_lru_lock to close
> + * the race with preemption/irq.
> + */
> +void invalidate_bh_lrus_cpu(void)
>  {
>  	struct bh_lru *b;
>  
>  	bh_lru_lock();
> -	b = per_cpu_ptr(&bh_lrus, cpu);
> +	b = this_cpu_ptr(&bh_lrus);
>  	__invalidate_bh_lrus(b);
>  	bh_lru_unlock();
>  }
> diff --git a/include/linux/buffer_head.h b/include/linux/buffer_head.h
> index 6486d3c19463..36f33685c8c0 100644
> --- a/include/linux/buffer_head.h
> +++ b/include/linux/buffer_head.h
> @@ -194,7 +194,7 @@ void __breadahead_gfp(struct block_device *, sector_t block, unsigned int size,
>  struct buffer_head *__bread_gfp(struct block_device *,
>  				sector_t block, unsigned size, gfp_t gfp);
>  void invalidate_bh_lrus(void);
> -void invalidate_bh_lrus_cpu(int cpu);
> +void invalidate_bh_lrus_cpu(void);
>  bool has_bh_in_lru(int cpu, void *dummy);
>  struct buffer_head *alloc_buffer_head(gfp_t gfp_flags);
>  void free_buffer_head(struct buffer_head * bh);
> @@ -408,7 +408,7 @@ static inline int inode_has_buffers(struct inode *inode) { return 0; }
>  static inline void invalidate_inode_buffers(struct inode *inode) {}
>  static inline int remove_inode_buffers(struct inode *inode) { return 1; }
>  static inline int sync_mapping_buffers(struct address_space *mapping) { return 0; }
> -static inline void invalidate_bh_lrus_cpu(int cpu) {}
> +static inline void invalidate_bh_lrus_cpu(void) {}
>  static inline bool has_bh_in_lru(int cpu, void *dummy) { return false; }
>  #define buffer_heads_over_limit 0
>  
> diff --git a/mm/swap.c b/mm/swap.c
> index 897200d27dd0..af3cad4e5378 100644
> --- a/mm/swap.c
> +++ b/mm/swap.c
> @@ -620,7 +620,6 @@ void lru_add_drain_cpu(int cpu)
>  		pagevec_lru_move_fn(pvec, lru_lazyfree_fn);
>  
>  	activate_page_drain(cpu);
> -	invalidate_bh_lrus_cpu(cpu);
>  }
>  
>  /**
> @@ -703,6 +702,20 @@ void lru_add_drain(void)
>  	local_unlock(&lru_pvecs.lock);
>  }
>  
> +/*
> + * It's called from per-cpu workqueue context in SMP case so
> + * lru_add_drain_cpu and invalidate_bh_lrus_cpu should run on
> + * the same cpu. It shouldn't be a problem in !SMP case since
> + * the core is only one and the locks will disable preemption.
> + */
> +static void lru_add_and_bh_lrus_drain(void)
> +{
> +	local_lock(&lru_pvecs.lock);
> +	lru_add_drain_cpu(smp_processor_id());
> +	local_unlock(&lru_pvecs.lock);
> +	invalidate_bh_lrus_cpu();
> +}
> +
>  void lru_add_drain_cpu_zone(struct zone *zone)
>  {
>  	local_lock(&lru_pvecs.lock);
> @@ -717,7 +730,7 @@ static DEFINE_PER_CPU(struct work_struct, lru_add_drain_work);
>  
>  static void lru_add_drain_per_cpu(struct work_struct *dummy)
>  {
> -	lru_add_drain();
> +	lru_add_and_bh_lrus_drain();
>  }
>  
>  /*
> @@ -858,7 +871,7 @@ void lru_cache_disable(void)
>  	 */
>  	__lru_add_drain_all(true);
>  #else
> -	lru_add_drain();
> +	lru_add_and_bh_lrus_drain();
>  #endif
>  }
>  
> -- 
> 2.33.0.309.g3052b89438-goog
> 

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
  2021-09-07 21:23 ` Minchan Kim
  (?)
@ 2021-09-20 23:29   ` Linus Torvalds
  -1 siblings, 0 replies; 14+ messages in thread
From: Linus Torvalds @ 2021-09-20 23:29 UTC (permalink / raw)
  To: Minchan Kim
  Cc: Andrew Morton, Laura Abbott, Oliver Sang, David Hildenbrand,
	John Dias, Matthew Wilcox, Michal Hocko, Suren Baghdasaryan,
	Vlastimil Babka, LKML, linux-mm, lkp, kernel test robot, Huang,
	Ying, Feng Tang, zhengjun.xing, Chris Goldsworthy

On Tue, Sep 7, 2021 at 2:24 PM Minchan Kim <minchan@kernel.org> wrote:
>
> kernel test robot reported the regression of fio.write_iops[1]
> with [2].
>
> Since lru_add_drain is called frequently, invalidate bh_lrus
> there could increase bh_lrus cache miss ratio, which needs
> more IO in the end.
>
> This patch moves the bh_lrus invalidation from the hot path(
> e.g., zap_page_range, pagevec_release) to cold path(i.e.,
> lru_add_drain_all, lru_cache_disable).

Was this confirmed to fix the regression?

I only see the "tested with 5.14" that the regression was still there

   https://lore.kernel.org/lkml/034fc860-d0d0-0c61-09d2-3c41c4f020c6@intel.com/

I don't see a confirmation that this patch fixed it.

It looks likely, but if you have the confirmation somewhere, it would
help to link that too.

Or did I miss it?

           Linus


^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
@ 2021-09-20 23:29   ` Linus Torvalds
  0 siblings, 0 replies; 14+ messages in thread
From: Linus Torvalds @ 2021-09-20 23:29 UTC (permalink / raw)
  To: Minchan Kim
  Cc: Andrew Morton, Laura Abbott, Oliver Sang, David Hildenbrand,
	John Dias, Matthew Wilcox, Michal Hocko, Suren Baghdasaryan,
	Vlastimil Babka, LKML, linux-mm, lkp, kernel test robot, Huang,
	Ying, Feng Tang, zhengjun.xing, Chris Goldsworthy

On Tue, Sep 7, 2021 at 2:24 PM Minchan Kim <minchan@kernel.org> wrote:
>
> kernel test robot reported the regression of fio.write_iops[1]
> with [2].
>
> Since lru_add_drain is called frequently, invalidate bh_lrus
> there could increase bh_lrus cache miss ratio, which needs
> more IO in the end.
>
> This patch moves the bh_lrus invalidation from the hot path(
> e.g., zap_page_range, pagevec_release) to cold path(i.e.,
> lru_add_drain_all, lru_cache_disable).

Was this confirmed to fix the regression?

I only see the "tested with 5.14" that the regression was still there

   https://lore.kernel.org/lkml/034fc860-d0d0-0c61-09d2-3c41c4f020c6@intel.com/

I don't see a confirmation that this patch fixed it.

It looks likely, but if you have the confirmation somewhere, it would
help to link that too.

Or did I miss it?

           Linus

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
@ 2021-09-20 23:29   ` Linus Torvalds
  0 siblings, 0 replies; 14+ messages in thread
From: Linus Torvalds @ 2021-09-20 23:29 UTC (permalink / raw)
  To: lkp

[-- Attachment #1: Type: text/plain, Size: 871 bytes --]

On Tue, Sep 7, 2021 at 2:24 PM Minchan Kim <minchan@kernel.org> wrote:
>
> kernel test robot reported the regression of fio.write_iops[1]
> with [2].
>
> Since lru_add_drain is called frequently, invalidate bh_lrus
> there could increase bh_lrus cache miss ratio, which needs
> more IO in the end.
>
> This patch moves the bh_lrus invalidation from the hot path(
> e.g., zap_page_range, pagevec_release) to cold path(i.e.,
> lru_add_drain_all, lru_cache_disable).

Was this confirmed to fix the regression?

I only see the "tested with 5.14" that the regression was still there

   https://lore.kernel.org/lkml/034fc860-d0d0-0c61-09d2-3c41c4f020c6(a)intel.com/

I don't see a confirmation that this patch fixed it.

It looks likely, but if you have the confirmation somewhere, it would
help to link that too.

Or did I miss it?

           Linus

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
  2021-09-20 23:29   ` Linus Torvalds
@ 2021-09-20 23:50     ` Minchan Kim
  -1 siblings, 0 replies; 14+ messages in thread
From: Minchan Kim @ 2021-09-20 23:50 UTC (permalink / raw)
  To: Linus Torvalds, zhengjun.xing
  Cc: Andrew Morton, Laura Abbott, Oliver Sang, David Hildenbrand,
	John Dias, Matthew Wilcox, Michal Hocko, Suren Baghdasaryan,
	Vlastimil Babka, LKML, linux-mm, lkp, kernel test robot, Huang,
	Ying, Feng Tang, zhengjun.xing, Chris Goldsworthy

On Mon, Sep 20, 2021 at 04:29:52PM -0700, Linus Torvalds wrote:
> On Tue, Sep 7, 2021 at 2:24 PM Minchan Kim <minchan@kernel.org> wrote:
> >
> > kernel test robot reported the regression of fio.write_iops[1]
> > with [2].
> >
> > Since lru_add_drain is called frequently, invalidate bh_lrus
> > there could increase bh_lrus cache miss ratio, which needs
> > more IO in the end.
> >
> > This patch moves the bh_lrus invalidation from the hot path(
> > e.g., zap_page_range, pagevec_release) to cold path(i.e.,
> > lru_add_drain_all, lru_cache_disable).
> 
> Was this confirmed to fix the regression?
> 
> I only see the "tested with 5.14" that the regression was still there
> 
>    https://lore.kernel.org/lkml/034fc860-d0d0-0c61-09d2-3c41c4f020c6@intel.com/
> 
> I don't see a confirmation that this patch fixed it.
> 
> It looks likely, but if you have the confirmation somewhere, it would
> help to link that too.
> 
> Or did I miss it?

I have no idea why I couldn't find the reply from lore.kernel.org/lkml/
The message id was 89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com
in the thread(https://lore.kernel.org/lkml/20210520083144.GD14190@xsang-OptiPlex-9020/)

Xing, Zhengjun zhengjun.xing@intel.com confirmed 

Quote -
```
Hi Minchan,

...

I test the patch, the regression reduced to -2.9%.
```
Since then, I modified the patch a little more and couldn't find the
message from the public URL. Thus, do we need to confirm it again against
latest kernel?

Xing, could you confirm whether this patch fixes the regression?

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
@ 2021-09-20 23:50     ` Minchan Kim
  0 siblings, 0 replies; 14+ messages in thread
From: Minchan Kim @ 2021-09-20 23:50 UTC (permalink / raw)
  To: lkp

[-- Attachment #1: Type: text/plain, Size: 1580 bytes --]

On Mon, Sep 20, 2021 at 04:29:52PM -0700, Linus Torvalds wrote:
> On Tue, Sep 7, 2021 at 2:24 PM Minchan Kim <minchan@kernel.org> wrote:
> >
> > kernel test robot reported the regression of fio.write_iops[1]
> > with [2].
> >
> > Since lru_add_drain is called frequently, invalidate bh_lrus
> > there could increase bh_lrus cache miss ratio, which needs
> > more IO in the end.
> >
> > This patch moves the bh_lrus invalidation from the hot path(
> > e.g., zap_page_range, pagevec_release) to cold path(i.e.,
> > lru_add_drain_all, lru_cache_disable).
> 
> Was this confirmed to fix the regression?
> 
> I only see the "tested with 5.14" that the regression was still there
> 
>    https://lore.kernel.org/lkml/034fc860-d0d0-0c61-09d2-3c41c4f020c6(a)intel.com/
> 
> I don't see a confirmation that this patch fixed it.
> 
> It looks likely, but if you have the confirmation somewhere, it would
> help to link that too.
> 
> Or did I miss it?

I have no idea why I couldn't find the reply from lore.kernel.org/lkml/
The message id was 89e2b66b-c706-c020-bff5-b815dcd5c461(a)intel.com
in the thread(https://lore.kernel.org/lkml/20210520083144.GD14190(a)xsang-OptiPlex-9020/)

Xing, Zhengjun zhengjun.xing(a)intel.com confirmed 

Quote -
```
Hi Minchan,

...

I test the patch, the regression reduced to -2.9%.
```
Since then, I modified the patch a little more and couldn't find the
message from the public URL. Thus, do we need to confirm it again against
latest kernel?

Xing, could you confirm whether this patch fixes the regression?

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
  2021-09-20 23:50     ` Minchan Kim
  (?)
@ 2021-09-21  0:23       ` Linus Torvalds
  -1 siblings, 0 replies; 14+ messages in thread
From: Linus Torvalds @ 2021-09-21  0:23 UTC (permalink / raw)
  To: Minchan Kim, Konstantin Ryabitsev
  Cc: zhengjun.xing, Andrew Morton, Laura Abbott, Oliver Sang,
	David Hildenbrand, John Dias, Matthew Wilcox, Michal Hocko,
	Suren Baghdasaryan, Vlastimil Babka, LKML, linux-mm, lkp,
	kernel test robot, Huang, Ying, Feng Tang, Chris Goldsworthy

On Mon, Sep 20, 2021 at 4:50 PM Minchan Kim <minchan@kernel.org> wrote:
>
> I have no idea why I couldn't find the reply from lore.kernel.org/lkml/
> The message id was 89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com
> in the thread(https://lore.kernel.org/lkml/20210520083144.GD14190@xsang-OptiPlex-9020/)

Hmm. That message-id isn't found anywhere.

Maybe it was sent just to you personally?

[ Goes off and looks ]

Nope, I can see it in my mails too, and yeah, I see

  To: Minchan Kim <minchan@kernel.org>, kernel test robot
<oliver.sang@intel.com>
  Cc: Linus Torvalds <torvalds@linux-foundation.org>, Chris
Goldsworthy <cgoldswo@codeaurora.org>, Laura Abbott
<labbott@kernel.org>, David Hildenbrand <david@redhat.com>, John Dias
<joaodias@google.com>, Matthew Wilcox <willy@infradead.org>, Michal
Hocko <mhocko@suse.com>, Suren Baghdasaryan <surenb@google.com>,
Vlastimil Babka <vbabka@suse.cz>, Andrew Morton
<akpm@linux-foundation.org>, LKML <linux-kernel@vger.kernel.org>,
lkp@lists.01.org, lkp@intel.com
  References: <20210520083144.GD14190@xsang-OptiPlex-9020>
<YKasEeXCr9R5yzCr@google.com>
  From: "Xing, Zhengjun" <zhengjun.xing@intel.com>
  Message-ID: <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com>

but lore has no idea about it:

  https://lore.kernel.org/all/89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com/

just says "Message-ID <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com>
not found".

Strange.

Something presumably caused the list to drop that email.

           Linus


^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
@ 2021-09-21  0:23       ` Linus Torvalds
  0 siblings, 0 replies; 14+ messages in thread
From: Linus Torvalds @ 2021-09-21  0:23 UTC (permalink / raw)
  To: Minchan Kim, Konstantin Ryabitsev
  Cc: zhengjun.xing, Andrew Morton, Laura Abbott, Oliver Sang,
	David Hildenbrand, John Dias, Matthew Wilcox, Michal Hocko,
	Suren Baghdasaryan, Vlastimil Babka, LKML, linux-mm, lkp,
	kernel test robot, Huang, Ying, Feng Tang, Chris Goldsworthy

On Mon, Sep 20, 2021 at 4:50 PM Minchan Kim <minchan@kernel.org> wrote:
>
> I have no idea why I couldn't find the reply from lore.kernel.org/lkml/
> The message id was 89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com
> in the thread(https://lore.kernel.org/lkml/20210520083144.GD14190@xsang-OptiPlex-9020/)

Hmm. That message-id isn't found anywhere.

Maybe it was sent just to you personally?

[ Goes off and looks ]

Nope, I can see it in my mails too, and yeah, I see

  To: Minchan Kim <minchan@kernel.org>, kernel test robot
<oliver.sang@intel.com>
  Cc: Linus Torvalds <torvalds@linux-foundation.org>, Chris
Goldsworthy <cgoldswo@codeaurora.org>, Laura Abbott
<labbott@kernel.org>, David Hildenbrand <david@redhat.com>, John Dias
<joaodias@google.com>, Matthew Wilcox <willy@infradead.org>, Michal
Hocko <mhocko@suse.com>, Suren Baghdasaryan <surenb@google.com>,
Vlastimil Babka <vbabka@suse.cz>, Andrew Morton
<akpm@linux-foundation.org>, LKML <linux-kernel@vger.kernel.org>,
lkp@lists.01.org, lkp@intel.com
  References: <20210520083144.GD14190@xsang-OptiPlex-9020>
<YKasEeXCr9R5yzCr@google.com>
  From: "Xing, Zhengjun" <zhengjun.xing@intel.com>
  Message-ID: <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com>

but lore has no idea about it:

  https://lore.kernel.org/all/89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com/

just says "Message-ID <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com>
not found".

Strange.

Something presumably caused the list to drop that email.

           Linus

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
@ 2021-09-21  0:23       ` Linus Torvalds
  0 siblings, 0 replies; 14+ messages in thread
From: Linus Torvalds @ 2021-09-21  0:23 UTC (permalink / raw)
  To: lkp

[-- Attachment #1: Type: text/plain, Size: 1547 bytes --]

On Mon, Sep 20, 2021 at 4:50 PM Minchan Kim <minchan@kernel.org> wrote:
>
> I have no idea why I couldn't find the reply from lore.kernel.org/lkml/
> The message id was 89e2b66b-c706-c020-bff5-b815dcd5c461(a)intel.com
> in the thread(https://lore.kernel.org/lkml/20210520083144.GD14190(a)xsang-OptiPlex-9020/)

Hmm. That message-id isn't found anywhere.

Maybe it was sent just to you personally?

[ Goes off and looks ]

Nope, I can see it in my mails too, and yeah, I see

  To: Minchan Kim <minchan@kernel.org>, kernel test robot
<oliver.sang@intel.com>
  Cc: Linus Torvalds <torvalds@linux-foundation.org>, Chris
Goldsworthy <cgoldswo@codeaurora.org>, Laura Abbott
<labbott@kernel.org>, David Hildenbrand <david@redhat.com>, John Dias
<joaodias@google.com>, Matthew Wilcox <willy@infradead.org>, Michal
Hocko <mhocko@suse.com>, Suren Baghdasaryan <surenb@google.com>,
Vlastimil Babka <vbabka@suse.cz>, Andrew Morton
<akpm@linux-foundation.org>, LKML <linux-kernel@vger.kernel.org>,
lkp(a)lists.01.org, lkp(a)intel.com
  References: <20210520083144.GD14190@xsang-OptiPlex-9020>
<YKasEeXCr9R5yzCr@google.com>
  From: "Xing, Zhengjun" <zhengjun.xing@intel.com>
  Message-ID: <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com>

but lore has no idea about it:

  https://lore.kernel.org/all/89e2b66b-c706-c020-bff5-b815dcd5c461(a)intel.com/

just says "Message-ID <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com>
not found".

Strange.

Something presumably caused the list to drop that email.

           Linus

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
  2021-09-21  0:23       ` Linus Torvalds
@ 2021-09-21  1:03         ` Konstantin Ryabitsev
  -1 siblings, 0 replies; 14+ messages in thread
From: Konstantin Ryabitsev @ 2021-09-21  1:03 UTC (permalink / raw)
  To: Linus Torvalds, warthog9
  Cc: Minchan Kim, zhengjun.xing, Andrew Morton, Laura Abbott,
	Oliver Sang, David Hildenbrand, John Dias, Matthew Wilcox,
	Michal Hocko, Suren Baghdasaryan, Vlastimil Babka, LKML,
	linux-mm, lkp, kernel test robot, Huang, Ying, Feng Tang,
	Chris Goldsworthy

On Mon, Sep 20, 2021 at 05:23:38PM -0700, Linus Torvalds wrote:
> Hmm. That message-id isn't found anywhere.
> 
> Maybe it was sent just to you personally?
> 
> [ Goes off and looks ]
> 
> Nope, I can see it in my mails too, and yeah, I see
> 
>   To: Minchan Kim <minchan@kernel.org>, kernel test robot
> <oliver.sang@intel.com>
>   Cc: Linus Torvalds <torvalds@linux-foundation.org>, Chris
> Goldsworthy <cgoldswo@codeaurora.org>, Laura Abbott
> <labbott@kernel.org>, David Hildenbrand <david@redhat.com>, John Dias
> <joaodias@google.com>, Matthew Wilcox <willy@infradead.org>, Michal
> Hocko <mhocko@suse.com>, Suren Baghdasaryan <surenb@google.com>,
> Vlastimil Babka <vbabka@suse.cz>, Andrew Morton
> <akpm@linux-foundation.org>, LKML <linux-kernel@vger.kernel.org>,
> lkp@lists.01.org, lkp@intel.com
>   References: <20210520083144.GD14190@xsang-OptiPlex-9020>
> <YKasEeXCr9R5yzCr@google.com>
>   From: "Xing, Zhengjun" <zhengjun.xing@intel.com>
>   Message-ID: <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com>
> 
> but lore has no idea about it:
> 
>   https://lore.kernel.org/all/89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com/
> 
> just says "Message-ID <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com>
> not found".
> 
> Strange.
> 
> Something presumably caused the list to drop that email.

I think vger is backed up. I've flagged this with John, hopefully he'll be
able to figure out what's keeping things from hitting the archives.

-K

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH v3] mm: fs: invalidate bh_lrus for only cold path
@ 2021-09-21  1:03         ` Konstantin Ryabitsev
  0 siblings, 0 replies; 14+ messages in thread
From: Konstantin Ryabitsev @ 2021-09-21  1:03 UTC (permalink / raw)
  To: lkp

[-- Attachment #1: Type: text/plain, Size: 1494 bytes --]

On Mon, Sep 20, 2021 at 05:23:38PM -0700, Linus Torvalds wrote:
> Hmm. That message-id isn't found anywhere.
> 
> Maybe it was sent just to you personally?
> 
> [ Goes off and looks ]
> 
> Nope, I can see it in my mails too, and yeah, I see
> 
>   To: Minchan Kim <minchan@kernel.org>, kernel test robot
> <oliver.sang@intel.com>
>   Cc: Linus Torvalds <torvalds@linux-foundation.org>, Chris
> Goldsworthy <cgoldswo@codeaurora.org>, Laura Abbott
> <labbott@kernel.org>, David Hildenbrand <david@redhat.com>, John Dias
> <joaodias@google.com>, Matthew Wilcox <willy@infradead.org>, Michal
> Hocko <mhocko@suse.com>, Suren Baghdasaryan <surenb@google.com>,
> Vlastimil Babka <vbabka@suse.cz>, Andrew Morton
> <akpm@linux-foundation.org>, LKML <linux-kernel@vger.kernel.org>,
> lkp(a)lists.01.org, lkp(a)intel.com
>   References: <20210520083144.GD14190@xsang-OptiPlex-9020>
> <YKasEeXCr9R5yzCr@google.com>
>   From: "Xing, Zhengjun" <zhengjun.xing@intel.com>
>   Message-ID: <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com>
> 
> but lore has no idea about it:
> 
>   https://lore.kernel.org/all/89e2b66b-c706-c020-bff5-b815dcd5c461(a)intel.com/
> 
> just says "Message-ID <89e2b66b-c706-c020-bff5-b815dcd5c461@intel.com>
> not found".
> 
> Strange.
> 
> Something presumably caused the list to drop that email.

I think vger is backed up. I've flagged this with John, hopefully he'll be
able to figure out what's keeping things from hitting the archives.

-K

^ permalink raw reply	[flat|nested] 14+ messages in thread

end of thread, other threads:[~2021-09-21  2:10 UTC | newest]

Thread overview: 14+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-09-07 21:23 [PATCH v3] mm: fs: invalidate bh_lrus for only cold path Minchan Kim
2021-09-07 21:23 ` Minchan Kim
2021-09-20 22:33 ` Minchan Kim
2021-09-20 22:33   ` Minchan Kim
2021-09-20 23:29 ` Linus Torvalds
2021-09-20 23:29   ` Linus Torvalds
2021-09-20 23:29   ` Linus Torvalds
2021-09-20 23:50   ` Minchan Kim
2021-09-20 23:50     ` Minchan Kim
2021-09-21  0:23     ` Linus Torvalds
2021-09-21  0:23       ` Linus Torvalds
2021-09-21  0:23       ` Linus Torvalds
2021-09-21  1:03       ` Konstantin Ryabitsev
2021-09-21  1:03         ` Konstantin Ryabitsev

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.