linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [PATCH] mm/swap: piggyback lru_add_drain_all() calls
@ 2019-10-04 10:11 Konstantin Khlebnikov
  2019-10-04 12:10 ` Matthew Wilcox
  0 siblings, 1 reply; 6+ messages in thread
From: Konstantin Khlebnikov @ 2019-10-04 10:11 UTC (permalink / raw)
  To: linux-mm, Andrew Morton; +Cc: linux-kernel

This is very slow operation. There is no reason to do it again if somebody
else already drained all per-cpu vectors after we waited for lock.

Signed-off-by: Konstantin Khlebnikov <khlebnikov@yandex-team.ru>
---
 mm/swap.c |   13 ++++++++++++-
 1 file changed, 12 insertions(+), 1 deletion(-)

diff --git a/mm/swap.c b/mm/swap.c
index 38c3fa4308e2..6203918e1316 100644
--- a/mm/swap.c
+++ b/mm/swap.c
@@ -708,9 +708,10 @@ static void lru_add_drain_per_cpu(struct work_struct *dummy)
  */
 void lru_add_drain_all(void)
 {
+	static seqcount_t seqcount = SEQCNT_ZERO(seqcount);
 	static DEFINE_MUTEX(lock);
 	static struct cpumask has_work;
-	int cpu;
+	int cpu, seq;
 
 	/*
 	 * Make sure nobody triggers this path before mm_percpu_wq is fully
@@ -719,7 +720,16 @@ void lru_add_drain_all(void)
 	if (WARN_ON(!mm_percpu_wq))
 		return;
 
+	seq = raw_read_seqcount_latch(&seqcount);
+
 	mutex_lock(&lock);
+
+	/* Piggyback on drain done by somebody else. */
+	if (__read_seqcount_retry(&seqcount, seq))
+		goto done;
+
+	raw_write_seqcount_latch(&seqcount);
+
 	cpumask_clear(&has_work);
 
 	for_each_online_cpu(cpu) {
@@ -740,6 +750,7 @@ void lru_add_drain_all(void)
 	for_each_cpu(cpu, &has_work)
 		flush_work(&per_cpu(lru_add_drain_work, cpu));
 
+done:
 	mutex_unlock(&lock);
 }
 #else


^ permalink raw reply related	[flat|nested] 6+ messages in thread

* Re: [PATCH] mm/swap: piggyback lru_add_drain_all() calls
  2019-10-04 10:11 [PATCH] mm/swap: piggyback lru_add_drain_all() calls Konstantin Khlebnikov
@ 2019-10-04 12:10 ` Matthew Wilcox
  2019-10-04 12:24   ` Konstantin Khlebnikov
  2019-10-04 12:27   ` Michal Hocko
  0 siblings, 2 replies; 6+ messages in thread
From: Matthew Wilcox @ 2019-10-04 12:10 UTC (permalink / raw)
  To: Konstantin Khlebnikov; +Cc: linux-mm, Andrew Morton, linux-kernel

On Fri, Oct 04, 2019 at 01:11:06PM +0300, Konstantin Khlebnikov wrote:
> This is very slow operation. There is no reason to do it again if somebody
> else already drained all per-cpu vectors after we waited for lock.
> +	seq = raw_read_seqcount_latch(&seqcount);
> +
>  	mutex_lock(&lock);
> +
> +	/* Piggyback on drain done by somebody else. */
> +	if (__read_seqcount_retry(&seqcount, seq))
> +		goto done;
> +
> +	raw_write_seqcount_latch(&seqcount);
> +

Do we really need the seqcount to do this?  Wouldn't a mutex_trylock()
have the same effect?

^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH] mm/swap: piggyback lru_add_drain_all() calls
  2019-10-04 12:10 ` Matthew Wilcox
@ 2019-10-04 12:24   ` Konstantin Khlebnikov
  2019-10-04 12:27   ` Michal Hocko
  1 sibling, 0 replies; 6+ messages in thread
From: Konstantin Khlebnikov @ 2019-10-04 12:24 UTC (permalink / raw)
  To: Matthew Wilcox; +Cc: linux-mm, Andrew Morton, linux-kernel



On 04/10/2019 15.10, Matthew Wilcox wrote:
> On Fri, Oct 04, 2019 at 01:11:06PM +0300, Konstantin Khlebnikov wrote:
>> This is very slow operation. There is no reason to do it again if somebody
>> else already drained all per-cpu vectors after we waited for lock.
>> +	seq = raw_read_seqcount_latch(&seqcount);
>> +
>>   	mutex_lock(&lock);
>> +
>> +	/* Piggyback on drain done by somebody else. */
>> +	if (__read_seqcount_retry(&seqcount, seq))
>> +		goto done;
>> +
>> +	raw_write_seqcount_latch(&seqcount);
>> +
> 
> Do we really need the seqcount to do this?  Wouldn't a mutex_trylock()
> have the same effect?
> 

No, this is completely different semantics.

Operation could be safely skipped only if somebody else started and
finished drain after current task called this function.

^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH] mm/swap: piggyback lru_add_drain_all() calls
  2019-10-04 12:10 ` Matthew Wilcox
  2019-10-04 12:24   ` Konstantin Khlebnikov
@ 2019-10-04 12:27   ` Michal Hocko
  2019-10-04 12:32     ` Konstantin Khlebnikov
  1 sibling, 1 reply; 6+ messages in thread
From: Michal Hocko @ 2019-10-04 12:27 UTC (permalink / raw)
  To: Matthew Wilcox
  Cc: Konstantin Khlebnikov, linux-mm, Andrew Morton, linux-kernel

On Fri 04-10-19 05:10:17, Matthew Wilcox wrote:
> On Fri, Oct 04, 2019 at 01:11:06PM +0300, Konstantin Khlebnikov wrote:
> > This is very slow operation. There is no reason to do it again if somebody
> > else already drained all per-cpu vectors after we waited for lock.
> > +	seq = raw_read_seqcount_latch(&seqcount);
> > +
> >  	mutex_lock(&lock);
> > +
> > +	/* Piggyback on drain done by somebody else. */
> > +	if (__read_seqcount_retry(&seqcount, seq))
> > +		goto done;
> > +
> > +	raw_write_seqcount_latch(&seqcount);
> > +
> 
> Do we really need the seqcount to do this?  Wouldn't a mutex_trylock()
> have the same effect?

Yeah, this makes sense. From correctness point of view it should be ok
because no caller can expect that per-cpu pvecs are empty on return.
This might have some runtime effects that some paths might retry more -
e.g. offlining path drains pcp pvces before migrating the range away, if
there are pages still waiting for a worker to drain them then the
migration would fail and we would retry. But this not a correctness
issue.

-- 
Michal Hocko
SUSE Labs

^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH] mm/swap: piggyback lru_add_drain_all() calls
  2019-10-04 12:27   ` Michal Hocko
@ 2019-10-04 12:32     ` Konstantin Khlebnikov
  2019-10-04 12:37       ` Michal Hocko
  0 siblings, 1 reply; 6+ messages in thread
From: Konstantin Khlebnikov @ 2019-10-04 12:32 UTC (permalink / raw)
  To: Michal Hocko, Matthew Wilcox; +Cc: linux-mm, Andrew Morton, linux-kernel



On 04/10/2019 15.27, Michal Hocko wrote:
> On Fri 04-10-19 05:10:17, Matthew Wilcox wrote:
>> On Fri, Oct 04, 2019 at 01:11:06PM +0300, Konstantin Khlebnikov wrote:
>>> This is very slow operation. There is no reason to do it again if somebody
>>> else already drained all per-cpu vectors after we waited for lock.
>>> +	seq = raw_read_seqcount_latch(&seqcount);
>>> +
>>>   	mutex_lock(&lock);
>>> +
>>> +	/* Piggyback on drain done by somebody else. */
>>> +	if (__read_seqcount_retry(&seqcount, seq))
>>> +		goto done;
>>> +
>>> +	raw_write_seqcount_latch(&seqcount);
>>> +
>>
>> Do we really need the seqcount to do this?  Wouldn't a mutex_trylock()
>> have the same effect?
> 
> Yeah, this makes sense. From correctness point of view it should be ok
> because no caller can expect that per-cpu pvecs are empty on return.
> This might have some runtime effects that some paths might retry more -
> e.g. offlining path drains pcp pvces before migrating the range away, if
> there are pages still waiting for a worker to drain them then the
> migration would fail and we would retry. But this not a correctness
> issue.
> 

Caller might expect that pages added by him before are drained.
Exiting after mutex_trylock() will not guarantee that.

For example POSIX_FADV_DONTNEED uses that.

^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH] mm/swap: piggyback lru_add_drain_all() calls
  2019-10-04 12:32     ` Konstantin Khlebnikov
@ 2019-10-04 12:37       ` Michal Hocko
  0 siblings, 0 replies; 6+ messages in thread
From: Michal Hocko @ 2019-10-04 12:37 UTC (permalink / raw)
  To: Konstantin Khlebnikov
  Cc: Matthew Wilcox, linux-mm, Andrew Morton, linux-kernel

On Fri 04-10-19 15:32:01, Konstantin Khlebnikov wrote:
> 
> 
> On 04/10/2019 15.27, Michal Hocko wrote:
> > On Fri 04-10-19 05:10:17, Matthew Wilcox wrote:
> > > On Fri, Oct 04, 2019 at 01:11:06PM +0300, Konstantin Khlebnikov wrote:
> > > > This is very slow operation. There is no reason to do it again if somebody
> > > > else already drained all per-cpu vectors after we waited for lock.
> > > > +	seq = raw_read_seqcount_latch(&seqcount);
> > > > +
> > > >   	mutex_lock(&lock);
> > > > +
> > > > +	/* Piggyback on drain done by somebody else. */
> > > > +	if (__read_seqcount_retry(&seqcount, seq))
> > > > +		goto done;
> > > > +
> > > > +	raw_write_seqcount_latch(&seqcount);
> > > > +
> > > 
> > > Do we really need the seqcount to do this?  Wouldn't a mutex_trylock()
> > > have the same effect?
> > 
> > Yeah, this makes sense. From correctness point of view it should be ok
> > because no caller can expect that per-cpu pvecs are empty on return.
> > This might have some runtime effects that some paths might retry more -
> > e.g. offlining path drains pcp pvces before migrating the range away, if
> > there are pages still waiting for a worker to drain them then the
> > migration would fail and we would retry. But this not a correctness
> > issue.
> > 
> 
> Caller might expect that pages added by him before are drained.
> Exiting after mutex_trylock() will not guarantee that.
> 
> For example POSIX_FADV_DONTNEED uses that.

OK, I was not aware of this case. Please make sure to document that in
the changelog and a comment in the code wouldn't hurt either. It would
certainly explain more thatn "Piggyback on drain done by somebody
else.".

Thanks!
-- 
Michal Hocko
SUSE Labs

^ permalink raw reply	[flat|nested] 6+ messages in thread

end of thread, other threads:[~2019-10-04 12:37 UTC | newest]

Thread overview: 6+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2019-10-04 10:11 [PATCH] mm/swap: piggyback lru_add_drain_all() calls Konstantin Khlebnikov
2019-10-04 12:10 ` Matthew Wilcox
2019-10-04 12:24   ` Konstantin Khlebnikov
2019-10-04 12:27   ` Michal Hocko
2019-10-04 12:32     ` Konstantin Khlebnikov
2019-10-04 12:37       ` Michal Hocko

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).