From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-ed1-f70.google.com (mail-ed1-f70.google.com [209.85.208.70]) by kanga.kvack.org (Postfix) with ESMTP id 5C8576B000A for ; Wed, 4 Jul 2018 03:16:35 -0400 (EDT) Received: by mail-ed1-f70.google.com with SMTP id c20-v6so1830273eds.21 for ; Wed, 04 Jul 2018 00:16:35 -0700 (PDT) Received: from mx1.suse.de (mx2.suse.de. [195.135.220.15]) by mx.google.com with ESMTPS id d1-v6si3116665edn.311.2018.07.04.00.16.33 for (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Wed, 04 Jul 2018 00:16:33 -0700 (PDT) Date: Wed, 4 Jul 2018 09:16:32 +0200 From: Michal Hocko Subject: Re: [PATCH 0/8] OOM killer/reaper changes for avoiding OOM lockup problem. Message-ID: <20180704071632.GB22503@dhcp22.suse.cz> References: <20180703151223.GP16767@dhcp22.suse.cz> <20180703152922.GR16767@dhcp22.suse.cz> <201807040222.w642Mtlr099513@www262.sakura.ne.jp> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <201807040222.w642Mtlr099513@www262.sakura.ne.jp> Sender: owner-linux-mm@kvack.org List-ID: To: penguin-kernel@i-love.sakura.ne.jp Cc: linux-mm@kvack.org, akpm@linux-foundation.org, torvalds@linux-foundation.org On Wed 04-07-18 11:22:55, Tetsuo Handa wrote: > Michal Hocko wrote: > > > On Tue 03-07-18 23:25:01, Tetsuo Handa wrote: > > > > This series provides > > > > > > > > (1) Mitigation and a fix for CVE-2016-10723. > > > > > > > > (2) A mitigation for needlessly selecting next OOM victim reported > > > > by David Rientjes and rejected by Michal Hocko. > > > > > > > > (3) A preparation for handling many concurrent OOM victims which > > > > could become real by introducing memcg-aware OOM killer. > > > > > > It would have been great to describe the overal design in the cover > > > letter. So let me summarize just to be sure I understand the proposal. > > You understood the proposal correctly. > > > > You are removing the oom_reaper and moving the oom victim tear down to > > > the oom path. > > Yes. This is for getting rid of the lie > > /* > * Acquire the oom lock. If that fails, somebody else is > * making progress for us. > */ > if (!mutex_trylock(&oom_lock)) { > *did_some_progress = 1; > schedule_timeout_uninterruptible(1); > return NULL; > } > > which is leading to CVE-2016-10723. By reclaiming from the OOM killer path, > we can eliminate this heuristic. > > Of course, we don't have to remove the OOM reaper kernel thread. The thing is that the current design uses the oom_reaper only as a backup to get situation unstuck. Once you move all that heavy lifting into the oom path directly then you will have to handle all sorts of issues. E.g. how do you handle that a random process hitting OOM path has to pay the full price to tear down multi TB process? This is a lot of time. > But the OOM > killer path knows better than the OOM reaper path regarding which domains need > to be reclaimed faster than other domains. How can it tell any priority? > > > To handle cases where we cannot get mmap_sem to do that > > > work you simply decay oom_badness over time if there are no changes in > > > the victims oom score. > > Yes. It is a feedback based approach which handles not only unable to get > mmap_sem case but also already started free_pgtables()/remove_vma() case. > > However, you are not understanding [PATCH 3/8] correctly. It IS A FEEDBACK > BASED APPROACH which corresponds to > > while (attempts++ < MAX_OOM_REAP_RETRIES && !oom_reap_task_mm(tsk, mm)) > schedule_timeout_idle(HZ/10); > > in current code (though the timeout is different from current code). Yes the timeout is different. The current solution is based on the retry loops while you are proposing a fuzzy amount of retries scheme. You can retry much more times if they happen often and there is _some_ change in the oom_score. I am not yet convinced that the oom_score based back off is a good idea. If you wanted to give some comparable results you could easily implement it on top of the current oom_reaper and argue with some numbers. > > Correction. You do not decay oom_badness. You simply increase a stall > > counter anytime oom_badness hasn't changed since the last check (if that > > check happend at least HZ/10 ago) and get the victim out of sight if the > > counter is larger than 30. This is where 3s are coming from. So in fact > > this is the low boundary while it might be considerably larger depending > > on how often we examine the victim. > > Yes, it can become larger. But if we don't need to examine the OOM victim, it means > that we are not OOM. We need to examine the OOM victim only if we are still OOM. > (If we are no longer OOM, the OOM victim will disappear via exit_oom_mm() before > the counter reaches 30.) > > Note that neither current > > #define MAX_OOM_REAP_RETRIES 10 > > /* Retry the down_read_trylock(mmap_sem) a few times */ > while (attempts++ < MAX_OOM_REAP_RETRIES && !oom_reap_task_mm(tsk, mm)) > schedule_timeout_idle(HZ/10); > > code guarantees that the OOM reaper will give up in 1 second. And that is the genuine design decision. Because the timeout is not really important. It is the retry loop feedback that matters. We do not play any promisses that something will happen after one or how many seconds. We only care about retrying as long as it makes sense but not for ever. That not for ever part is an implementation detail that we can tune based on real bug reports. > If oom_reap_task_mm() > for one OOM domain takes e.g. one minute when there are OOM victims in multiple OOM > domains, we will needlessly block another OOM domain where allocating threads are > wasting CPU resources due to the lie mentioned above. Not at all. This would only happen when oom victims from other domains are stuck not able to make forward progress. Is this something to optimize the behavior for? > By allowing allocating threads to do direct OOM reaping, we won't needlessly block > allocating threads and we can eliminate the lie mentioned above. Yes, this would help _but_ there are other downsides as mentioned above. You have to carefuly weight them. > > > > > In order to not block in the oom context for too > > > long because the address space might be quite large, you allow to > > > direct oom reap from multiple contexts. > > Yes. Since the thread which acquired the oom_lock can be SCHED_IDLE priority, > this patch is trying to speed up direct OOM reaping by allowing other threads > who acquired the oom_lock with !SCHED_IDLE priority. Thus, [PATCH 7/8] tries to > reduce the duration of holding the oom_lock, by releasing the oom_lock before > doing direct OOM reaping. > > We could make oom_lock a spinlock (or "mutex_lock(&oom_lock); preempt_disable();" > and "preempt_enable(); mutex_unlock(&oom_lock);") but there is no perfect answer > for whether we should do so. But there is a good example, why we shouldn't do that. Crawling over the task list can take a long time. And you definitely do not want to stall even more for PREEMPT kernels. > > > You fail to explain why is this approach more appropriate and how you > > > have settled with your current tuning with 3s timeout etc... > > > > > > Considering how subtle this whole area is I am not overly happy about > > > another rewrite without a really strong reasoning behind. There is none > > > here, unfortunately. Well, except for statements how I reject something > > > without telling the whole story etc... > > However, you are not understanding [PATCH 1/8] correctly. You are simply > refusing my patch instead of proving that removing the short sleep _is_ safe. > Since you don't prove it, I won't remove the short sleep until [PATCH 8/8]. No, you are misunderstanding my point. Do not fiddle with the code you are not willing to understand. The whole sleeping has been done optimistically in the oom path. Removing it perfectly reasonable with the current code. We do have preemption points in the reclaim path. If they are not sufficient then this is another story. But no sane design relies on random sleeps in the OOM path which is probably the most cold path in the kernel. > In the kernel space, not yielding enough CPU resources for other threads to > make forward progress is a BUG. You learned it with warn_alloc() for allocation > stall reporting, didn't you? And again, you are blaming a wrong part of the pipe. The whole warn_alloc fiasco is the result of printk allowing unbound amount of work on behalf of the caller. -- Michal Hocko SUSE Labs