From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1754809AbdEQOsh (ORCPT ); Wed, 17 May 2017 10:48:37 -0400 Received: from resqmta-ch2-03v.sys.comcast.net ([69.252.207.35]:37538 "EHLO resqmta-ch2-03v.sys.comcast.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1754421AbdEQOse (ORCPT ); Wed, 17 May 2017 10:48:34 -0400 Date: Wed, 17 May 2017 09:48:25 -0500 (CDT) From: Christoph Lameter X-X-Sender: cl@east.gentwo.org To: Michal Hocko cc: Vlastimil Babka , linux-mm@kvack.org, linux-kernel@vger.kernel.org, cgroups@vger.kernel.org, Li Zefan , Mel Gorman , David Rientjes , Hugh Dickins , Andrea Arcangeli , Anshuman Khandual , "Kirill A. Shutemov" , linux-api@vger.kernel.org Subject: Re: [RFC 1/6] mm, page_alloc: fix more premature OOM due to race with cpuset update In-Reply-To: <20170517140501.GM18247@dhcp22.suse.cz> Message-ID: References: <20170411140609.3787-2-vbabka@suse.cz> <20170517092042.GH18247@dhcp22.suse.cz> <20170517140501.GM18247@dhcp22.suse.cz> Content-Type: text/plain; charset=US-ASCII X-CMAE-Envelope: MS4wfAn+/yNiuyCjgv6LFs0laZORKNNAuS5Sz+5U77swMfkBB/MGKYuczeZKc07P1EQOJPE15sDkxhT0nh8SAFgjkR1re0AvE9YCDL2I+XEknhyIFv6F+InS Ea5+acAFLaME91YvbPm3o1fXbWynZ2uKYtdYBoqABtNZ49ZO5ALs+OLy0xl/ht812VSoG3FUBhoG41q09cV0wgC+SDLdZ8SP+FMMCVSNRf/ne20nAA3DMqEX U9F6i/3+KhtpLqD5TefWplHyKHhPBNdPLJ5727Zpi6bsqVMVBWp3+ZbitSbeCeP3yPkgYcvUSt6zPcOZUs/lq7GRPPHlAgTYTS1BqxOgMkx3WnAaFDZ6YhvV ygyB0ESw+IEsihw+C8GTqSxq3edaDccs04RkbJG93rrEfqi3hxYQWVduD+SFiCjP0ikYblFvbVRK+crcnp68vIZCpJscBqEoYB+PhV5EsES/mAKIk2oyDvQX PovHLY0dka5BhjtZ5cQg45nDNpZljywqt0IEpzAYanBZjsPEy5XjC71Cvtq02kJJfLBU1BDVoVAyNGDZukEv0pe1ywqVp7VEJcMvpA== Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, 17 May 2017, Michal Hocko wrote: > > > So how are you going to distinguish VM_FAULT_OOM from an empty mempolicy > > > case in a raceless way? > > > > You dont have to do that if you do not create an empty mempolicy in the > > first place. The current kernel code avoids that by first allowing access > > to the new set of nodes and removing the old ones from the set when done. > > which is racy and as Vlastimil pointed out. If we simply fail such an > allocation the failure will go up the call chain until we hit the OOM > killer due to VM_FAULT_OOM. How would you want to handle that? The race is where? If you expand the node set during the move of the application then you are safe in terms of the legacy apps that did not include static bindings. If you have screwy things like static mbinds in there then you are hopelessly lost anyways. You may have moved the process to another set of nodes but the static bindings may refer to a node no longer available. Thus the OOM is legitimate. At least a user space app could inspect the situation and come up with custom ways of dealing with the mess.