From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-2.3 required=3.0 tests=DKIM_INVALID,DKIM_SIGNED, HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI,SPF_PASS,USER_AGENT_MUTT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id B028BC4360F for ; Wed, 3 Apr 2019 16:01:40 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id 680B3206B7 for ; Wed, 3 Apr 2019 16:01:40 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=fail reason="signature verification failed" (2048-bit key) header.d=infradead.org header.i=@infradead.org header.b="s8sW6KUN" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726770AbfDCQBi (ORCPT ); Wed, 3 Apr 2019 12:01:38 -0400 Received: from merlin.infradead.org ([205.233.59.134]:43528 "EHLO merlin.infradead.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726396AbfDCQBi (ORCPT ); Wed, 3 Apr 2019 12:01:38 -0400 DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=infradead.org; s=merlin.20170209; h=In-Reply-To:Content-Transfer-Encoding: Content-Type:MIME-Version:References:Message-ID:Subject:Cc:To:From:Date: Sender:Reply-To:Content-ID:Content-Description:Resent-Date:Resent-From: Resent-Sender:Resent-To:Resent-Cc:Resent-Message-ID:List-Id:List-Help: List-Unsubscribe:List-Subscribe:List-Post:List-Owner:List-Archive; bh=egIDybEuB51DrNyJlt/PcAuz3Iil78G3z2tbRZ7T7Zk=; b=s8sW6KUNt8FeI8iM2AxETX8YIA Uu6b9xPYXuMhgGPhSes5pfTUf3H4hfN1OpPbB0DylzdOYYnxBA8ATmpPEpbk29/i82slHf9PlRqV/ SC/AUW5fTristV9ADcMU7sfdKdVAZXn9fyjrFa7ETWjZz7jMkehzKmSeKczyNX9a+zEMN9fdm2M6H ZKizW9pPtmNQ35LlZCI2aYevlWfH3Hh7tGvdE09RyNvrW0++vL6sxNUM28sWzDjNZvxe1w8AbjgaL 4iYWeDj0vEDY3ZJL/F+sVI3N/v3r5/rm7NYK+ag+98fjKADtSdSSof8UwjU4UWXIN0dN+hRjmeRuj wJlGXsVA==; Received: from j217100.upc-j.chello.nl ([24.132.217.100] helo=hirez.programming.kicks-ass.net) by merlin.infradead.org with esmtpsa (Exim 4.90_1 #2 (Red Hat Linux)) id 1hBiKB-0003nq-Uz; Wed, 03 Apr 2019 16:01:16 +0000 Received: by hirez.programming.kicks-ass.net (Postfix, from userid 1000) id 061F72038C245; Wed, 3 Apr 2019 18:01:13 +0200 (CEST) Date: Wed, 3 Apr 2019 18:01:12 +0200 From: Peter Zijlstra To: Alex Kogan Cc: Waiman Long , linux@armlinux.org.uk, mingo@redhat.com, will.deacon@arm.com, arnd@arndb.de, linux-arch@vger.kernel.org, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org, tglx@linutronix.de, bp@alien8.de, hpa@zytor.com, x86@kernel.org, steven.sistare@oracle.com, daniel.m.jordan@oracle.com, dave.dice@oracle.com, rahul.x.yadav@oracle.com Subject: Re: [PATCH v2 3/5] locking/qspinlock: Introduce CNA into the slow path of qspinlock Message-ID: <20190403160112.GK4038@hirez.programming.kicks-ass.net> References: <20190329152006.110370-1-alex.kogan@oracle.com> <20190329152006.110370-4-alex.kogan@oracle.com> <60a3a2d8-d222-73aa-2df1-64c9d3fa3241@redhat.com> <20190402094320.GM11158@hirez.programming.kicks-ass.net> <6AEDE4F2-306A-4DF9-9307-9E3517C68A2B@oracle.com> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: <6AEDE4F2-306A-4DF9-9307-9E3517C68A2B@oracle.com> User-Agent: Mutt/1.10.1 (2018-07-13) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, Apr 03, 2019 at 11:39:09AM -0400, Alex Kogan wrote: > >> The patch that I am looking for is to have a separate > >> numa_queued_spinlock_slowpath() that coexists with > >> native_queued_spinlock_slowpath() and > >> paravirt_queued_spinlock_slowpath(). At boot time, we select the most > >> appropriate one for the system at hand. > Is this how this selection works today for paravirt? > I see a PARAVIRT_SPINLOCKS config option, but IIUC you are talking about a different mechanism here. > Can you, please, elaborate or give me a link to a page that explains that? Oh man, you ask us to explain how paravirt patching works... that's magic :-) Basically, the compiler will emit a bunch of indirect calls to the various pv_ops.*.* functions. Then, at alternative_instructions() <- apply_paravirt() it will rewrite all these indirect calls to direct calls to the function pointers that are in the pv_ops structure at that time (+- more magic). So we initialize the pv_ops.lock.* methods to the normal native_queued_spin*() stuff, if KVM/Xen/whatever setup detectors pv spnlock support changes the methods to the paravirt_queued_*() stuff. If you wnt more details, you'll just have to read arch/x86/include/asm/paravirt*.h and arch/x86/kernel/paravirt*.c, I don't think there's a coherent writeup of all that. > > Agreed; and until we have static_call, I think we can abuse the paravirt > > stuff for this. > > > > By the time we patch the paravirt stuff: > > > > check_bugs() > > alternative_instructions() > > apply_paravirt() > > > > we should already have enumerated the NODE topology and so nr_node_ids() > > should be set. > > > > So if we frob pv_ops.lock.queued_spin_lock_slowpath to > > numa_queued_spin_lock_slowpath before that, it should all get patched > > just right. > > > > That of course means the whole NUMA_AWARE_SPINLOCKS thing depends on > > PARAVIRT_SPINLOCK, which is a bit awkward… > Just to mention here, the patch so far does not address paravirt, but > our goal is to add this support once we address all the concerns for > the native version. So we will end up with four variants for the > queued_spinlock_slowpath() — one for each combination of > native/paravirt and NUMA/non-NUMA. Or perhaps we do not need a > NUMA/paravirt variant? I wouldn't bother with a pv version of the numa aware code at all. If you have overcommitted guests, topology is likely irrelevant anyway. If you have 1:1 pinned guests, they'll not use pv spinlocks anyway. So keep it to tertiary choice: - native - native/numa - paravirt From mboxrd@z Thu Jan 1 00:00:00 1970 From: Peter Zijlstra Subject: Re: [PATCH v2 3/5] locking/qspinlock: Introduce CNA into the slow path of qspinlock Date: Wed, 3 Apr 2019 18:01:12 +0200 Message-ID: <20190403160112.GK4038@hirez.programming.kicks-ass.net> References: <20190329152006.110370-1-alex.kogan@oracle.com> <20190329152006.110370-4-alex.kogan@oracle.com> <60a3a2d8-d222-73aa-2df1-64c9d3fa3241@redhat.com> <20190402094320.GM11158@hirez.programming.kicks-ass.net> <6AEDE4F2-306A-4DF9-9307-9E3517C68A2B@oracle.com> Mime-Version: 1.0 Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: base64 Return-path: Content-Disposition: inline In-Reply-To: <6AEDE4F2-306A-4DF9-9307-9E3517C68A2B@oracle.com> List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Sender: "linux-arm-kernel" Errors-To: linux-arm-kernel-bounces+linux-arm-kernel=m.gmane.org@lists.infradead.org To: Alex Kogan Cc: linux-arch@vger.kernel.org, arnd@arndb.de, dave.dice@oracle.com, x86@kernel.org, will.deacon@arm.com, linux@armlinux.org.uk, steven.sistare@oracle.com, linux-kernel@vger.kernel.org, rahul.x.yadav@oracle.com, mingo@redhat.com, bp@alien8.de, hpa@zytor.com, Waiman Long , tglx@linutronix.de, daniel.m.jordan@oracle.com, linux-arm-kernel@lists.infradead.org List-Id: linux-arch.vger.kernel.org T24gV2VkLCBBcHIgMDMsIDIwMTkgYXQgMTE6Mzk6MDlBTSAtMDQwMCwgQWxleCBLb2dhbiB3cm90 ZToKCj4gPj4gVGhlIHBhdGNoIHRoYXQgSSBhbSBsb29raW5nIGZvciBpcyB0byBoYXZlIGEgc2Vw YXJhdGUKPiA+PiBudW1hX3F1ZXVlZF9zcGlubG9ja19zbG93cGF0aCgpIHRoYXQgY29leGlzdHMg d2l0aAo+ID4+IG5hdGl2ZV9xdWV1ZWRfc3BpbmxvY2tfc2xvd3BhdGgoKSBhbmQKPiA+PiBwYXJh dmlydF9xdWV1ZWRfc3BpbmxvY2tfc2xvd3BhdGgoKS4gQXQgYm9vdCB0aW1lLCB3ZSBzZWxlY3Qg dGhlIG1vc3QKPiA+PiBhcHByb3ByaWF0ZSBvbmUgZm9yIHRoZSBzeXN0ZW0gYXQgaGFuZC4KPiBJ cyB0aGlzIGhvdyB0aGlzIHNlbGVjdGlvbiB3b3JrcyB0b2RheSBmb3IgcGFyYXZpcnQ/Cj4gSSBz ZWUgYSBQQVJBVklSVF9TUElOTE9DS1MgY29uZmlnIG9wdGlvbiwgYnV0IElJVUMgeW91IGFyZSB0 YWxraW5nIGFib3V0IGEgZGlmZmVyZW50IG1lY2hhbmlzbSBoZXJlLgo+IENhbiB5b3UsIHBsZWFz ZSwgZWxhYm9yYXRlIG9yIGdpdmUgbWUgYSBsaW5rIHRvIGEgcGFnZSB0aGF0IGV4cGxhaW5zIHRo YXQ/CgpPaCBtYW4sIHlvdSBhc2sgdXMgdG8gZXhwbGFpbiBob3cgcGFyYXZpcnQgcGF0Y2hpbmcg d29ya3MuLi4gdGhhdCdzCm1hZ2ljIDotKQoKQmFzaWNhbGx5LCB0aGUgY29tcGlsZXIgd2lsbCBl bWl0IGEgYnVuY2ggb2YgaW5kaXJlY3QgY2FsbHMgdG8gdGhlCnZhcmlvdXMgcHZfb3BzLiouKiBm dW5jdGlvbnMuCgpUaGVuLCBhdCBhbHRlcm5hdGl2ZV9pbnN0cnVjdGlvbnMoKSA8LSBhcHBseV9w YXJhdmlydCgpIGl0IHdpbGwgcmV3cml0ZQphbGwgdGhlc2UgaW5kaXJlY3QgY2FsbHMgdG8gZGly ZWN0IGNhbGxzIHRvIHRoZSBmdW5jdGlvbiBwb2ludGVycyB0aGF0CmFyZSBpbiB0aGUgcHZfb3Bz IHN0cnVjdHVyZSBhdCB0aGF0IHRpbWUgKCstIG1vcmUgbWFnaWMpLgoKU28gd2UgaW5pdGlhbGl6 ZSB0aGUgcHZfb3BzLmxvY2suKiBtZXRob2RzIHRvIHRoZSBub3JtYWwKbmF0aXZlX3F1ZXVlZF9z cGluKigpIHN0dWZmLCBpZiBLVk0vWGVuL3doYXRldmVyIHNldHVwIGRldGVjdG9ycyBwdgpzcG5s b2NrIHN1cHBvcnQgY2hhbmdlcyB0aGUgbWV0aG9kcyB0byB0aGUgcGFyYXZpcnRfcXVldWVkXyoo KSBzdHVmZi4KCklmIHlvdSB3bnQgbW9yZSBkZXRhaWxzLCB5b3UnbGwganVzdCBoYXZlIHRvIHJl YWQKYXJjaC94ODYvaW5jbHVkZS9hc20vcGFyYXZpcnQqLmggYW5kIGFyY2gveDg2L2tlcm5lbC9w YXJhdmlydCouYywgSQpkb24ndCB0aGluayB0aGVyZSdzIGEgY29oZXJlbnQgd3JpdGV1cCBvZiBh bGwgdGhhdC4KCj4gPiBBZ3JlZWQ7IGFuZCB1bnRpbCB3ZSBoYXZlIHN0YXRpY19jYWxsLCBJIHRo aW5rIHdlIGNhbiBhYnVzZSB0aGUgcGFyYXZpcnQKPiA+IHN0dWZmIGZvciB0aGlzLgo+ID4gCj4g PiBCeSB0aGUgdGltZSB3ZSBwYXRjaCB0aGUgcGFyYXZpcnQgc3R1ZmY6Cj4gPiAKPiA+ICBjaGVj a19idWdzKCkKPiA+ICAgIGFsdGVybmF0aXZlX2luc3RydWN0aW9ucygpCj4gPiAgICAgIGFwcGx5 X3BhcmF2aXJ0KCkKPiA+IAo+ID4gd2Ugc2hvdWxkIGFscmVhZHkgaGF2ZSBlbnVtZXJhdGVkIHRo ZSBOT0RFIHRvcG9sb2d5IGFuZCBzbyBucl9ub2RlX2lkcygpCj4gPiBzaG91bGQgYmUgc2V0Lgo+ ID4gCj4gPiBTbyBpZiB3ZSBmcm9iIHB2X29wcy5sb2NrLnF1ZXVlZF9zcGluX2xvY2tfc2xvd3Bh dGggdG8KPiA+IG51bWFfcXVldWVkX3NwaW5fbG9ja19zbG93cGF0aCBiZWZvcmUgdGhhdCwgaXQg c2hvdWxkIGFsbCBnZXQgcGF0Y2hlZAo+ID4ganVzdCByaWdodC4KPiA+IAo+ID4gVGhhdCBvZiBj b3Vyc2UgbWVhbnMgdGhlIHdob2xlIE5VTUFfQVdBUkVfU1BJTkxPQ0tTIHRoaW5nIGRlcGVuZHMg b24KPiA+IFBBUkFWSVJUX1NQSU5MT0NLLCB3aGljaCBpcyBhIGJpdCBhd2t3YXJk4oCmCgo+IEp1 c3QgdG8gbWVudGlvbiBoZXJlLCB0aGUgcGF0Y2ggc28gZmFyIGRvZXMgbm90IGFkZHJlc3MgcGFy YXZpcnQsIGJ1dAo+IG91ciBnb2FsIGlzIHRvIGFkZCB0aGlzIHN1cHBvcnQgb25jZSB3ZSBhZGRy ZXNzIGFsbCB0aGUgY29uY2VybnMgZm9yCj4gdGhlIG5hdGl2ZSB2ZXJzaW9uLiAgU28gd2Ugd2ls bCBlbmQgdXAgd2l0aCBmb3VyIHZhcmlhbnRzIGZvciB0aGUKPiBxdWV1ZWRfc3BpbmxvY2tfc2xv d3BhdGgoKSDigJQgb25lIGZvciBlYWNoIGNvbWJpbmF0aW9uIG9mCj4gbmF0aXZlL3BhcmF2aXJ0 IGFuZCBOVU1BL25vbi1OVU1BLiAgT3IgcGVyaGFwcyB3ZSBkbyBub3QgbmVlZCBhCj4gTlVNQS9w YXJhdmlydCB2YXJpYW50PwoKSSB3b3VsZG4ndCBib3RoZXIgd2l0aCBhIHB2IHZlcnNpb24gb2Yg dGhlIG51bWEgYXdhcmUgY29kZSBhdCBhbGwuIElmCnlvdSBoYXZlIG92ZXJjb21taXR0ZWQgZ3Vl c3RzLCB0b3BvbG9neSBpcyBsaWtlbHkgaXJyZWxldmFudCBhbnl3YXkuIElmCnlvdSBoYXZlIDE6 MSBwaW5uZWQgZ3Vlc3RzLCB0aGV5J2xsIG5vdCB1c2UgcHYgc3BpbmxvY2tzIGFueXdheS4KClNv IGtlZXAgaXQgdG8gdGVydGlhcnkgY2hvaWNlOgoKIC0gbmF0aXZlCiAtIG5hdGl2ZS9udW1hCiAt IHBhcmF2aXJ0CgpfX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX19f XwpsaW51eC1hcm0ta2VybmVsIG1haWxpbmcgbGlzdApsaW51eC1hcm0ta2VybmVsQGxpc3RzLmlu ZnJhZGVhZC5vcmcKaHR0cDovL2xpc3RzLmluZnJhZGVhZC5vcmcvbWFpbG1hbi9saXN0aW5mby9s aW51eC1hcm0ta2VybmVsCg==