From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 5EFBBC433EF for ; Fri, 28 Jan 2022 00:29:23 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1344838AbiA1A3V (ORCPT ); Thu, 27 Jan 2022 19:29:21 -0500 Received: from mga18.intel.com ([134.134.136.126]:64736 "EHLO mga18.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S241256AbiA1A3M (ORCPT ); Thu, 27 Jan 2022 19:29:12 -0500 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1643329751; x=1674865751; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=qnDCWzwqZN3I15w/s8Qb2KClLbLckceUFz3RVBc72fw=; b=I8zubUAx6xO8/iyZskDFNTf+ez6tgARp9H0OBtUSFrT1oUaBGts/s0YN 6BqlVvylQOXYpFSzjSZZGcObktZ0liDbXmSxT94xoJ1Pv7ojYi2TAtACC wmnV8xj4O7fYBZ+2I+KEAtkwtkrQabVMicFme1+Y5AxScCR087Z5//zXy DpMtPPV9iIkMNdYRAgaaFJlGmAVnDhdhwP0vh2UXAVR3/2GSp3fxkN1oF hX96cuUUCJstacg5SncQr9Wy8y7zBxrW9n2bDkJEpC/8fc0yPy1x5RU18 4nImmXGmqPAt8HIMEGjwDD44L7GtKI5moKobqm/kcD2CJDfMu3Xl6O/L+ w==; X-IronPort-AV: E=McAfee;i="6200,9189,10239"; a="230580034" X-IronPort-AV: E=Sophos;i="5.88,322,1635231600"; d="scan'208";a="230580034" Received: from fmsmga001.fm.intel.com ([10.253.24.23]) by orsmga106.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 27 Jan 2022 16:27:30 -0800 X-IronPort-AV: E=Sophos;i="5.88,322,1635231600"; d="scan'208";a="674909665" Received: from vrao2-mobl1.gar.corp.intel.com (HELO localhost.localdomain) ([10.252.129.6]) by fmsmga001-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 27 Jan 2022 16:27:30 -0800 From: Ben Widawsky To: linux-cxl@vger.kernel.org Cc: patches@lists.linux.dev, Ben Widawsky , kernel test robot , Alison Schofield , Dan Williams , Ira Weiny , Jonathan Cameron , Vishal Verma , Bjorn Helgaas , nvdimm@lists.linux.dev, linux-pci@vger.kernel.org Subject: [PATCH v3 14/14] cxl/region: Create an nd_region Date: Thu, 27 Jan 2022 16:27:07 -0800 Message-Id: <20220128002707.391076-15-ben.widawsky@intel.com> X-Mailer: git-send-email 2.35.0 In-Reply-To: <20220128002707.391076-1-ben.widawsky@intel.com> References: <20220128002707.391076-1-ben.widawsky@intel.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: X-Mailing-List: linux-pci@vger.kernel.org LIBNVDIMM supports the creation of regions for both persistent and volatile memory ranges. The cxl_region driver is capable of handling the CXL side of region creation but will reuse LIBVDIMM for interfacing with the rest of the kernel. TODO: CXL regions can go away. As a result the nd_region must also be torn down. TODO2: Handle mappings. LIBNVDIMM is capable of being informed about which parts of devices contribute to a region and validating whether or not the region is configured properly. To do this properly requires tracking allocations per device. Reported-by: kernel test robot (v2) Signed-off-by: Ben Widawsky --- Changes since v2: - Check nvb is non-null - Give a dev_dbg for non-existent nvdimm_bus --- drivers/cxl/Kconfig | 3 ++- drivers/cxl/core/pmem.c | 16 ++++++++++++ drivers/cxl/cxl.h | 1 + drivers/cxl/region.c | 58 +++++++++++++++++++++++++++++++++++++++++ 4 files changed, 77 insertions(+), 1 deletion(-) diff --git a/drivers/cxl/Kconfig b/drivers/cxl/Kconfig index 742847503c16..054dc78d6f7d 100644 --- a/drivers/cxl/Kconfig +++ b/drivers/cxl/Kconfig @@ -99,7 +99,8 @@ config CXL_PORT tristate config CXL_REGION - default CXL_PORT + depends on CXL_PMEM + default CXL_BUS tristate endif diff --git a/drivers/cxl/core/pmem.c b/drivers/cxl/core/pmem.c index 7e431667ade1..58dc6fba3130 100644 --- a/drivers/cxl/core/pmem.c +++ b/drivers/cxl/core/pmem.c @@ -220,6 +220,22 @@ struct cxl_nvdimm *to_cxl_nvdimm(struct device *dev) } EXPORT_SYMBOL_NS_GPL(to_cxl_nvdimm, CXL); +static int match_cxl_nvdimm(struct device *dev, void *data) +{ + return is_cxl_nvdimm(dev); +} + +struct cxl_nvdimm *cxl_find_nvdimm(struct cxl_memdev *cxlmd) +{ + struct device *dev; + + dev = device_find_child(&cxlmd->dev, NULL, match_cxl_nvdimm); + if (!dev) + return NULL; + return to_cxl_nvdimm(dev); +} +EXPORT_SYMBOL_NS_GPL(cxl_find_nvdimm, CXL); + static struct cxl_nvdimm *cxl_nvdimm_alloc(struct cxl_memdev *cxlmd) { struct cxl_nvdimm *cxl_nvd; diff --git a/drivers/cxl/cxl.h b/drivers/cxl/cxl.h index 062654204eca..7eb8f36af30b 100644 --- a/drivers/cxl/cxl.h +++ b/drivers/cxl/cxl.h @@ -480,6 +480,7 @@ bool is_cxl_nvdimm(struct device *dev); bool is_cxl_nvdimm_bridge(struct device *dev); int devm_cxl_add_nvdimm(struct device *host, struct cxl_memdev *cxlmd); struct cxl_nvdimm_bridge *cxl_find_nvdimm_bridge(struct device *dev); +struct cxl_nvdimm *cxl_find_nvdimm(struct cxl_memdev *cxlmd); /* * Unit test builds overrides this to __weak, find the 'strong' version diff --git a/drivers/cxl/region.c b/drivers/cxl/region.c index ac290677534d..be472560fc6a 100644 --- a/drivers/cxl/region.c +++ b/drivers/cxl/region.c @@ -708,6 +708,58 @@ static int bind_region(struct cxl_region *cxlr) return rc; } +static int connect_to_libnvdimm(struct cxl_region *region) +{ + struct nd_region_desc ndr_desc; + struct cxl_nvdimm_bridge *nvb; + struct nd_region *ndr; + int rc = 0; + + nvb = cxl_find_nvdimm_bridge(®ion->config.targets[0]->dev); + if (!nvb) { + dev_dbg(®ion->dev, "Couldn't find nvdimm bridge\n"); + return -ENODEV; + } + + device_lock(&nvb->dev); + if (!nvb->nvdimm_bus) { + dev_dbg(&nvb->dev, "Couldn't find nvdimm bridge's bus\n"); + rc = -ENXIO; + goto out; + } + + memset(&ndr_desc, 0, sizeof(ndr_desc)); + + ndr_desc.res = region->res; + + ndr_desc.numa_node = memory_add_physaddr_to_nid(region->res->start); + ndr_desc.target_node = phys_to_target_node(region->res->start); + if (ndr_desc.numa_node == NUMA_NO_NODE) { + ndr_desc.numa_node = + memory_add_physaddr_to_nid(region->res->start); + dev_info(®ion->dev, + "changing numa node from %d to %d for CXL region %pR", + NUMA_NO_NODE, ndr_desc.numa_node, region->res); + } + if (ndr_desc.target_node == NUMA_NO_NODE) { + ndr_desc.target_node = ndr_desc.numa_node; + dev_info(®ion->dev, + "changing target node from %d to %d for CXL region %pR", + NUMA_NO_NODE, ndr_desc.target_node, region->res); + } + + ndr = nvdimm_pmem_region_create(nvb->nvdimm_bus, &ndr_desc); + if (IS_ERR(ndr)) + rc = PTR_ERR(ndr); + else + dev_set_drvdata(®ion->dev, ndr); + +out: + device_unlock(&nvb->dev); + put_device(&nvb->dev); + return rc; +} + static void region_unregister(void *dev) { struct cxl_region *region = to_cxl_region(dev); @@ -791,6 +843,12 @@ static int cxl_region_probe(struct device *dev) return ret; } + ret = connect_to_libnvdimm(cxlr); + if (ret) { + region_unregister(dev); + return ret; + } + cxlr->active = true; dev_info(dev, "Bound"); return devm_add_action_or_reset(dev, region_unregister, dev); -- 2.35.0