From patchwork Mon Aug 8 09:18:16 2016 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: "Leizhen \(ThunderTown\)" X-Patchwork-Id: 73423 Delivered-To: patch@linaro.org Received: by 10.140.29.52 with SMTP id a49csp3116952qga; Mon, 8 Aug 2016 02:22:50 -0700 (PDT) X-Received: by 10.98.9.194 with SMTP id 63mr160967190pfj.56.1470648170743; Mon, 08 Aug 2016 02:22:50 -0700 (PDT) Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id x21si36109316pfj.106.2016.08.08.02.22.50; Mon, 08 Aug 2016 02:22:50 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of devicetree-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of devicetree-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=devicetree-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752274AbcHHJWt (ORCPT + 7 others); Mon, 8 Aug 2016 05:22:49 -0400 Received: from szxga03-in.huawei.com ([119.145.14.66]:20594 "EHLO szxga03-in.huawei.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752134AbcHHJUA (ORCPT ); Mon, 8 Aug 2016 05:20:00 -0400 Received: from 172.24.1.136 (EHLO szxeml434-hub.china.huawei.com) ([172.24.1.136]) by szxrg03-dlp.huawei.com (MOS 4.4.3-GA FastPath queued) with ESMTP id CGA52042; Mon, 08 Aug 2016 17:19:55 +0800 (CST) Received: from localhost (10.177.23.164) by szxeml434-hub.china.huawei.com (10.82.67.225) with Microsoft SMTP Server id 14.3.235.1; Mon, 8 Aug 2016 17:19:47 +0800 From: Zhen Lei To: Catalin Marinas , Will Deacon , linux-arm-kernel , linux-kernel , Rob Herring , "Frank Rowand" , devicetree CC: Zefan Li , Xinwei Hu , "Tianhong Ding" , Hanjun Guo , Zhen Lei Subject: [PATCH v5 11/14] arm64/numa: support HAVE_MEMORYLESS_NODES Date: Mon, 8 Aug 2016 17:18:16 +0800 Message-ID: <1470647899-6324-12-git-send-email-thunder.leizhen@huawei.com> X-Mailer: git-send-email 1.9.5.msysgit.1 In-Reply-To: <1470647899-6324-1-git-send-email-thunder.leizhen@huawei.com> References: <1470647899-6324-1-git-send-email-thunder.leizhen@huawei.com> MIME-Version: 1.0 X-Originating-IP: [10.177.23.164] X-CFilter-Loop: Reflected X-Mirapoint-Virus-RAPID-Raw: score=unknown(0), refid=str=0001.0A020206.57A84EBC.000A, ss=1, re=0.000, recu=0.000, reip=0.000, cl=1, cld=1, fgs=0, ip=0.0.0.0, so=2013-05-26 15:14:31, dmn=2013-03-21 17:37:32 X-Mirapoint-Loop-Id: f0aaeff36cb780771545f64ce62c2952 Sender: devicetree-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: devicetree@vger.kernel.org Some numa nodes may have no memory. For example: 1. cpu0 on node0 2. cpu1 on node1 3. device0 access the momory from node0 and node1 take the same time. So, we can not simply classify device0 to node0 or node1, but we can define a node2 which distances to node0 and node1 are the same. Signed-off-by: Zhen Lei --- arch/arm64/Kconfig | 4 ++++ arch/arm64/kernel/smp.c | 1 + arch/arm64/mm/numa.c | 43 +++++++++++++++++++++++++++++++++++++++++-- 3 files changed, 46 insertions(+), 2 deletions(-) -- 2.5.0 -- To unsubscribe from this list: send the line "unsubscribe devicetree" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html diff --git a/arch/arm64/Kconfig b/arch/arm64/Kconfig index 05c1bf1..5904a62 100644 --- a/arch/arm64/Kconfig +++ b/arch/arm64/Kconfig @@ -581,6 +581,10 @@ config NEED_PER_CPU_EMBED_FIRST_CHUNK def_bool y depends on NUMA +config HAVE_MEMORYLESS_NODES + def_bool y + depends on NUMA + source kernel/Kconfig.preempt source kernel/Kconfig.hz diff --git a/arch/arm64/kernel/smp.c b/arch/arm64/kernel/smp.c index d099306..9e15297 100644 --- a/arch/arm64/kernel/smp.c +++ b/arch/arm64/kernel/smp.c @@ -620,6 +620,7 @@ static void __init of_parse_and_init_cpus(void) } bootcpu_valid = true; + early_map_cpu_to_node(0, of_node_to_nid(dn)); /* * cpu_logical_map has already been diff --git a/arch/arm64/mm/numa.c b/arch/arm64/mm/numa.c index df5c842..d73b0a0 100644 --- a/arch/arm64/mm/numa.c +++ b/arch/arm64/mm/numa.c @@ -128,6 +128,14 @@ void __init early_map_cpu_to_node(unsigned int cpu, int nid) nid = 0; cpu_to_node_map[cpu] = nid; + + /* + * We should set the numa node of cpu0 as soon as possible, because it + * has already been set up online before. cpu_to_node(0) will soon be + * called. + */ + if (!cpu) + set_cpu_numa_node(cpu, nid); } #ifdef CONFIG_HAVE_SETUP_PER_CPU_AREA @@ -215,6 +223,35 @@ int __init numa_add_memblk(int nid, u64 start, u64 end) return ret; } +static u64 __init alloc_node_data_from_nearest_node(int nid, const size_t size) +{ + int i, best_nid, distance; + u64 pa; + DECLARE_BITMAP(nodes_map, MAX_NUMNODES); + + bitmap_zero(nodes_map, MAX_NUMNODES); + bitmap_set(nodes_map, nid, 1); + +find_nearest_node: + best_nid = NUMA_NO_NODE; + distance = INT_MAX; + + for_each_clear_bit(i, nodes_map, MAX_NUMNODES) + if (numa_distance[nid][i] < distance) { + best_nid = i; + distance = numa_distance[nid][i]; + } + + pa = memblock_alloc_nid(size, SMP_CACHE_BYTES, best_nid); + if (!pa) { + BUG_ON(best_nid == NUMA_NO_NODE); + bitmap_set(nodes_map, best_nid, 1); + goto find_nearest_node; + } + + return pa; +} + /** * Initialize NODE_DATA for a node on the local memory */ @@ -228,7 +265,9 @@ static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) pr_info("Initmem setup node %d [mem %#010Lx-%#010Lx]\n", nid, start_pfn << PAGE_SHIFT, (end_pfn << PAGE_SHIFT) - 1); - nd_pa = memblock_alloc_try_nid(nd_size, SMP_CACHE_BYTES, nid); + nd_pa = memblock_alloc_nid(nd_size, SMP_CACHE_BYTES, nid); + if (!nd_pa) + nd_pa = alloc_node_data_from_nearest_node(nid, nd_size); nd = __va(nd_pa); /* report and initialize */ @@ -238,7 +277,7 @@ static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) if (tnid != nid) pr_info(" NODE_DATA(%d) on node %d\n", nid, tnid); - node_data[nid] = nd; + NODE_DATA(nid) = nd; memset(NODE_DATA(nid), 0, sizeof(pg_data_t)); NODE_DATA(nid)->node_id = nid; NODE_DATA(nid)->node_start_pfn = start_pfn;