From patchwork Mon Mar 20 17:26:11 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: James Morse X-Patchwork-Id: 72333 Return-Path: Delivered-To: ouuuleilei@gmail.com Received: by 2002:a5d:604a:0:0:0:0:0 with SMTP id j10csp1355672wrt; Mon, 20 Mar 2023 11:03:40 -0700 (PDT) X-Google-Smtp-Source: AK7set/NBRUjqW4Y8nB/3IcAadUmsIIPv3cuQEN7dY2J9KbdAQKq/CMwi9EQZAAyd2Ymrg7wRQsd X-Received: by 2002:a05:6a20:1b30:b0:d3:a13a:4c35 with SMTP id ch48-20020a056a201b3000b000d3a13a4c35mr348554pzb.6.1679335419825; Mon, 20 Mar 2023 11:03:39 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1679335419; cv=none; d=google.com; s=arc-20160816; b=dloLXKizPDXuzrqbhL9W30BUS4AsmkihFqcbG/iDHRyrfY61aXuncKQtpJBtUu8xde pjEWYvRYfQVdEVlt95wRu7T9Q43uNGN9emqRU/cfwpqOGQSpZ6iyeFgyA0ZclQ5kyZAm nvn/72h4VSdOcqloSrU5BnpB+Vl3VbgaXqRwOHxCu7lK6c0wM/SksHqJtlDtODHnzd/K AH5ahiOjuhkUTbVQyjXfLv+HF4/+tKbSBiYougN9rk1RqtarAmPqk2bmVkq5Q57rQ/yP K8x9gaw7UAyP+noNBvEFzgqxVEWoAUPTxNYzZh+VSBOzK5TzoCojvUMudXfsEI07VQIT utGA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from; bh=a6lmvxTaH/CgYqU6cQ9uWiCyW+tuo0ab4N1J0wE19Jc=; b=egVk5ZSxkULxdEo5AUrPer3BxYnDRI6t7ytdFB3/m4ypYepILQJRc+rQc07fbteDdP SBhZnSHkykP7hGqE4X1SHBBhvbjjXL+bJKZt1GZOvVxdrAV80LON9+oCSG0AV8uiH11L w6Oy2OZD/zsp1JgjJrSdJEIBcewe1gV9s0JJIJ6NIA8wJ2ARjczJ/NdThb51OY731/7F it7tWJoELXEqV3f/kQ8mBJ25uM9bP09G7Q6lB6cAnr9zORnzmGJBsduju6XMYEkLCkLP FPDs7MVSPym++eQ0sZUe/UN390w3vxefysFs4WCzZJXrak5Down4A3gW3IFdCitKLQqE KPLA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id h11-20020a65480b000000b00502d6b2edd4si10392243pgs.804.2023.03.20.11.03.22; Mon, 20 Mar 2023 11:03:39 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229993AbjCTRrQ (ORCPT + 99 others); Mon, 20 Mar 2023 13:47:16 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:33590 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229872AbjCTRqb (ORCPT ); Mon, 20 Mar 2023 13:46:31 -0400 Received: from foss.arm.com (foss.arm.com [217.140.110.172]) by lindbergh.monkeyblade.net (Postfix) with ESMTP id 76C1A3B220 for ; Mon, 20 Mar 2023 10:42:48 -0700 (PDT) Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 220151682; Mon, 20 Mar 2023 10:28:23 -0700 (PDT) Received: from merodach.members.linode.com (unknown [172.31.20.19]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id A0DC53F67D; Mon, 20 Mar 2023 10:27:36 -0700 (PDT) From: James Morse To: x86@kernel.org, linux-kernel@vger.kernel.org Cc: Fenghua Yu , Reinette Chatre , Thomas Gleixner , Ingo Molnar , Borislav Petkov , H Peter Anvin , Babu Moger , James Morse , shameerali.kolothum.thodi@huawei.com, D Scott Phillips OS , carl@os.amperecomputing.com, lcherian@marvell.com, bobo.shaobowang@huawei.com, tan.shaopeng@fujitsu.com, xingxin.hx@openanolis.org, baolin.wang@linux.alibaba.com, Jamie Iles , Xin Hao , peternewman@google.com Subject: [PATCH v3 10/19] x86/resctrl: Allow resctrl_arch_rmid_read() to sleep Date: Mon, 20 Mar 2023 17:26:11 +0000 Message-Id: <20230320172620.18254-11-james.morse@arm.com> X-Mailer: git-send-email 2.20.1 In-Reply-To: <20230320172620.18254-1-james.morse@arm.com> References: <20230320172620.18254-1-james.morse@arm.com> MIME-Version: 1.0 X-Spam-Status: No, score=-4.2 required=5.0 tests=BAYES_00,RCVD_IN_DNSWL_MED, SPF_HELO_NONE,SPF_NONE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X-getmail-retrieved-from-mailbox: =?utf-8?q?INBOX?= X-GMAIL-THRID: =?utf-8?q?1760910817116735123?= X-GMAIL-MSGID: =?utf-8?q?1760910817116735123?= MPAM's cache occupancy counters can take a little while to settle once the monitor has been configured. The maximum settling time is described to the driver via a firmware table. The value could be large enough that it makes sense to sleep. To avoid exposing this to resctrl, it should be hidden behind MPAM's resctrl_arch_rmid_read(). But add_rmid_to_limbo() calls resctrl_arch_rmid_read() from a non-preemptible context. add_rmid_to_limbo() is opportunistically reading the L3 occupancy counter on this domain to avoid adding the RMID to limbo if this domain's value has drifted below resctrl_rmid_realloc_threshold since the limbo handler last ran. Determining 'this domain' involves disabling preeption to prevent the thread being migrated to CPUs in a different domain between the check and resctrl_arch_rmid_read() call. The check is skipped for all remote domains. Instead, call resctrl_arch_rmid_read() for each domain, and get it to read the arch specific counter via IPI if its called on a CPU outside the target domain. By covering remote domains, this change stops the limbo handler from being started unnecessarily if a remote domain is below the threshold. This also allows resctrl_arch_rmid_read() to sleep. Tested-by: Shaopeng Tan Signed-off-by: James Morse --- The alternative is to remove the counter read from this path altogether, and assume user-space would never try to re-allocate the last RMID before the limbo handler runs next. --- arch/x86/kernel/cpu/resctrl/ctrlmondata.c | 12 +----- arch/x86/kernel/cpu/resctrl/monitor.c | 48 +++++++++++++++-------- 2 files changed, 33 insertions(+), 27 deletions(-) diff --git a/arch/x86/kernel/cpu/resctrl/ctrlmondata.c b/arch/x86/kernel/cpu/resctrl/ctrlmondata.c index b06e86839d00..9161bc95eea7 100644 --- a/arch/x86/kernel/cpu/resctrl/ctrlmondata.c +++ b/arch/x86/kernel/cpu/resctrl/ctrlmondata.c @@ -543,16 +543,8 @@ void mon_event_read(struct rmid_read *rr, struct rdt_resource *r, rr->val = 0; rr->first = first; - cpu = get_cpu(); - if (cpumask_test_cpu(cpu, &d->cpu_mask)) { - mon_event_count(rr); - put_cpu(); - } else { - put_cpu(); - - cpu = cpumask_any_housekeeping(&d->cpu_mask); - smp_call_on_cpu(cpu, mon_event_count, rr, false); - } + cpu = cpumask_any_housekeeping(&d->cpu_mask); + smp_call_on_cpu(cpu, mon_event_count, rr, false); } int rdtgroup_mondata_show(struct seq_file *m, void *arg) diff --git a/arch/x86/kernel/cpu/resctrl/monitor.c b/arch/x86/kernel/cpu/resctrl/monitor.c index 5e9e876c3409..de72df06b37b 100644 --- a/arch/x86/kernel/cpu/resctrl/monitor.c +++ b/arch/x86/kernel/cpu/resctrl/monitor.c @@ -253,22 +253,42 @@ static u64 mbm_overflow_count(u64 prev_msr, u64 cur_msr, unsigned int width) return chunks >> shift; } +struct __rmid_read_arg +{ + u32 rmid; + enum resctrl_event_id eventid; + + u64 msr_val; + int err; +}; + +static void smp_call_rmid_read(void *_arg) +{ + struct __rmid_read_arg *arg = _arg; + + arg->err = __rmid_read(arg->rmid, arg->eventid, &arg->msr_val); +} + int resctrl_arch_rmid_read(struct rdt_resource *r, struct rdt_domain *d, u32 closid, u32 rmid, enum resctrl_event_id eventid, u64 *val) { struct rdt_hw_resource *hw_res = resctrl_to_arch_res(r); struct rdt_hw_domain *hw_dom = resctrl_to_arch_dom(d); + struct __rmid_read_arg arg; struct arch_mbm_state *am; u64 msr_val, chunks; - int ret; + int err; - if (!cpumask_test_cpu(smp_processor_id(), &d->cpu_mask)) - return -EINVAL; + arg.rmid = rmid; + arg.eventid = eventid; - ret = __rmid_read(rmid, eventid, &msr_val); - if (ret) - return ret; + err = smp_call_function_any(&d->cpu_mask, smp_call_rmid_read, &arg, true); + if (err) + return err; + if (arg.err) + return arg.err; + msr_val = arg.msr_val; am = get_arch_mbm_state(hw_dom, rmid, eventid); if (am) { @@ -424,23 +444,18 @@ static void add_rmid_to_limbo(struct rmid_entry *entry) { struct rdt_resource *r = &rdt_resources_all[RDT_RESOURCE_L3].r_resctrl; struct rdt_domain *d; - int cpu, err; u64 val = 0; u32 idx; + int err; idx = resctrl_arch_rmid_idx_encode(entry->closid, entry->rmid); entry->busy = 0; - cpu = get_cpu(); list_for_each_entry(d, &r->domains, list) { - if (cpumask_test_cpu(cpu, &d->cpu_mask)) { - err = resctrl_arch_rmid_read(r, d, entry->closid, - entry->rmid, - QOS_L3_OCCUP_EVENT_ID, - &val); - if (err || val <= resctrl_rmid_realloc_threshold) - continue; - } + err = resctrl_arch_rmid_read(r, d, entry->closid, entry->rmid, + QOS_L3_OCCUP_EVENT_ID, &val); + if (err || val <= resctrl_rmid_realloc_threshold) + continue; /* * For the first limbo RMID in the domain, @@ -451,7 +466,6 @@ static void add_rmid_to_limbo(struct rmid_entry *entry) set_bit(idx, d->rmid_busy_llc); entry->busy++; } - put_cpu(); if (entry->busy) rmid_limbo_count++;