From patchwork Mon Feb 19 06:04:02 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Byungchul Park X-Patchwork-Id: 202902 Return-Path: Delivered-To: ouuuleilei@gmail.com Received: by 2002:a05:693c:2685:b0:108:e6aa:91d0 with SMTP id mn5csp1103163dyc; Sun, 18 Feb 2024 22:05:06 -0800 (PST) X-Forwarded-Encrypted: i=3; AJvYcCUHhYxDGdPzXqx/HZQjdb/CHnjZpVQGObiwIgPrtxXL2TxWSsoCNi+N4faXxgajh23Y76NTKziD2Yl3KWLJNPueS8lluA== X-Google-Smtp-Source: AGHT+IHV70Hf6Tlp85t26axkqePmxthKYH8u8RQfIHgyaJbqeZBXSrNDWdBQS7tXXZRnG4QoU2GF X-Received: by 2002:ac8:5787:0:b0:42d:ab95:a55e with SMTP id v7-20020ac85787000000b0042dab95a55emr16744023qta.49.1708322706772; Sun, 18 Feb 2024 22:05:06 -0800 (PST) ARC-Seal: i=2; a=rsa-sha256; t=1708322706; cv=pass; d=google.com; s=arc-20160816; b=SPPn2hSWV7M3QOY0FY9llTXm6vsp+5IMTZQRX5iFmFcqd+wvzvtMR3cuE+f/LdIX4x KuewFDBlxYA1yU3IJw0XUlmd+GLdO/aI2oJKGV6roFENY/MYOigB3ba4hrqaW5RMhk65 9go5sPCdrpJT1d21XyfoE6tsK31KTu7yvrSMGTl2NwGWwR6f9ZQEpuJ57KI5JIAGQFWu +RJBAAqJsEXLzywpJCPIFWBteZZYwKQnQBkalzHFLAfWh1UdkqOJrsBU8LqwiOQ6M+Wc Jb7XhNw9E96mBxtjTrcJGQmPyQUt/hPfuHOpbJbbcy66f9u1O2stqhJGudU961GKtHv6 X3DQ== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-unsubscribe:list-subscribe:list-id:precedence:references :in-reply-to:message-id:date:subject:cc:to:from; bh=deHIYqkuaMMowWAwSsScVOmIndLCnwrqbZwxQiYz+QA=; fh=I/fJb8ryqmmGEnsMy9AhDuMN2ego/v5YAM2Gbv30rAg=; b=HDkZhLgacnhNVK/KPHy6sL0CHifYbp2t7Ku9UCm2dzg97xg8qUpyXkWDYwLlfHOX8P LNINHtmhhm+JgFthxHV2xZToeYsAIz9Sn5YWs/e/QibcF1uag68cQfoAMasyEcLCV0FI si3EueyRKeSBbDqeTjTuSzP8yB7KwAuvMUICzQnhrh/rwJPod8et5NFVOrMQrGu5Ah1V HYfHKEJFuRi1K3ongjw//GTrVqiXDBHs/+aVbHtXzJlN+PPAaktzpvADEKHM3Tsj6DK1 /XE2grXK5TUfYNvsZBkbNHWoe9FAymWH/huHhWhLkwgddNzTu9tftjdlOVbgjyAG6oCs rpXw==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; arc=pass (i=1 spf=pass spfdomain=sk.com); spf=pass (google.com: domain of linux-kernel+bounces-70750-ouuuleilei=gmail.com@vger.kernel.org designates 2604:1380:45d1:ec00::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-70750-ouuuleilei=gmail.com@vger.kernel.org" Received: from ny.mirrors.kernel.org (ny.mirrors.kernel.org. [2604:1380:45d1:ec00::1]) by mx.google.com with ESMTPS id h9-20020ac85149000000b0042e1845dbc1si369076qtn.77.2024.02.18.22.05.06 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Sun, 18 Feb 2024 22:05:06 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel+bounces-70750-ouuuleilei=gmail.com@vger.kernel.org designates 2604:1380:45d1:ec00::1 as permitted sender) client-ip=2604:1380:45d1:ec00::1; Authentication-Results: mx.google.com; arc=pass (i=1 spf=pass spfdomain=sk.com); spf=pass (google.com: domain of linux-kernel+bounces-70750-ouuuleilei=gmail.com@vger.kernel.org designates 2604:1380:45d1:ec00::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-70750-ouuuleilei=gmail.com@vger.kernel.org" Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ny.mirrors.kernel.org (Postfix) with ESMTPS id 8FEC61C21057 for ; Mon, 19 Feb 2024 06:05:06 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id A0018208DA; Mon, 19 Feb 2024 06:04:33 +0000 (UTC) Received: from invmail4.hynix.com (exvmail4.hynix.com [166.125.252.92]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 3BBB6200BA for ; Mon, 19 Feb 2024 06:04:27 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=166.125.252.92 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1708322671; cv=none; b=ZXopp4TV7ji4rCVJ8vvP0w2EST2l2liQTN02uImbGotTmToSpYb974eVbXDAzKnZ3LbDlYat7ilgz170b2oN3VP842di8dHJmAIt1Te0vqc3WlySNVFe5qgAa6g/6VSWZYa/1FvnI+GJXgV2gUekIYgOg/P795/cE21W8mTvup8= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1708322671; c=relaxed/simple; bh=WHuF8WryLfu8JX/60dlBPk8CVS136b4yvUfGBE1UZcQ=; h=From:To:Cc:Subject:Date:Message-Id:In-Reply-To:References; b=ThTTRtG8QA70afD+s9595ZI540IyZAfWVcpjlcIwHKGlXxklEd3cFgT93G2rFZhfjzOaplZH0xzglxc4dBkNgthAlA1XbdQ5bEGkQRCiDqTJuPD8kt4rNPJ5urP+gvmH2RmpXlGQJeUn0mU/kpJitaisqLYyOOODiQIrZqh/Huw= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=none (p=none dis=none) header.from=sk.com; spf=pass smtp.mailfrom=sk.com; arc=none smtp.client-ip=166.125.252.92 Authentication-Results: smtp.subspace.kernel.org; dmarc=none (p=none dis=none) header.from=sk.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=sk.com X-AuditID: a67dfc5b-d6dff70000001748-c7-65d2ef6120f8 From: Byungchul Park To: linux-kernel@vger.kernel.org, linux-mm@kvack.org Cc: kernel_team@skhynix.com, akpm@linux-foundation.org, ying.huang@intel.com, namit@vmware.com, vernhao@tencent.com, mgorman@techsingularity.net, hughd@google.com, willy@infradead.org, david@redhat.com, peterz@infradead.org, luto@kernel.org, tglx@linutronix.de, mingo@redhat.com, bp@alien8.de, dave.hansen@linux.intel.com, rjgolo@gmail.com Subject: [PATCH v8 3/8] mm/rmap: Recognize read-only TLB entries during batched TLB flush Date: Mon, 19 Feb 2024 15:04:02 +0900 Message-Id: <20240219060407.25254-4-byungchul@sk.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20240219060407.25254-1-byungchul@sk.com> References: <20240219060407.25254-1-byungchul@sk.com> X-Brightmail-Tracker: H4sIAAAAAAAAA+NgFnrPLMWRmVeSWpSXmKPExsXC9ZZnkW7i+0upBhsW6lvMWb+GzeLzhn9s Fi82tDNafF3/i9ni6ac+FovLu+awWdxb85/V4vyutawWO5buY7K4dGABk8X1XQ8ZLY73HmCy mH/vM5vF5k1TmS2OT5nKaPH7B1DHyVmTWRwEPb639rF47Jx1l91jwaZSj80rtDwW73nJ5LFp VSebx6ZPk9g93p07x+5xYsZvFo95JwM93u+7yuax9ZedR+PUa2wenzfJebyb/5YtgD+KyyYl NSezLLVI3y6BK+PrPKeC49IVHz7tYGxgfCDWxcjJISFgIrHl3SQmGPv4yimsIDabgLrEjRs/ mUFsEQEziYOtf9i7GLk4mAU+Mkms/t7BApIQFoiUOLTpPJjNIqAq8fTlH7BmXgFTiRuf9kAN lZdYveEA0CAODk6gQbsPCoGEhYBKTlydzAQyU0KgmV3iyNE2Foh6SYmDK26wTGDkXcDIsIpR KDOvLDcxM8dEL6MyL7NCLzk/dxMjMBqW1f6J3sH46ULwIUYBDkYlHt4MkUupQqyJZcWVuYcY JTiYlUR43ZsupArxpiRWVqUW5ccXleakFh9ilOZgURLnNfpWniIkkJ5YkpqdmlqQWgSTZeLg lGpgjNytd05he82hCWc/NTn08igeecM5/fwcB7FbUsf2pPCW+HP1GyfoCsd1XXA40yVUV9h7 0ne6BdefOZUnI9jNcisFirZL2/ZbnuS1/iPOGbUo0qQyZyFXZPBx8xkf05Yv4t6mwyqqHCcs fjI2sWC1/5NPzjVvfW/a+8qmfzuVtUjcPPnxn3dKLMUZiYZazEXFiQD2/cjxggIAAA== X-Brightmail-Tracker: H4sIAAAAAAAAA+NgFrrGLMWRmVeSWpSXmKPExsXC5WfdrJv4/lKqQfNKZYs569ewWXze8I/N 4sWGdkaLr+t/MVs8/dTHYnF47klWi8u75rBZ3Fvzn9Xi/K61rBY7lu5jsrh0YAGTxfVdDxkt jvceYLKYf+8zm8XmTVOZLY5Pmcpo8fsHUMfJWZNZHIQ8vrf2sXjsnHWX3WPBplKPzSu0PBbv ecnksWlVJ5vHpk+T2D3enTvH7nFixm8Wj3knAz3e77vK5rH4xQcmj62/7Dwap15j8/i8Sc7j 3fy3bAECUVw2Kak5mWWpRfp2CVwZX+c5FRyXrvjwaQdjA+MDsS5GTg4JAROJ4yunsILYbALq Ejdu/GQGsUUEzCQOtv5h72Lk4mAW+Mgksfp7BwtIQlggUuLQpvNgNouAqsTTl3/AmnkFTCVu fNrDBDFUXmL1hgNAgzg4OIEG7T4oBBIWAio5cXUy0wRGrgWMDKsYRTLzynITM3NM9YqzMyrz Miv0kvNzNzECQ3tZ7Z+JOxi/XHY/xCjAwajEw5shcilViDWxrLgy9xCjBAezkgive9OFVCHe lMTKqtSi/Pii0pzU4kOM0hwsSuK8XuGpCUIC6YklqdmpqQWpRTBZJg5OqQbG5swyM7PeKvlC BZMJuvN+7F/6aNG+y5VFn/fe63+0VGOC6RLGaxOauVS28YckLCgpFLU7GXu3oaZA9g3DA1lW pobWLS0neyb/byu+U51Y/u/k2TKfz2G3k2xWnBU+YF/kEVtxWfzns9+b+HQSTe7kzlWcav/l mOqHvrZaF4ZLZ1aFa/14sJRPiaU4I9FQi7moOBEANl9Y0GkCAAA= X-CFilter-Loop: Reflected Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: X-getmail-retrieved-from-mailbox: INBOX X-GMAIL-THRID: 1791306190676133414 X-GMAIL-MSGID: 1791306190676133414 Functionally, no change. This is a preparation for migrc mechanism that requires to recognize read-only TLB entries and makes use of them to batch more aggressively. Plus, the newly introduced API, fold_ubc() will be used by migrc mechanism when manipulating tlb batch data. Signed-off-by: Byungchul Park --- include/linux/sched.h | 1 + mm/internal.h | 4 ++++ mm/rmap.c | 31 ++++++++++++++++++++++++++++++- 3 files changed, 35 insertions(+), 1 deletion(-) diff --git a/include/linux/sched.h b/include/linux/sched.h index 292c31697248..0317e7a65151 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h @@ -1328,6 +1328,7 @@ struct task_struct { #endif struct tlbflush_unmap_batch tlb_ubc; + struct tlbflush_unmap_batch tlb_ubc_ro; /* Cache last used pipe for splice(): */ struct pipe_inode_info *splice_pipe; diff --git a/mm/internal.h b/mm/internal.h index b61034bd50f5..b880f1e78700 100644 --- a/mm/internal.h +++ b/mm/internal.h @@ -923,6 +923,7 @@ extern struct workqueue_struct *mm_percpu_wq; void try_to_unmap_flush(void); void try_to_unmap_flush_dirty(void); void flush_tlb_batched_pending(struct mm_struct *mm); +void fold_ubc(struct tlbflush_unmap_batch *dst, struct tlbflush_unmap_batch *src); #else static inline void try_to_unmap_flush(void) { @@ -933,6 +934,9 @@ static inline void try_to_unmap_flush_dirty(void) static inline void flush_tlb_batched_pending(struct mm_struct *mm) { } +static inline void fold_ubc(struct tlbflush_unmap_batch *dst, struct tlbflush_unmap_batch *src) +{ +} #endif /* CONFIG_ARCH_WANT_BATCHED_UNMAP_TLB_FLUSH */ extern const struct trace_print_flags pageflag_names[]; diff --git a/mm/rmap.c b/mm/rmap.c index 7a27a2b41802..da36f23ff7b0 100644 --- a/mm/rmap.c +++ b/mm/rmap.c @@ -605,6 +605,28 @@ struct anon_vma *folio_lock_anon_vma_read(struct folio *folio, } #ifdef CONFIG_ARCH_WANT_BATCHED_UNMAP_TLB_FLUSH + +void fold_ubc(struct tlbflush_unmap_batch *dst, + struct tlbflush_unmap_batch *src) +{ + if (!src->flush_required) + return; + + /* + * Fold src to dst. + */ + arch_tlbbatch_fold(&dst->arch, &src->arch); + dst->writable = dst->writable || src->writable; + dst->flush_required = true; + + /* + * Reset src. + */ + arch_tlbbatch_clear(&src->arch); + src->flush_required = false; + src->writable = false; +} + /* * Flush TLB entries for recently unmapped pages from remote CPUs. It is * important if a PTE was dirty when it was unmapped that it's flushed @@ -614,7 +636,9 @@ struct anon_vma *folio_lock_anon_vma_read(struct folio *folio, void try_to_unmap_flush(void) { struct tlbflush_unmap_batch *tlb_ubc = ¤t->tlb_ubc; + struct tlbflush_unmap_batch *tlb_ubc_ro = ¤t->tlb_ubc_ro; + fold_ubc(tlb_ubc, tlb_ubc_ro); if (!tlb_ubc->flush_required) return; @@ -645,13 +669,18 @@ void try_to_unmap_flush_dirty(void) static void set_tlb_ubc_flush_pending(struct mm_struct *mm, pte_t pteval, unsigned long uaddr) { - struct tlbflush_unmap_batch *tlb_ubc = ¤t->tlb_ubc; + struct tlbflush_unmap_batch *tlb_ubc; int batch; bool writable = pte_dirty(pteval); if (!pte_accessible(mm, pteval)) return; + if (pte_write(pteval) || writable) + tlb_ubc = ¤t->tlb_ubc; + else + tlb_ubc = ¤t->tlb_ubc_ro; + arch_tlbbatch_add_pending(&tlb_ubc->arch, mm, uaddr); tlb_ubc->flush_required = true;