From patchwork Fri Dec 9 02:10:41 2022 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Kefeng Wang X-Patchwork-Id: 31589 Return-Path: Delivered-To: ouuuleilei@gmail.com Received: by 2002:adf:f944:0:0:0:0:0 with SMTP id q4csp529890wrr; Thu, 8 Dec 2022 17:58:16 -0800 (PST) X-Google-Smtp-Source: AA0mqf47ge4q9056YK/8F0sJ+sD8I3GF1WmAcWtpGoy24SAwSEYEVX/Qb/OU2x4CShP73BqyfKsS X-Received: by 2002:a05:6402:5505:b0:467:9912:8e11 with SMTP id fi5-20020a056402550500b0046799128e11mr3618887edb.13.1670551096248; Thu, 08 Dec 2022 17:58:16 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1670551096; cv=none; d=google.com; s=arc-20160816; b=YkqsakruFOeZsVZn1s4GG3S3CLSe0SHRf4k3pixzSVLG9QzVU2i9CarCGQllZ9Cbk4 NxaJ3sOjw0UMNYF/qvzfMDwBzaZuDzhCbqPFV5xUf+xrY1G2l5l36VuwDonj2zOK9h4K 5rrJRzpzJIHLmE1Yngc7dMg5ucIpF7NBWn/Kx/sv1YNQcjKSNbyUAgv6P0j5CzviiJkK h/u5nMJpeM57zgo0XtN7pUFDdxlkwz0rfNFK7+/XNqUCGMtmTKuy5OX1P+ceZazRkD+S S11dkl0cKJF6j9xWaH2JEcds6Bc1hylzXW46U6m3w/yKJkDsiJgPx49r3Rc1CAdS/U2f Nl9w== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from; bh=dVqj/pLfI1Quj4499c9IaSQehTndRmqLV1l34UFeCbY=; b=mhDPFZlZTySCh6MEbQ6X6jcdEK2SJ8KC4gtsw8qVk9K1cyDF04h5k5BUjt668GGFTL OJ1VVblSQH9hnDadA/gi8aDqk64OVp0optviklrSbi0hlrBULltNbToBNZNcdH87MhrM lg+bEnEMcpG7g+ax9a9IL1mzx3GOlFRpxA/6f0+ptJa5ksRrysbpA1JVxhtjFlDNRWCu dDJ8ZIGHNFPQxewXN6ih/l4g/b0LdamuLmfTGYaFRXeG8sqa/gflGpAvy6m1sH1i3bXv d7skX4xpqtat8IIJCnncY7zteb9JRGJYj6ie28UHmcOJiw+vhCzdBbPsEqaK8TI+rlM4 RALA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=huawei.com Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id b6-20020aa7c6c6000000b004699e89601dsi151144eds.588.2022.12.08.17.57.52; Thu, 08 Dec 2022 17:58:16 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=huawei.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230109AbiLIBx5 (ORCPT + 99 others); Thu, 8 Dec 2022 20:53:57 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:53080 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230022AbiLIBxm (ORCPT ); Thu, 8 Dec 2022 20:53:42 -0500 Received: from szxga02-in.huawei.com (szxga02-in.huawei.com [45.249.212.188]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id EDE3A98961 for ; Thu, 8 Dec 2022 17:53:40 -0800 (PST) Received: from dggpemm500001.china.huawei.com (unknown [172.30.72.55]) by szxga02-in.huawei.com (SkyGuard) with ESMTP id 4NSvDF6rh9zRpnP; Fri, 9 Dec 2022 09:52:45 +0800 (CST) Received: from localhost.localdomain.localdomain (10.175.113.25) by dggpemm500001.china.huawei.com (7.185.36.107) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.2375.31; Fri, 9 Dec 2022 09:53:38 +0800 From: Kefeng Wang To: , , CC: , , , Kefeng Wang Subject: [PATCH] mm: hwposion: support recovery from ksm_might_need_to_copy() Date: Fri, 9 Dec 2022 10:10:41 +0800 Message-ID: <20221209021041.192835-2-wangkefeng.wang@huawei.com> X-Mailer: git-send-email 2.35.3 In-Reply-To: <20221209021041.192835-1-wangkefeng.wang@huawei.com> References: <20221209021041.192835-1-wangkefeng.wang@huawei.com> MIME-Version: 1.0 X-Originating-IP: [10.175.113.25] X-ClientProxiedBy: dggems706-chm.china.huawei.com (10.3.19.183) To dggpemm500001.china.huawei.com (7.185.36.107) X-CFilter-Loop: Reflected X-Spam-Status: No, score=-4.2 required=5.0 tests=BAYES_00,RCVD_IN_DNSWL_MED, SPF_HELO_NONE,SPF_PASS autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X-getmail-retrieved-from-mailbox: =?utf-8?q?INBOX?= X-GMAIL-THRID: =?utf-8?q?1751699711580564857?= X-GMAIL-MSGID: =?utf-8?q?1751699786037377191?= When the kernel copy a page from ksm_might_need_to_copy(), but runs into an uncorrectable error, it will crash since poisoned page is consumed by kernel, this is similar to Copy-on-write poison recovery, When an error is detected during the page copy, return VM_FAULT_HWPOISON, which help us to avoid system crash. Note, memory failure on a KSM page will be skipped, but still call memory_failure_queue() to be consistent with general memory failure process. Signed-off-by: Kefeng Wang --- mm/ksm.c | 8 ++++++-- mm/memory.c | 3 +++ mm/swapfile.c | 2 +- 3 files changed, 10 insertions(+), 3 deletions(-) diff --git a/mm/ksm.c b/mm/ksm.c index f1e06b1d47f3..356e93b85287 100644 --- a/mm/ksm.c +++ b/mm/ksm.c @@ -2629,8 +2629,12 @@ struct page *ksm_might_need_to_copy(struct page *page, new_page = NULL; } if (new_page) { - copy_user_highpage(new_page, page, address, vma); - + if (copy_mc_user_highpage(new_page, page, address, vma)) { + put_page(new_page); + new_page = ERR_PTR(-EHWPOISON); + memory_failure_queue(page_to_pfn(page), 0); + return new_page; + } SetPageDirty(new_page); __SetPageUptodate(new_page); __SetPageLocked(new_page); diff --git a/mm/memory.c b/mm/memory.c index 2615fa615be4..bb7b35e42297 100644 --- a/mm/memory.c +++ b/mm/memory.c @@ -3840,6 +3840,9 @@ vm_fault_t do_swap_page(struct vm_fault *vmf) if (unlikely(!page)) { ret = VM_FAULT_OOM; goto out_page; + } els if (unlikely(PTR_ERR(page) == -EHWPOISON)) { + ret = VM_FAULT_HWPOISON; + goto out_page; } folio = page_folio(page); diff --git a/mm/swapfile.c b/mm/swapfile.c index f670ffb7df7e..763ff6a8a576 100644 --- a/mm/swapfile.c +++ b/mm/swapfile.c @@ -1767,7 +1767,7 @@ static int unuse_pte(struct vm_area_struct *vma, pmd_t *pmd, swapcache = page; page = ksm_might_need_to_copy(page, vma, addr); - if (unlikely(!page)) + if (IS_ERR_OR_NULL(!page)) return -ENOMEM; pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);