Message ID | 20221021073025.18831-1-wenchao.chen666@gmail.com |
---|---|
Headers |
Return-Path: <linux-kernel-owner@vger.kernel.org> Delivered-To: ouuuleilei@gmail.com Received: by 2002:a5d:4242:0:0:0:0:0 with SMTP id s2csp545928wrr; Fri, 21 Oct 2022 00:38:38 -0700 (PDT) X-Google-Smtp-Source: AMsMyM7yOgC/+b759NBoZ4Y2h8tPsoE49s00+YS0xhn90px+3G1baRgLbc0udjhB0GKnnDWsaonu X-Received: by 2002:a63:2b41:0:b0:46e:9364:eb07 with SMTP id r62-20020a632b41000000b0046e9364eb07mr7067083pgr.46.1666337917961; Fri, 21 Oct 2022 00:38:37 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1666337917; cv=none; d=google.com; s=arc-20160816; b=t0OfmroD1ni/ZmYLv/7p1etnPsGAXot8nUoYi250rPmH8oA60wWud5xStyhyw4wj1e cNaRXi8dmpWfRm6gvbYyTMUObW332tIXCImXn+InOILuB4AxWa2rnRB+BdHvjQ1zMLsC 9fN3m3BE+O7224NeHSWCQIpIl6DS3BxE+7Aq/Gfeubj6Pw5ZnoihB+B5EzxW0gux41Pn 5E/TxNPRQcWXhDAUmgZ3jDz+CvEmTVZypNfYseNwWD6l6hYMb+83mR36clAd6WswZ6XW Eg6ne1F4Q1aDeutUhVq+21IWQF+rL/AndxDEeO955aDQPpDrHMZH1VBkhqws1fF1IAJW FedQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :message-id:date:subject:cc:to:from:dkim-signature; bh=FWybCkNqc7BOIlhGtPQ8GIKC3VKRfzSYLOuETquJ7MQ=; b=ivEEzuIRiEn/8aMxIfgJMM6ZMzDD0i1DosRme4thWg083D+VrgZs9/ThE/ic5aBYjW /ZbEXaKgHazMI+wOlgdFJkwHgBoNP2vGFD7X6IGUwwAUfCoIiN57UhKNRYaOUxjbYGFB MkK+i3iKfsFZt8Vc5uT9nwg2if+LU0I9CDRpoN8hE67flSA6q9wNJoWTAxtzoXtzrW8e 9hiLb5eU9tHES2lH9D2+//s9jKquMtOJR8w/cttGxq+eiyAA9a9ZHJ36zw8rQDiwzm+9 iOptE57c2E8Nq87QPOH0dFVF0g5N60muSddZOXuWf0j5gXwa7tbfkksIsySuKf+t82Ul TXvA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20210112 header.b=U0yBPLAd; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id ob4-20020a17090b390400b0020d9da656e7si7507068pjb.30.2022.10.21.00.38.25; Fri, 21 Oct 2022 00:38:37 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20210112 header.b=U0yBPLAd; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229716AbiJUHar (ORCPT <rfc822;pwkd43@gmail.com> + 99 others); Fri, 21 Oct 2022 03:30:47 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:60492 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229849AbiJUHao (ORCPT <rfc822;linux-kernel@vger.kernel.org>); Fri, 21 Oct 2022 03:30:44 -0400 Received: from mail-pf1-x42e.google.com (mail-pf1-x42e.google.com [IPv6:2607:f8b0:4864:20::42e]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 8FF923B704; Fri, 21 Oct 2022 00:30:43 -0700 (PDT) Received: by mail-pf1-x42e.google.com with SMTP id h13so1823674pfr.7; Fri, 21 Oct 2022 00:30:43 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=content-transfer-encoding:mime-version:message-id:date:subject:cc :to:from:from:to:cc:subject:date:message-id:reply-to; bh=FWybCkNqc7BOIlhGtPQ8GIKC3VKRfzSYLOuETquJ7MQ=; b=U0yBPLAdkBv+Vxw5j7hfc4m2f/ZQWy2xecE2yMlGvLa4lkwPKKJQEZJkHKLDtlyCsq n+lNpkOANQPYyfjPPw/Anx7BbZE/dhzIUJqRXy3EEVYkby7CPDAoYvnL8ogyUA2BxWVN tUXY3jxFwjeNhJtWImPNrYXX/BqxHdov6WtUe7WbySC1VYf1xhm7/aVYU4zsc+GhNB1P LYfMF2uXYKsm91HF0y0/0UB43gZHzXmAAD7Kk0nh1DpPmmerC2ieIyK0WYZAdIVH/kZ4 ZNlPmOvkZCO62SVP95oWUFIUSSFKgjRLwOM0+VQc9rC8u1f638WqiM9DHoBu+U/E78VI MVNg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=content-transfer-encoding:mime-version:message-id:date:subject:cc :to:from:x-gm-message-state:from:to:cc:subject:date:message-id :reply-to; bh=FWybCkNqc7BOIlhGtPQ8GIKC3VKRfzSYLOuETquJ7MQ=; b=iHDySUZ662l62dDa5pL9J82Clx+XEaR5F/RGYlbcZ+18deRTCaUIEpb1omjmyC65IK bIbuA6R9LPgbsn22ZCfNHpQP3AKfXtmb5pvJlcTcuKe1Ky7ecMOP2sk/38rHKh0UvSqm FlDgLm+iDDxg67kMzyEavDmrdWjyqHYO+121+fWO4sRFnk9TbgI4zwM7JP+InJRyNhhl 3CSoF3i0pICuGW1TtHK1dhNthIWozRIv6XeXrq9Y9m74aUp4DLT5p3czofSe2l7W+r86 /uoxbl04HjULuHH3rcLitvr7URpWwCvrlPhMYU1er0/L/oGTub6Jt9hW/1adwLN/a7z/ O4ew== X-Gm-Message-State: ACrzQf1reopul9mKuj2mSv9ikueDRhX789p/KIU+tJHG/uNAFg9IZE+q uEto9zQxNZqe2w4dS1xMVNw= X-Received: by 2002:a05:6a00:4396:b0:563:6fd7:9c98 with SMTP id bt22-20020a056a00439600b005636fd79c98mr18284870pfb.13.1666337443014; Fri, 21 Oct 2022 00:30:43 -0700 (PDT) Received: from xm06403pcu.spreadtrum.com ([117.18.48.102]) by smtp.gmail.com with ESMTPSA id d4-20020a62f804000000b005628a30a500sm14531929pfh.41.2022.10.21.00.30.36 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 21 Oct 2022 00:30:42 -0700 (PDT) From: Wenchao Chen <wenchao.chen666@gmail.com> To: ulf.hansson@linaro.org, adrian.hunter@intel.com, orsonzhai@gmail.com, baolin.wang@linux.alibaba.com, zhang.lyra@gmail.com, axboe@kernel.dk, avri.altman@wdc.com, kch@nvidia.com Cc: CLoehle@hyperstone.com, vincent.whitchurch@axis.com, bigeasy@linutronix.de, s.shtylyov@omp.ru, michael@allwinnertech.com, linux-mmc@vger.kernel.org, linux-kernel@vger.kernel.org, megoo.tang@gmail.com, lzx.stg@gmail.com Subject: [PATCH V2 0/2] mmc: block: Support Host to control FUA Date: Fri, 21 Oct 2022 15:30:23 +0800 Message-Id: <20221021073025.18831-1-wenchao.chen666@gmail.com> X-Mailer: git-send-email 2.17.1 MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit X-Spam-Status: No, score=-1.8 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,FREEMAIL_ENVFROM_END_DIGIT, FREEMAIL_FROM,RCVD_IN_DNSWL_NONE,SPF_HELO_NONE,SPF_PASS autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: <linux-kernel.vger.kernel.org> X-Mailing-List: linux-kernel@vger.kernel.org X-getmail-retrieved-from-mailbox: =?utf-8?q?INBOX?= X-GMAIL-THRID: =?utf-8?q?1747281948324336918?= X-GMAIL-MSGID: =?utf-8?q?1747281948324336918?= |
Series |
mmc: block: Support Host to control FUA
|
|
Message
Wenchao Chen
Oct. 21, 2022, 7:30 a.m. UTC
From: Wenchao Chen <wenchao.chen@unisoc.com>
Summary
=======
These patches[1] supports the host to turn off FUA.
About FUA, roughly deal with the following two parts:
1) FUA(Forced Unit Access):
- The REQ_FUA flag can be OR ed into the r/w flags of a bio submitted from the
filesystem and will make sure that I/O completion for this request is only
signaled after the data has been committed to non-volatile storage.
2) In emmc, FUA is represented as Reliable write. code show as below:
static void mmc_blk_data_prep(struct mmc_queue *mq, struct mmc_queue_req *mqrq,
int recovery_mode, bool *do_rel_wr_p, bool *do_data_tag_p)
{
...
/*
* Reliable writes are used to implement Forced Unit Access and
* are supported only on MMCs.
*/
do_rel_wr = (req->cmd_flags & REQ_FUA) &&
rq_data_dir(req) == WRITE &&
(md->flags & MMC_BLK_REL_WR);
...
}
Patch structure
===============
patch#1: for block
patch#2: for sdhci-sprd
Tests
=====
Ran 'AndroBench' to evaluate the performance:
1. fua_disable = 1
/sys/block/mmcblk0/queue # cat fua 0
I tested 5 times for each case and output a average speed.
1) Sequential read:
Speed: 266.8MiB/s, 265.1MiB/s, 262.9MiB/s, 268.7MiB/s, 265.2MiB/s
Average speed: 265.74MiB/s
2) Random read:
Speed: 98.75MiB/s, 98.7MiB/s, 98.5MiB/s, 99.4MiB/s, 98.7MiB/s
Average speed: 98.81MiB/s
3) Sequential write:
Speed: 199.94MiB/s, 199.1MiB/s, 205.5MiB/s, 206.5MiB/s, 191.5MiB/s
Average speed: 200.5MiB/s
4) Random write:
Speed: 68.6MiB/s, 71.8MiB/s, 77.1MiB/s, 64.8MiB/s, 69.3MiB/s
Average speed: 70.32MiB/s
2. fua_disable = 0 (default 0)
/sys/block/mmcblk0/queue # cat fua 1
I tested 5 times for each case and output a average speed.
1) Sequential read:
Speed: 259.3MiB/s, 258.8MiB/s, 258.2MiB/s, 259.5MiB/s, 253.5MiB/s
Average speed: 257.86MiB/s
2) Random read:
Speed: 98.9MiB/s, 101MiB/s, 101MiB/s, 99MiB/s, 101.1MiB/s
Average speed: 100.2MiB/s
3) Sequential write:
Speed: 153.7MiB/s, 146.2MiB/s, 151.2MiB/s, 148.8MiB/s, 147.5MiB/s
Average speed: 149.48MiB/s
4) Random write:
Speed: 12.9MiB/s, 12.3MiB/s, 12.6MiB/s, 12.8MiB/s, 12.8MiB/s
Average speed: 12.68MiB/s
According to the above data, disable FUA (fua_disable = 1) improves the
performance:
1)Sequential read improved by 3%.
2)Random read were down 1%.
3)Sequential write improved by 34%.
4)Random write improved by 454%.
Therefore, it is recommended to support the host to control FUA.
Reference
=========
[1] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/Documentation/block/writeback_cache_control.rst
[2] Embedded Multi-Media Card (e•MMC) Electrical Standard (5.1)''
Wenchao Chen (2):
mmc: block: Support Host to control FUA
mmc: sdhci-sprd: enable fua_disable for SPRDSDHCI
drivers/mmc/core/block.c | 3 ++-
drivers/mmc/host/sdhci-sprd.c | 2 ++
include/linux/mmc/host.h | 3 +++
3 files changed, 7 insertions(+), 1 deletion(-)
Comments
On 21/10/22 10:30, Wenchao Chen wrote: > From: Wenchao Chen <wenchao.chen@unisoc.com> > > Summary > ======= > These patches[1] supports the host to turn off FUA. > > About FUA, roughly deal with the following two parts: > 1) FUA(Forced Unit Access): > - The REQ_FUA flag can be OR ed into the r/w flags of a bio submitted from the > filesystem and will make sure that I/O completion for this request is only > signaled after the data has been committed to non-volatile storage. > > 2) In emmc, FUA is represented as Reliable write. code show as below: > static void mmc_blk_data_prep(struct mmc_queue *mq, struct mmc_queue_req *mqrq, > int recovery_mode, bool *do_rel_wr_p, bool *do_data_tag_p) > { > ... > /* > * Reliable writes are used to implement Forced Unit Access and > * are supported only on MMCs. > */ > do_rel_wr = (req->cmd_flags & REQ_FUA) && > rq_data_dir(req) == WRITE && > (md->flags & MMC_BLK_REL_WR); > ... > } > > Patch structure > =============== > patch#1: for block > patch#2: for sdhci-sprd > > Tests > ===== > Ran 'AndroBench' to evaluate the performance: It would be good to have more details e.g. What file system? What block size? What journal size? What file size? What record size? > 1. fua_disable = 1 > /sys/block/mmcblk0/queue # cat fua 0 > I tested 5 times for each case and output a average speed. > > 1) Sequential read: > Speed: 266.8MiB/s, 265.1MiB/s, 262.9MiB/s, 268.7MiB/s, 265.2MiB/s > Average speed: 265.74MiB/s > > 2) Random read: > Speed: 98.75MiB/s, 98.7MiB/s, 98.5MiB/s, 99.4MiB/s, 98.7MiB/s > Average speed: 98.81MiB/s > > 3) Sequential write: > Speed: 199.94MiB/s, 199.1MiB/s, 205.5MiB/s, 206.5MiB/s, 191.5MiB/s > Average speed: 200.5MiB/s > > 4) Random write: > Speed: 68.6MiB/s, 71.8MiB/s, 77.1MiB/s, 64.8MiB/s, 69.3MiB/s > Average speed: 70.32MiB/s > > 2. fua_disable = 0 (default 0) > /sys/block/mmcblk0/queue # cat fua 1 > I tested 5 times for each case and output a average speed. > > 1) Sequential read: > Speed: 259.3MiB/s, 258.8MiB/s, 258.2MiB/s, 259.5MiB/s, 253.5MiB/s > Average speed: 257.86MiB/s > > 2) Random read: > Speed: 98.9MiB/s, 101MiB/s, 101MiB/s, 99MiB/s, 101.1MiB/s > Average speed: 100.2MiB/s > > 3) Sequential write: > Speed: 153.7MiB/s, 146.2MiB/s, 151.2MiB/s, 148.8MiB/s, 147.5MiB/s > Average speed: 149.48MiB/s > > 4) Random write: > Speed: 12.9MiB/s, 12.3MiB/s, 12.6MiB/s, 12.8MiB/s, 12.8MiB/s > Average speed: 12.68MiB/s Is every write being sync'ed of just sync at the end? > > According to the above data, disable FUA (fua_disable = 1) improves the > performance: > 1)Sequential read improved by 3%. > 2)Random read were down 1%. FUA should not affect reads. If it is, you may want to investigate how. > 3)Sequential write improved by 34%. > 4)Random write improved by 454%. > Therefore, it is recommended to support the host to control FUA. > > Reference > ========= > [1] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/Documentation/block/writeback_cache_control.rst > [2] Embedded Multi-Media Card (e•MMC) Electrical Standard (5.1)'' You do not seem to have considered data integrity. Regular disks are assumed to provide atomic sector writes. That is, a sector has either the old data or the new data, but not some corrupt mixture. mmc does not have that assumption, which is presumably why Reliable Write has been used instead. Although that idea appears to have been thrown away for devices with no cache by commit 08ebf903af57 ("mmc: core: Fixup support for writeback-cache for eMMC and SD"). File systems can use FUA to mark a successful journal flush. Whether or not getting a torn sector at that point will corrupt the file system recovery is presumably file system specific, and maybe specific to file system options e.g. the use of checksums. It may well be that a file system can survive a torn sector at that point, or that user space would prefer to take the risk in order to get better performance. In either of those cases, it is not really a decision for the host controller driver. > > Wenchao Chen (2): > mmc: block: Support Host to control FUA > mmc: sdhci-sprd: enable fua_disable for SPRDSDHCI > > drivers/mmc/core/block.c | 3 ++- > drivers/mmc/host/sdhci-sprd.c | 2 ++ > include/linux/mmc/host.h | 3 +++ > 3 files changed, 7 insertions(+), 1 deletion(-) >
> These patches[1] supports the host to turn off FUA. > > About FUA, roughly deal with the following two parts: > 1) FUA(Forced Unit Access): > - The REQ_FUA flag can be OR ed into the r/w flags of a bio submitted from the > filesystem and will make sure that I/O completion for this request is only > signaled after the data has been committed to non-volatile storage. > > 2) In emmc, FUA is represented as Reliable write. code show as below: > static void mmc_blk_data_prep(struct mmc_queue *mq, struct mmc_queue_req *mqrq, > int recovery_mode, bool *do_rel_wr_p, bool *do_data_tag_p) { > ... > /* > * Reliable writes are used to implement Forced Unit Access and > * are supported only on MMCs. > */ > do_rel_wr = (req->cmd_flags & REQ_FUA) && > rq_data_dir(req) == WRITE && > (md->flags & MMC_BLK_REL_WR); > ... > } So reliable writes are only issed when FUA flag is set, so as it should be? > > Patch structure > =============== > patch#1: for block > patch#2: for sdhci-sprd > > Tests > ===== > Ran 'AndroBench' to evaluate the performance: > 1. fua_disable = 1 > /sys/block/mmcblk0/queue # cat fua 0 > I tested 5 times for each case and output a average speed. > > 1) Sequential read: > Speed: 266.8MiB/s, 265.1MiB/s, 262.9MiB/s, 268.7MiB/s, 265.2MiB/s Average speed: 265.74MiB/s > > 2) Random read: > Speed: 98.75MiB/s, 98.7MiB/s, 98.5MiB/s, 99.4MiB/s, 98.7MiB/s Average speed: 98.81MiB/s > > 3) Sequential write: > Speed: 199.94MiB/s, 199.1MiB/s, 205.5MiB/s, 206.5MiB/s, 191.5MiB/s Average speed: 200.5MiB/s > > 4) Random write: > Speed: 68.6MiB/s, 71.8MiB/s, 77.1MiB/s, 64.8MiB/s, 69.3MiB/s Average speed: 70.32MiB/s > Unless there is something special / wrong with sdhci-sprd (unlikely? Its just a flag) or your eMMC (maybe more likely?) then Id expect this or similar performance degradation for any host controller and eMMC. I can redo some measurement if you provide your workload. But I'd say if you don’t want to pay the price of reliable write then make sure to not issue them, by not issuing FUA? Maybe I'm misunderstanding, but why would the host controller driver control FUA? > 2. fua_disable = 0 (default 0) > /sys/block/mmcblk0/queue # cat fua 1 > I tested 5 times for each case and output a average speed. > > 1) Sequential read: > Speed: 259.3MiB/s, 258.8MiB/s, 258.2MiB/s, 259.5MiB/s, 253.5MiB/s Average speed: 257.86MiB/s > > 2) Random read: > Speed: 98.9MiB/s, 101MiB/s, 101MiB/s, 99MiB/s, 101.1MiB/s Average speed: 100.2MiB/s > > 3) Sequential write: > Speed: 153.7MiB/s, 146.2MiB/s, 151.2MiB/s, 148.8MiB/s, 147.5MiB/s Average speed: 149.48MiB/s > > 4) Random write: > Speed: 12.9MiB/s, 12.3MiB/s, 12.6MiB/s, 12.8MiB/s, 12.8MiB/s Average speed: 12.68MiB/s > > According to the above data, disable FUA (fua_disable = 1) improves the > performance: > 1)Sequential read improved by 3%. > 2)Random read were down 1%. > 3)Sequential write improved by 34%. > 4)Random write improved by 454%. > Therefore, it is recommended to support the host to control FUA. > > Hyperstone GmbH | Reichenaustr. 39a | 78467 Konstanz Managing Director: Dr. Jan Peter Berns. Commercial register of local courts: Freiburg HRB381782
> > Unless there is something special / wrong with sdhci-sprd (unlikely? Its just a flag) or your eMMC (maybe more likely?) then Id expect this or similar performance degradation for any host controller and eMMC. > I can redo some measurement if you provide your workload. > But I'd say if you don’t want to pay the price of reliable write then make sure to not issue them, by not issuing FUA? > Maybe I'm misunderstanding, but why would the host controller driver control FUA? > Maybe one more point on that: Compare performance with fua disabled but cache off, should be comparable unless there is an actual reliable write problem. If reliable write is performing horribly on that eMMC maybe a quirk could be thought of. Anyway your 13MB/s random with fua, if a reliable write is basically always active for a given cache size, doesn't sound totally unreasonable (over 3000 IOPS, assuming 4k writes). Hyperstone GmbH | Reichenaustr. 39a | 78467 Konstanz Managing Director: Dr. Jan Peter Berns. Commercial register of local courts: Freiburg HRB381782
Hi Hunter Thank you for your review! I'm sorry to reply you so late because I've been too busy lately. On Fri, Oct 21, 2022 at 11:50 PM Adrian Hunter <adrian.hunter@intel.com> wrote: > > On 21/10/22 10:30, Wenchao Chen wrote: > > From: Wenchao Chen <wenchao.chen@unisoc.com> > > > > Summary > > ======= > > These patches[1] supports the host to turn off FUA. > > > > About FUA, roughly deal with the following two parts: > > 1) FUA(Forced Unit Access): > > - The REQ_FUA flag can be OR ed into the r/w flags of a bio submitted from the > > filesystem and will make sure that I/O completion for this request is only > > signaled after the data has been committed to non-volatile storage. > > > > 2) In emmc, FUA is represented as Reliable write. code show as below: > > static void mmc_blk_data_prep(struct mmc_queue *mq, struct mmc_queue_req *mqrq, > > int recovery_mode, bool *do_rel_wr_p, bool *do_data_tag_p) > > { > > ... > > /* > > * Reliable writes are used to implement Forced Unit Access and > > * are supported only on MMCs. > > */ > > do_rel_wr = (req->cmd_flags & REQ_FUA) && > > rq_data_dir(req) == WRITE && > > (md->flags & MMC_BLK_REL_WR); > > ... > > } > > > > Patch structure > > =============== > > patch#1: for block > > patch#2: for sdhci-sprd > > > > Tests > > ===== > > Ran 'AndroBench' to evaluate the performance: > > It would be good to have more details e.g. > What file system? What block size? What journal size? > What file size? What record size? > What file system? F2FS What block size? Sequential: 32768KB, Random: 4KB What file size? 64MB > > 1. fua_disable = 1 > > /sys/block/mmcblk0/queue # cat fua 0 > > I tested 5 times for each case and output a average speed. > > > > 1) Sequential read: > > Speed: 266.8MiB/s, 265.1MiB/s, 262.9MiB/s, 268.7MiB/s, 265.2MiB/s > > Average speed: 265.74MiB/s > > > > 2) Random read: > > Speed: 98.75MiB/s, 98.7MiB/s, 98.5MiB/s, 99.4MiB/s, 98.7MiB/s > > Average speed: 98.81MiB/s > > > > 3) Sequential write: > > Speed: 199.94MiB/s, 199.1MiB/s, 205.5MiB/s, 206.5MiB/s, 191.5MiB/s > > Average speed: 200.5MiB/s > > > > 4) Random write: > > Speed: 68.6MiB/s, 71.8MiB/s, 77.1MiB/s, 64.8MiB/s, 69.3MiB/s > > Average speed: 70.32MiB/s > > > > 2. fua_disable = 0 (default 0) > > /sys/block/mmcblk0/queue # cat fua 1 > > I tested 5 times for each case and output a average speed. > > > > 1) Sequential read: > > Speed: 259.3MiB/s, 258.8MiB/s, 258.2MiB/s, 259.5MiB/s, 253.5MiB/s > > Average speed: 257.86MiB/s > > > > 2) Random read: > > Speed: 98.9MiB/s, 101MiB/s, 101MiB/s, 99MiB/s, 101.1MiB/s > > Average speed: 100.2MiB/s > > > > 3) Sequential write: > > Speed: 153.7MiB/s, 146.2MiB/s, 151.2MiB/s, 148.8MiB/s, 147.5MiB/s > > Average speed: 149.48MiB/s > > > > 4) Random write: > > Speed: 12.9MiB/s, 12.3MiB/s, 12.6MiB/s, 12.8MiB/s, 12.8MiB/s > > Average speed: 12.68MiB/s > > Is every write being sync'ed of just sync at the end? > /* * Reliable writes are used to implement Forced Unit Access and * are supported only on MMCs. */ do_rel_wr = (req->cmd_flags & REQ_FUA) && rq_data_dir(req) == WRITE && (md->flags & MMC_BLK_REL_WR); A Reliable Write access shall force the data to be written to the nonvolatile storage。 It will consume more time. > > > > According to the above data, disable FUA (fua_disable = 1) improves the > > performance: > > 1)Sequential read improved by 3%. > > 2)Random read were down 1%. > > FUA should not affect reads. If it is, you may want to investigate how. > > > 3)Sequential write improved by 34%. > > 4)Random write improved by 454%. > > Therefore, it is recommended to support the host to control FUA. > > > > Reference > > ========= > > [1] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/Documentation/block/writeback_cache_control.rst > > [2] Embedded Multi-Media Card (e•MMC) Electrical Standard (5.1)'' > > You do not seem to have considered data integrity. > > Regular disks are assumed to provide atomic sector writes. That is, a sector has either the old data or the new data, but not some corrupt mixture. > > mmc does not have that assumption, which is presumably why Reliable Write has been used instead. Although that idea appears to have been thrown away for devices with no cache by commit 08ebf903af57 ("mmc: core: Fixup support for writeback-cache for eMMC and SD"). > > File systems can use FUA to mark a successful journal flush. Whether or not getting a torn sector at that point will corrupt the file system recovery is presumably file system specific, and maybe specific to file system options e.g. the use of checksums. > > It may well be that a file system can survive a torn sector at that point, or that user space would prefer to take the risk in order to get better performance. In either of those cases, it is not really a decision for the host controller driver. > Considering the data integrity, we did a random power-down test, and the experimental results were good. FUA can only reduce data loss under abnormal conditions, but cannot prevent data loss under abnormal conditions. I think there should be a balance between FUA and NO FUA, but filesystems seem to favor FUA. FUA brings a drop in random write performance. If enough tests are done, NO FUA is acceptable. I found a discussion about FUA: https://lore.kernel.org/linux-f2fs-devel/20220528051238.GX1098723@dread.disaster.area/ UFS reference: https://lore.kernel.org/linux-scsi/20220531201053.3300018-1-jaegeuk@kernel.org/ > > > > Wenchao Chen (2): > > mmc: block: Support Host to control FUA > > mmc: sdhci-sprd: enable fua_disable for SPRDSDHCI > > > > drivers/mmc/core/block.c | 3 ++- > > drivers/mmc/host/sdhci-sprd.c | 2 ++ > > include/linux/mmc/host.h | 3 +++ > > 3 files changed, 7 insertions(+), 1 deletion(-) > > >
On Fri, 11 Nov 2022 at 08:59, Wenchao Chen <wenchao.chen666@gmail.com> wrote: > > Hi Hunter > Thank you for your review! > I'm sorry to reply you so late because I've been too busy lately. > > On Fri, Oct 21, 2022 at 11:50 PM Adrian Hunter <adrian.hunter@intel.com> wrote: > > > > On 21/10/22 10:30, Wenchao Chen wrote: > > > From: Wenchao Chen <wenchao.chen@unisoc.com> > > > > > > Summary > > > ======= > > > These patches[1] supports the host to turn off FUA. > > > > > > About FUA, roughly deal with the following two parts: > > > 1) FUA(Forced Unit Access): > > > - The REQ_FUA flag can be OR ed into the r/w flags of a bio submitted from the > > > filesystem and will make sure that I/O completion for this request is only > > > signaled after the data has been committed to non-volatile storage. > > > > > > 2) In emmc, FUA is represented as Reliable write. code show as below: > > > static void mmc_blk_data_prep(struct mmc_queue *mq, struct mmc_queue_req *mqrq, > > > int recovery_mode, bool *do_rel_wr_p, bool *do_data_tag_p) > > > { > > > ... > > > /* > > > * Reliable writes are used to implement Forced Unit Access and > > > * are supported only on MMCs. > > > */ > > > do_rel_wr = (req->cmd_flags & REQ_FUA) && > > > rq_data_dir(req) == WRITE && > > > (md->flags & MMC_BLK_REL_WR); > > > ... > > > } > > > > > > Patch structure > > > =============== > > > patch#1: for block > > > patch#2: for sdhci-sprd > > > > > > Tests > > > ===== > > > Ran 'AndroBench' to evaluate the performance: > > > > It would be good to have more details e.g. > > What file system? What block size? What journal size? > > What file size? What record size? > > > > What file system? > F2FS > What block size? > Sequential: 32768KB, Random: 4KB > What file size? > 64MB > > > > 1. fua_disable = 1 > > > /sys/block/mmcblk0/queue # cat fua 0 > > > I tested 5 times for each case and output a average speed. > > > > > > 1) Sequential read: > > > Speed: 266.8MiB/s, 265.1MiB/s, 262.9MiB/s, 268.7MiB/s, 265.2MiB/s > > > Average speed: 265.74MiB/s > > > > > > 2) Random read: > > > Speed: 98.75MiB/s, 98.7MiB/s, 98.5MiB/s, 99.4MiB/s, 98.7MiB/s > > > Average speed: 98.81MiB/s > > > > > > 3) Sequential write: > > > Speed: 199.94MiB/s, 199.1MiB/s, 205.5MiB/s, 206.5MiB/s, 191.5MiB/s > > > Average speed: 200.5MiB/s > > > > > > 4) Random write: > > > Speed: 68.6MiB/s, 71.8MiB/s, 77.1MiB/s, 64.8MiB/s, 69.3MiB/s > > > Average speed: 70.32MiB/s > > > > > > 2. fua_disable = 0 (default 0) > > > /sys/block/mmcblk0/queue # cat fua 1 > > > I tested 5 times for each case and output a average speed. > > > > > > 1) Sequential read: > > > Speed: 259.3MiB/s, 258.8MiB/s, 258.2MiB/s, 259.5MiB/s, 253.5MiB/s > > > Average speed: 257.86MiB/s > > > > > > 2) Random read: > > > Speed: 98.9MiB/s, 101MiB/s, 101MiB/s, 99MiB/s, 101.1MiB/s > > > Average speed: 100.2MiB/s > > > > > > 3) Sequential write: > > > Speed: 153.7MiB/s, 146.2MiB/s, 151.2MiB/s, 148.8MiB/s, 147.5MiB/s > > > Average speed: 149.48MiB/s > > > > > > 4) Random write: > > > Speed: 12.9MiB/s, 12.3MiB/s, 12.6MiB/s, 12.8MiB/s, 12.8MiB/s > > > Average speed: 12.68MiB/s > > > > Is every write being sync'ed of just sync at the end? > > > > /* > * Reliable writes are used to implement Forced Unit Access and > * are supported only on MMCs. > */ > do_rel_wr = (req->cmd_flags & REQ_FUA) && > rq_data_dir(req) == WRITE && > (md->flags & MMC_BLK_REL_WR); > > A Reliable Write access shall force the data to be written to the > nonvolatile storage。 > It will consume more time. My apologies for side-stepping the discussion. Yes, REQ_FUA is per definition a write and a flush of the write-back-cache to non-volatile storage. So, this is indeed the correct behaviour, even if it consumes more time to complete the operation. > > > > > > > According to the above data, disable FUA (fua_disable = 1) improves the > > > performance: > > > 1)Sequential read improved by 3%. > > > 2)Random read were down 1%. > > > > FUA should not affect reads. If it is, you may want to investigate how. > > > > > 3)Sequential write improved by 34%. > > > 4)Random write improved by 454%. > > > Therefore, it is recommended to support the host to control FUA. > > > > > > Reference > > > ========= > > > [1] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/Documentation/block/writeback_cache_control.rst > > > [2] Embedded Multi-Media Card (e•MMC) Electrical Standard (5.1)'' > > > > You do not seem to have considered data integrity. > > > > Regular disks are assumed to provide atomic sector writes. That is, a sector has either the old data or the new data, but not some corrupt mixture. > > > > mmc does not have that assumption, which is presumably why Reliable Write has been used instead. Although that idea appears to have been thrown away for devices with no cache by commit 08ebf903af57 ("mmc: core: Fixup support for writeback-cache for eMMC and SD"). > > > > File systems can use FUA to mark a successful journal flush. Whether or not getting a torn sector at that point will corrupt the file system recovery is presumably file system specific, and maybe specific to file system options e.g. the use of checksums. > > > > It may well be that a file system can survive a torn sector at that point, or that user space would prefer to take the risk in order to get better performance. In either of those cases, it is not really a decision for the host controller driver. > > > > Considering the data integrity, we did a random power-down test, and > the experimental results were good. > > FUA can only reduce data loss under abnormal conditions, but cannot > prevent data loss under abnormal conditions. > > I think there should be a balance between FUA and NO FUA, but > filesystems seem to favor FUA. > > FUA brings a drop in random write performance. If enough tests are > done, NO FUA is acceptable. Testing this isn't entirely easy. It requires you to hook up electrical switches to allow you to automate the powering on/off of the platform(s). Then at each cycle, really make sure to stress test the data integrity of the flash memory. Is that what the tests did - or can you elaborate a bit on what was really tested? In any case, the performance impact boils down to how each eMMC/SD card internally manages reliable writes vs regular writes. Some vendors may treat them very similarly, while others do not. That said, trying to disable REQ_FUA from an mmc host driver is the wrong approach, as also pointed out by Adrian above. These types of decisions belong solely in the mmc core layer. Instead of what the $subject series proposes, I would rather suggest we discuss (and test) whether it could make sense to disable REQ_FUA - *if* the eMMC/SD card supports a write-back-cache (REQ_OP_FLUSH) too. Hence, the mmc core could then announce only REQ_OP_FLUSH. Kind regards Uffe
[...] > > > > Considering the data integrity, we did a random power-down test, and > > the experimental results were good. > > > > FUA can only reduce data loss under abnormal conditions, but cannot > > prevent data loss under abnormal conditions. > > > > I think there should be a balance between FUA and NO FUA, but > > filesystems seem to favor FUA. > > > > FUA brings a drop in random write performance. If enough tests are > > done, NO FUA is acceptable. > > Testing this isn't entirely easy. It requires you to hook up > electrical switches to allow you to automate the powering on/off of > the platform(s). Then at each cycle, really make sure to stress test > the data integrity of the flash memory. Is that what the tests did - > or can you elaborate a bit on what was really tested? > > In any case, the performance impact boils down to how each eMMC/SD > card internally manages reliable writes vs regular writes. Some > vendors may treat them very similarly, while others do not. > > That said, trying to disable REQ_FUA from an mmc host driver is the > wrong approach, as also pointed out by Adrian above. These types of > decisions belong solely in the mmc core layer. > > Instead of what the $subject series proposes, I would rather suggest > we discuss (and test) whether it could make sense to disable REQ_FUA - > *if* the eMMC/SD card supports a write-back-cache (REQ_OP_FLUSH) too. > Hence, the mmc core could then announce only REQ_OP_FLUSH. > Below is a simple patch that does the above. We may not want to enable this for *all* eMMC/SD cards, but it works fine for testing and to continue the discussions here. From: Ulf Hansson <ulf.hansson@linaro.org> Date: Fri, 11 Nov 2022 12:48:02 +0100 Subject: [PATCH] mmc: core: Disable REQ_FUA if the card supports an internal cache !!!! This is not for merge, but only for test and discussions!!! It has been reported that REQ_FUA can be costly on some eMMC devices. A potential option that could mitigate this problem, is to rely solely on REQ_OP_FLUSH instead, but that requires the eMMC/SD to support an internal cache. This is an attempt to try this out to see how it behaves. Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org> --- drivers/mmc/core/block.c | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/drivers/mmc/core/block.c b/drivers/mmc/core/block.c index db6d8a099910..197e9f6cdaad 100644 --- a/drivers/mmc/core/block.c +++ b/drivers/mmc/core/block.c @@ -2494,15 +2494,15 @@ static struct mmc_blk_data *mmc_blk_alloc_req(struct mmc_card *card, md->flags |= MMC_BLK_CMD23; } - if (md->flags & MMC_BLK_CMD23 && - ((card->ext_csd.rel_param & EXT_CSD_WR_REL_PARAM_EN) || - card->ext_csd.rel_sectors)) { + if (mmc_cache_enabled(card->host)) { + cache_enabled = true; + } else if (md->flags & MMC_BLK_CMD23 && + (card->ext_csd.rel_param & EXT_CSD_WR_REL_PARAM_EN || + card->ext_csd.rel_sectors)) { md->flags |= MMC_BLK_REL_WR; fua_enabled = true; cache_enabled = true; } - if (mmc_cache_enabled(card->host)) - cache_enabled = true; blk_queue_write_cache(md->queue.queue, cache_enabled, fua_enabled);
On 11/11/22 09:58, Wenchao Chen wrote: > Hi Hunter > Thank you for your review! > I'm sorry to reply you so late because I've been too busy lately. > > On Fri, Oct 21, 2022 at 11:50 PM Adrian Hunter <adrian.hunter@intel.com> wrote: >> >> On 21/10/22 10:30, Wenchao Chen wrote: >>> From: Wenchao Chen <wenchao.chen@unisoc.com> >>> >>> Summary >>> ======= >>> These patches[1] supports the host to turn off FUA. >>> >>> About FUA, roughly deal with the following two parts: >>> 1) FUA(Forced Unit Access): >>> - The REQ_FUA flag can be OR ed into the r/w flags of a bio submitted from the >>> filesystem and will make sure that I/O completion for this request is only >>> signaled after the data has been committed to non-volatile storage. >>> >>> 2) In emmc, FUA is represented as Reliable write. code show as below: >>> static void mmc_blk_data_prep(struct mmc_queue *mq, struct mmc_queue_req *mqrq, >>> int recovery_mode, bool *do_rel_wr_p, bool *do_data_tag_p) >>> { >>> ... >>> /* >>> * Reliable writes are used to implement Forced Unit Access and >>> * are supported only on MMCs. >>> */ >>> do_rel_wr = (req->cmd_flags & REQ_FUA) && >>> rq_data_dir(req) == WRITE && >>> (md->flags & MMC_BLK_REL_WR); >>> ... >>> } >>> >>> Patch structure >>> =============== >>> patch#1: for block >>> patch#2: for sdhci-sprd >>> >>> Tests >>> ===== >>> Ran 'AndroBench' to evaluate the performance: >> >> It would be good to have more details e.g. >> What file system? What block size? What journal size? >> What file size? What record size? >> > > What file system? > F2FS > What block size? > Sequential: 32768KB, Random: 4KB > What file size? > 64MB > >>> 1. fua_disable = 1 >>> /sys/block/mmcblk0/queue # cat fua 0 >>> I tested 5 times for each case and output a average speed. >>> >>> 1) Sequential read: >>> Speed: 266.8MiB/s, 265.1MiB/s, 262.9MiB/s, 268.7MiB/s, 265.2MiB/s >>> Average speed: 265.74MiB/s >>> >>> 2) Random read: >>> Speed: 98.75MiB/s, 98.7MiB/s, 98.5MiB/s, 99.4MiB/s, 98.7MiB/s >>> Average speed: 98.81MiB/s >>> >>> 3) Sequential write: >>> Speed: 199.94MiB/s, 199.1MiB/s, 205.5MiB/s, 206.5MiB/s, 191.5MiB/s >>> Average speed: 200.5MiB/s >>> >>> 4) Random write: >>> Speed: 68.6MiB/s, 71.8MiB/s, 77.1MiB/s, 64.8MiB/s, 69.3MiB/s >>> Average speed: 70.32MiB/s >>> >>> 2. fua_disable = 0 (default 0) >>> /sys/block/mmcblk0/queue # cat fua 1 >>> I tested 5 times for each case and output a average speed. >>> >>> 1) Sequential read: >>> Speed: 259.3MiB/s, 258.8MiB/s, 258.2MiB/s, 259.5MiB/s, 253.5MiB/s >>> Average speed: 257.86MiB/s >>> >>> 2) Random read: >>> Speed: 98.9MiB/s, 101MiB/s, 101MiB/s, 99MiB/s, 101.1MiB/s >>> Average speed: 100.2MiB/s >>> >>> 3) Sequential write: >>> Speed: 153.7MiB/s, 146.2MiB/s, 151.2MiB/s, 148.8MiB/s, 147.5MiB/s >>> Average speed: 149.48MiB/s >>> >>> 4) Random write: >>> Speed: 12.9MiB/s, 12.3MiB/s, 12.6MiB/s, 12.8MiB/s, 12.8MiB/s >>> Average speed: 12.68MiB/s >> >> Is every write being sync'ed of just sync at the end? >> > > /* > * Reliable writes are used to implement Forced Unit Access and > * are supported only on MMCs. > */ > do_rel_wr = (req->cmd_flags & REQ_FUA) && > rq_data_dir(req) == WRITE && > (md->flags & MMC_BLK_REL_WR); > > A Reliable Write access shall force the data to be written to the > nonvolatile storage。 > It will consume more time. Reliable write is slow because it guarantees not to tear the write. The issue is torn writes, not just FUA. > >>> >>> According to the above data, disable FUA (fua_disable = 1) improves the >>> performance: >>> 1)Sequential read improved by 3%. >>> 2)Random read were down 1%. >> >> FUA should not affect reads. If it is, you may want to investigate how. >> >>> 3)Sequential write improved by 34%. >>> 4)Random write improved by 454%. >>> Therefore, it is recommended to support the host to control FUA. >>> >>> Reference >>> ========= >>> [1] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/Documentation/block/writeback_cache_control.rst >>> [2] Embedded Multi-Media Card (e•MMC) Electrical Standard (5.1)'' >> >> You do not seem to have considered data integrity. >> >> Regular disks are assumed to provide atomic sector writes. That is, a sector has either the old data or the new data, but not some corrupt mixture. >> >> mmc does not have that assumption, which is presumably why Reliable Write has been used instead. Although that idea appears to have been thrown away for devices with no cache by commit 08ebf903af57 ("mmc: core: Fixup support for writeback-cache for eMMC and SD"). >> >> File systems can use FUA to mark a successful journal flush. Whether or not getting a torn sector at that point will corrupt the file system recovery is presumably file system specific, and maybe specific to file system options e.g. the use of checksums. >> >> It may well be that a file system can survive a torn sector at that point, or that user space would prefer to take the risk in order to get better performance. In either of those cases, it is not really a decision for the host controller driver. >> > > Considering the data integrity, we did a random power-down test, and > the experimental results were good. > > FUA can only reduce data loss under abnormal conditions, but cannot > prevent data loss under abnormal conditions. > > I think there should be a balance between FUA and NO FUA, but > filesystems seem to favor FUA. > > FUA brings a drop in random write performance. If enough tests are > done, NO FUA is acceptable. > > I found a discussion about FUA: > https://lore.kernel.org/linux-f2fs-devel/20220528051238.GX1098723@dread.disaster.area/ > > UFS reference: > https://lore.kernel.org/linux-scsi/20220531201053.3300018-1-jaegeuk@kernel.org/ > You really need buy-in from more people, especially file system developers. I suggest you try F2FS people to start with. Please be clear though: Reliable Write protects against torn writes. If enough stakeholders agree that file systems can handle the torn writes anyway, then we could presumably drop using Reliable Write for FUA. >>> >>> Wenchao Chen (2): >>> mmc: block: Support Host to control FUA >>> mmc: sdhci-sprd: enable fua_disable for SPRDSDHCI >>> >>> drivers/mmc/core/block.c | 3 ++- >>> drivers/mmc/host/sdhci-sprd.c | 2 ++ >>> include/linux/mmc/host.h | 3 +++ >>> 3 files changed, 7 insertions(+), 1 deletion(-) >>> >>
On Fri, Nov 18, 2022 at 6:12 PM Adrian Hunter <adrian.hunter@intel.com> wrote: > > On 11/11/22 09:58, Wenchao Chen wrote: > > Hi Hunter > > Thank you for your review! > > I'm sorry to reply you so late because I've been too busy lately. > > > > On Fri, Oct 21, 2022 at 11:50 PM Adrian Hunter <adrian.hunter@intel.com> wrote: > >> > >> On 21/10/22 10:30, Wenchao Chen wrote: > >>> From: Wenchao Chen <wenchao.chen@unisoc.com> > >>> > >>> Summary > >>> ======= > >>> These patches[1] supports the host to turn off FUA. > >>> > >>> About FUA, roughly deal with the following two parts: > >>> 1) FUA(Forced Unit Access): > >>> - The REQ_FUA flag can be OR ed into the r/w flags of a bio submitted from the > >>> filesystem and will make sure that I/O completion for this request is only > >>> signaled after the data has been committed to non-volatile storage. > >>> > >>> 2) In emmc, FUA is represented as Reliable write. code show as below: > >>> static void mmc_blk_data_prep(struct mmc_queue *mq, struct mmc_queue_req *mqrq, > >>> int recovery_mode, bool *do_rel_wr_p, bool *do_data_tag_p) > >>> { > >>> ... > >>> /* > >>> * Reliable writes are used to implement Forced Unit Access and > >>> * are supported only on MMCs. > >>> */ > >>> do_rel_wr = (req->cmd_flags & REQ_FUA) && > >>> rq_data_dir(req) == WRITE && > >>> (md->flags & MMC_BLK_REL_WR); > >>> ... > >>> } > >>> > >>> Patch structure > >>> =============== > >>> patch#1: for block > >>> patch#2: for sdhci-sprd > >>> > >>> Tests > >>> ===== > >>> Ran 'AndroBench' to evaluate the performance: > >> > >> It would be good to have more details e.g. > >> What file system? What block size? What journal size? > >> What file size? What record size? > >> > > > > What file system? > > F2FS > > What block size? > > Sequential: 32768KB, Random: 4KB > > What file size? > > 64MB > > > >>> 1. fua_disable = 1 > >>> /sys/block/mmcblk0/queue # cat fua 0 > >>> I tested 5 times for each case and output a average speed. > >>> > >>> 1) Sequential read: > >>> Speed: 266.8MiB/s, 265.1MiB/s, 262.9MiB/s, 268.7MiB/s, 265.2MiB/s > >>> Average speed: 265.74MiB/s > >>> > >>> 2) Random read: > >>> Speed: 98.75MiB/s, 98.7MiB/s, 98.5MiB/s, 99.4MiB/s, 98.7MiB/s > >>> Average speed: 98.81MiB/s > >>> > >>> 3) Sequential write: > >>> Speed: 199.94MiB/s, 199.1MiB/s, 205.5MiB/s, 206.5MiB/s, 191.5MiB/s > >>> Average speed: 200.5MiB/s > >>> > >>> 4) Random write: > >>> Speed: 68.6MiB/s, 71.8MiB/s, 77.1MiB/s, 64.8MiB/s, 69.3MiB/s > >>> Average speed: 70.32MiB/s > >>> > >>> 2. fua_disable = 0 (default 0) > >>> /sys/block/mmcblk0/queue # cat fua 1 > >>> I tested 5 times for each case and output a average speed. > >>> > >>> 1) Sequential read: > >>> Speed: 259.3MiB/s, 258.8MiB/s, 258.2MiB/s, 259.5MiB/s, 253.5MiB/s > >>> Average speed: 257.86MiB/s > >>> > >>> 2) Random read: > >>> Speed: 98.9MiB/s, 101MiB/s, 101MiB/s, 99MiB/s, 101.1MiB/s > >>> Average speed: 100.2MiB/s > >>> > >>> 3) Sequential write: > >>> Speed: 153.7MiB/s, 146.2MiB/s, 151.2MiB/s, 148.8MiB/s, 147.5MiB/s > >>> Average speed: 149.48MiB/s > >>> > >>> 4) Random write: > >>> Speed: 12.9MiB/s, 12.3MiB/s, 12.6MiB/s, 12.8MiB/s, 12.8MiB/s > >>> Average speed: 12.68MiB/s > >> > >> Is every write being sync'ed of just sync at the end? > >> > > > > /* > > * Reliable writes are used to implement Forced Unit Access and > > * are supported only on MMCs. > > */ > > do_rel_wr = (req->cmd_flags & REQ_FUA) && > > rq_data_dir(req) == WRITE && > > (md->flags & MMC_BLK_REL_WR); > > > > A Reliable Write access shall force the data to be written to the > > nonvolatile storage。 > > It will consume more time. > > Reliable write is slow because it guarantees not to tear the write. > The issue is torn writes, not just FUA. > If you'd like, could you introduce Reliable write that doesn't tear writes? > > > >>> > >>> According to the above data, disable FUA (fua_disable = 1) improves the > >>> performance: > >>> 1)Sequential read improved by 3%. > >>> 2)Random read were down 1%. > >> > >> FUA should not affect reads. If it is, you may want to investigate how. > >> > >>> 3)Sequential write improved by 34%. > >>> 4)Random write improved by 454%. > >>> Therefore, it is recommended to support the host to control FUA. > >>> > >>> Reference > >>> ========= > >>> [1] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/Documentation/block/writeback_cache_control.rst > >>> [2] Embedded Multi-Media Card (e•MMC) Electrical Standard (5.1)'' > >> > >> You do not seem to have considered data integrity. > >> > >> Regular disks are assumed to provide atomic sector writes. That is, a sector has either the old data or the new data, but not some corrupt mixture. > >> > >> mmc does not have that assumption, which is presumably why Reliable Write has been used instead. Although that idea appears to have been thrown away for devices with no cache by commit 08ebf903af57 ("mmc: core: Fixup support for writeback-cache for eMMC and SD"). > >> > >> File systems can use FUA to mark a successful journal flush. Whether or not getting a torn sector at that point will corrupt the file system recovery is presumably file system specific, and maybe specific to file system options e.g. the use of checksums. > >> > >> It may well be that a file system can survive a torn sector at that point, or that user space would prefer to take the risk in order to get better performance. In either of those cases, it is not really a decision for the host controller driver. > >> > > > > Considering the data integrity, we did a random power-down test, and > > the experimental results were good. > > > > FUA can only reduce data loss under abnormal conditions, but cannot > > prevent data loss under abnormal conditions. > > > > I think there should be a balance between FUA and NO FUA, but > > filesystems seem to favor FUA. > > > > FUA brings a drop in random write performance. If enough tests are > > done, NO FUA is acceptable. > > > > I found a discussion about FUA: > > https://lore.kernel.org/linux-f2fs-devel/20220528051238.GX1098723@dread.disaster.area/ > > > > UFS reference: > > https://lore.kernel.org/linux-scsi/20220531201053.3300018-1-jaegeuk@kernel.org/ > > > > You really need buy-in from more people, especially file system > developers. I suggest you try F2FS people to start with. > Please be clear though: Reliable Write protects against torn > writes. If enough stakeholders agree that file systems can > handle the torn writes anyway, then we could presumably drop > using Reliable Write for FUA. > > >>> > >>> Wenchao Chen (2): > >>> mmc: block: Support Host to control FUA > >>> mmc: sdhci-sprd: enable fua_disable for SPRDSDHCI > >>> > >>> drivers/mmc/core/block.c | 3 ++- > >>> drivers/mmc/host/sdhci-sprd.c | 2 ++ > >>> include/linux/mmc/host.h | 3 +++ > >>> 3 files changed, 7 insertions(+), 1 deletion(-) > >>> > >> >
On 18/11/22 12:54, Wenchao Chen wrote: > On Fri, Nov 18, 2022 at 6:12 PM Adrian Hunter <adrian.hunter@intel.com> wrote: >> >> On 11/11/22 09:58, Wenchao Chen wrote: >>> Hi Hunter >>> Thank you for your review! >>> I'm sorry to reply you so late because I've been too busy lately. >>> >>> On Fri, Oct 21, 2022 at 11:50 PM Adrian Hunter <adrian.hunter@intel.com> wrote: >>>> >>>> On 21/10/22 10:30, Wenchao Chen wrote: >>>>> From: Wenchao Chen <wenchao.chen@unisoc.com> >>>>> >>>>> Summary >>>>> ======= >>>>> These patches[1] supports the host to turn off FUA. >>>>> >>>>> About FUA, roughly deal with the following two parts: >>>>> 1) FUA(Forced Unit Access): >>>>> - The REQ_FUA flag can be OR ed into the r/w flags of a bio submitted from the >>>>> filesystem and will make sure that I/O completion for this request is only >>>>> signaled after the data has been committed to non-volatile storage. >>>>> >>>>> 2) In emmc, FUA is represented as Reliable write. code show as below: >>>>> static void mmc_blk_data_prep(struct mmc_queue *mq, struct mmc_queue_req *mqrq, >>>>> int recovery_mode, bool *do_rel_wr_p, bool *do_data_tag_p) >>>>> { >>>>> ... >>>>> /* >>>>> * Reliable writes are used to implement Forced Unit Access and >>>>> * are supported only on MMCs. >>>>> */ >>>>> do_rel_wr = (req->cmd_flags & REQ_FUA) && >>>>> rq_data_dir(req) == WRITE && >>>>> (md->flags & MMC_BLK_REL_WR); >>>>> ... >>>>> } >>>>> >>>>> Patch structure >>>>> =============== >>>>> patch#1: for block >>>>> patch#2: for sdhci-sprd >>>>> >>>>> Tests >>>>> ===== >>>>> Ran 'AndroBench' to evaluate the performance: >>>> >>>> It would be good to have more details e.g. >>>> What file system? What block size? What journal size? >>>> What file size? What record size? >>>> >>> >>> What file system? >>> F2FS >>> What block size? >>> Sequential: 32768KB, Random: 4KB >>> What file size? >>> 64MB >>> >>>>> 1. fua_disable = 1 >>>>> /sys/block/mmcblk0/queue # cat fua 0 >>>>> I tested 5 times for each case and output a average speed. >>>>> >>>>> 1) Sequential read: >>>>> Speed: 266.8MiB/s, 265.1MiB/s, 262.9MiB/s, 268.7MiB/s, 265.2MiB/s >>>>> Average speed: 265.74MiB/s >>>>> >>>>> 2) Random read: >>>>> Speed: 98.75MiB/s, 98.7MiB/s, 98.5MiB/s, 99.4MiB/s, 98.7MiB/s >>>>> Average speed: 98.81MiB/s >>>>> >>>>> 3) Sequential write: >>>>> Speed: 199.94MiB/s, 199.1MiB/s, 205.5MiB/s, 206.5MiB/s, 191.5MiB/s >>>>> Average speed: 200.5MiB/s >>>>> >>>>> 4) Random write: >>>>> Speed: 68.6MiB/s, 71.8MiB/s, 77.1MiB/s, 64.8MiB/s, 69.3MiB/s >>>>> Average speed: 70.32MiB/s >>>>> >>>>> 2. fua_disable = 0 (default 0) >>>>> /sys/block/mmcblk0/queue # cat fua 1 >>>>> I tested 5 times for each case and output a average speed. >>>>> >>>>> 1) Sequential read: >>>>> Speed: 259.3MiB/s, 258.8MiB/s, 258.2MiB/s, 259.5MiB/s, 253.5MiB/s >>>>> Average speed: 257.86MiB/s >>>>> >>>>> 2) Random read: >>>>> Speed: 98.9MiB/s, 101MiB/s, 101MiB/s, 99MiB/s, 101.1MiB/s >>>>> Average speed: 100.2MiB/s >>>>> >>>>> 3) Sequential write: >>>>> Speed: 153.7MiB/s, 146.2MiB/s, 151.2MiB/s, 148.8MiB/s, 147.5MiB/s >>>>> Average speed: 149.48MiB/s >>>>> >>>>> 4) Random write: >>>>> Speed: 12.9MiB/s, 12.3MiB/s, 12.6MiB/s, 12.8MiB/s, 12.8MiB/s >>>>> Average speed: 12.68MiB/s >>>> >>>> Is every write being sync'ed of just sync at the end? >>>> >>> >>> /* >>> * Reliable writes are used to implement Forced Unit Access and >>> * are supported only on MMCs. >>> */ >>> do_rel_wr = (req->cmd_flags & REQ_FUA) && >>> rq_data_dir(req) == WRITE && >>> (md->flags & MMC_BLK_REL_WR); >>> >>> A Reliable Write access shall force the data to be written to the >>> nonvolatile storage。 >>> It will consume more time. >> >> Reliable write is slow because it guarantees not to tear the write. >> The issue is torn writes, not just FUA. >> > > If you'd like, could you introduce Reliable write that doesn't tear writes? Not following you. Reliable Write doesn't tear writes, that is why it is used. > >>> >>>>> >>>>> According to the above data, disable FUA (fua_disable = 1) improves the >>>>> performance: >>>>> 1)Sequential read improved by 3%. >>>>> 2)Random read were down 1%. >>>> >>>> FUA should not affect reads. If it is, you may want to investigate how. >>>> >>>>> 3)Sequential write improved by 34%. >>>>> 4)Random write improved by 454%. >>>>> Therefore, it is recommended to support the host to control FUA. >>>>> >>>>> Reference >>>>> ========= >>>>> [1] https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/Documentation/block/writeback_cache_control.rst >>>>> [2] Embedded Multi-Media Card (e•MMC) Electrical Standard (5.1)'' >>>> >>>> You do not seem to have considered data integrity. >>>> >>>> Regular disks are assumed to provide atomic sector writes. That is, a sector has either the old data or the new data, but not some corrupt mixture. >>>> >>>> mmc does not have that assumption, which is presumably why Reliable Write has been used instead. Although that idea appears to have been thrown away for devices with no cache by commit 08ebf903af57 ("mmc: core: Fixup support for writeback-cache for eMMC and SD"). >>>> >>>> File systems can use FUA to mark a successful journal flush. Whether or not getting a torn sector at that point will corrupt the file system recovery is presumably file system specific, and maybe specific to file system options e.g. the use of checksums. >>>> >>>> It may well be that a file system can survive a torn sector at that point, or that user space would prefer to take the risk in order to get better performance. In either of those cases, it is not really a decision for the host controller driver. >>>> >>> >>> Considering the data integrity, we did a random power-down test, and >>> the experimental results were good. >>> >>> FUA can only reduce data loss under abnormal conditions, but cannot >>> prevent data loss under abnormal conditions. >>> >>> I think there should be a balance between FUA and NO FUA, but >>> filesystems seem to favor FUA. >>> >>> FUA brings a drop in random write performance. If enough tests are >>> done, NO FUA is acceptable. >>> >>> I found a discussion about FUA: >>> https://lore.kernel.org/linux-f2fs-devel/20220528051238.GX1098723@dread.disaster.area/ >>> >>> UFS reference: >>> https://lore.kernel.org/linux-scsi/20220531201053.3300018-1-jaegeuk@kernel.org/ >>> >> >> You really need buy-in from more people, especially file system >> developers. I suggest you try F2FS people to start with. >> Please be clear though: Reliable Write protects against torn >> writes. If enough stakeholders agree that file systems can >> handle the torn writes anyway, then we could presumably drop >> using Reliable Write for FUA. >> >>>>> >>>>> Wenchao Chen (2): >>>>> mmc: block: Support Host to control FUA >>>>> mmc: sdhci-sprd: enable fua_disable for SPRDSDHCI >>>>> >>>>> drivers/mmc/core/block.c | 3 ++- >>>>> drivers/mmc/host/sdhci-sprd.c | 2 ++ >>>>> include/linux/mmc/host.h | 3 +++ >>>>> 3 files changed, 7 insertions(+), 1 deletion(-) >>>>> >>>> >>
On Fri, 11 Nov 2022 at 13:04, Ulf Hansson <ulf.hansson@linaro.org> wrote: > > [...] > > > > > > > Considering the data integrity, we did a random power-down test, and > > > the experimental results were good. > > > > > > FUA can only reduce data loss under abnormal conditions, but cannot > > > prevent data loss under abnormal conditions. > > > > > > I think there should be a balance between FUA and NO FUA, but > > > filesystems seem to favor FUA. > > > > > > FUA brings a drop in random write performance. If enough tests are > > > done, NO FUA is acceptable. > > > > Testing this isn't entirely easy. It requires you to hook up > > electrical switches to allow you to automate the powering on/off of > > the platform(s). Then at each cycle, really make sure to stress test > > the data integrity of the flash memory. Is that what the tests did - > > or can you elaborate a bit on what was really tested? > > > > In any case, the performance impact boils down to how each eMMC/SD > > card internally manages reliable writes vs regular writes. Some > > vendors may treat them very similarly, while others do not. > > > > That said, trying to disable REQ_FUA from an mmc host driver is the > > wrong approach, as also pointed out by Adrian above. These types of > > decisions belong solely in the mmc core layer. > > > > Instead of what the $subject series proposes, I would rather suggest > > we discuss (and test) whether it could make sense to disable REQ_FUA - > > *if* the eMMC/SD card supports a write-back-cache (REQ_OP_FLUSH) too. > > Hence, the mmc core could then announce only REQ_OP_FLUSH. > > > > Below is a simple patch that does the above. We may not want to enable > this for *all* eMMC/SD cards, but it works fine for testing and to > continue the discussions here. > > > From: Ulf Hansson <ulf.hansson@linaro.org> > Date: Fri, 11 Nov 2022 12:48:02 +0100 > Subject: [PATCH] mmc: core: Disable REQ_FUA if the card supports an internal > cache > > !!!! This is not for merge, but only for test and discussions!!! > > It has been reported that REQ_FUA can be costly on some eMMC devices. A > potential option that could mitigate this problem, is to rely solely on > REQ_OP_FLUSH instead, but that requires the eMMC/SD to support an internal > cache. This is an attempt to try this out to see how it behaves. > > Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org> > --- > drivers/mmc/core/block.c | 10 +++++----- > 1 file changed, 5 insertions(+), 5 deletions(-) > > diff --git a/drivers/mmc/core/block.c b/drivers/mmc/core/block.c > index db6d8a099910..197e9f6cdaad 100644 > --- a/drivers/mmc/core/block.c > +++ b/drivers/mmc/core/block.c > @@ -2494,15 +2494,15 @@ static struct mmc_blk_data > *mmc_blk_alloc_req(struct mmc_card *card, > md->flags |= MMC_BLK_CMD23; > } > > - if (md->flags & MMC_BLK_CMD23 && > - ((card->ext_csd.rel_param & EXT_CSD_WR_REL_PARAM_EN) || > - card->ext_csd.rel_sectors)) { > + if (mmc_cache_enabled(card->host)) { > + cache_enabled = true; > + } else if (md->flags & MMC_BLK_CMD23 && > + (card->ext_csd.rel_param & EXT_CSD_WR_REL_PARAM_EN || > + card->ext_csd.rel_sectors)) { > md->flags |= MMC_BLK_REL_WR; > fua_enabled = true; > cache_enabled = true; > } > - if (mmc_cache_enabled(card->host)) > - cache_enabled = true; > > blk_queue_write_cache(md->queue.queue, cache_enabled, fua_enabled); > > -- > 2.34.1 Wenchao, Did you manage to try the above patch to see if that could improve the situation? Kind regards Uffe
On Thu, Feb 9, 2023 at 10:51 PM Ulf Hansson <ulf.hansson@linaro.org> wrote: > > On Fri, 11 Nov 2022 at 13:04, Ulf Hansson <ulf.hansson@linaro.org> wrote: > > > > [...] > > > > > > > > > > Considering the data integrity, we did a random power-down test, and > > > > the experimental results were good. > > > > > > > > FUA can only reduce data loss under abnormal conditions, but cannot > > > > prevent data loss under abnormal conditions. > > > > > > > > I think there should be a balance between FUA and NO FUA, but > > > > filesystems seem to favor FUA. > > > > > > > > FUA brings a drop in random write performance. If enough tests are > > > > done, NO FUA is acceptable. > > > > > > Testing this isn't entirely easy. It requires you to hook up > > > electrical switches to allow you to automate the powering on/off of > > > the platform(s). Then at each cycle, really make sure to stress test > > > the data integrity of the flash memory. Is that what the tests did - > > > or can you elaborate a bit on what was really tested? > > > > > > In any case, the performance impact boils down to how each eMMC/SD > > > card internally manages reliable writes vs regular writes. Some > > > vendors may treat them very similarly, while others do not. > > > > > > That said, trying to disable REQ_FUA from an mmc host driver is the > > > wrong approach, as also pointed out by Adrian above. These types of > > > decisions belong solely in the mmc core layer. > > > > > > Instead of what the $subject series proposes, I would rather suggest > > > we discuss (and test) whether it could make sense to disable REQ_FUA - > > > *if* the eMMC/SD card supports a write-back-cache (REQ_OP_FLUSH) too. > > > Hence, the mmc core could then announce only REQ_OP_FLUSH. > > > > > > > Below is a simple patch that does the above. We may not want to enable > > this for *all* eMMC/SD cards, but it works fine for testing and to > > continue the discussions here. > > > > > > From: Ulf Hansson <ulf.hansson@linaro.org> > > Date: Fri, 11 Nov 2022 12:48:02 +0100 > > Subject: [PATCH] mmc: core: Disable REQ_FUA if the card supports an internal > > cache > > > > !!!! This is not for merge, but only for test and discussions!!! > > > > It has been reported that REQ_FUA can be costly on some eMMC devices. A > > potential option that could mitigate this problem, is to rely solely on > > REQ_OP_FLUSH instead, but that requires the eMMC/SD to support an internal > > cache. This is an attempt to try this out to see how it behaves. > > > > Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org> > > --- > > drivers/mmc/core/block.c | 10 +++++----- > > 1 file changed, 5 insertions(+), 5 deletions(-) > > > > diff --git a/drivers/mmc/core/block.c b/drivers/mmc/core/block.c > > index db6d8a099910..197e9f6cdaad 100644 > > --- a/drivers/mmc/core/block.c > > +++ b/drivers/mmc/core/block.c > > @@ -2494,15 +2494,15 @@ static struct mmc_blk_data > > *mmc_blk_alloc_req(struct mmc_card *card, > > md->flags |= MMC_BLK_CMD23; > > } > > > > - if (md->flags & MMC_BLK_CMD23 && > > - ((card->ext_csd.rel_param & EXT_CSD_WR_REL_PARAM_EN) || > > - card->ext_csd.rel_sectors)) { > > + if (mmc_cache_enabled(card->host)) { > > + cache_enabled = true; > > + } else if (md->flags & MMC_BLK_CMD23 && > > + (card->ext_csd.rel_param & EXT_CSD_WR_REL_PARAM_EN || > > + card->ext_csd.rel_sectors)) { > > md->flags |= MMC_BLK_REL_WR; > > fua_enabled = true; > > cache_enabled = true; > > } > > - if (mmc_cache_enabled(card->host)) > > - cache_enabled = true; > > > > blk_queue_write_cache(md->queue.queue, cache_enabled, fua_enabled); > > > > -- > > 2.34.1 > > Wenchao, > > Did you manage to try the above patch to see if that could improve the > situation? > Hi Uffe, Yes, it can solve my problem. Thank you very much. > Kind regards > Uffe
On Mon, 13 Feb 2023 at 09:21, Wenchao Chen <wenchao.chen666@gmail.com> wrote: > > On Thu, Feb 9, 2023 at 10:51 PM Ulf Hansson <ulf.hansson@linaro.org> wrote: > > > > On Fri, 11 Nov 2022 at 13:04, Ulf Hansson <ulf.hansson@linaro.org> wrote: > > > > > > [...] > > > > > > > > > > > > > Considering the data integrity, we did a random power-down test, and > > > > > the experimental results were good. > > > > > > > > > > FUA can only reduce data loss under abnormal conditions, but cannot > > > > > prevent data loss under abnormal conditions. > > > > > > > > > > I think there should be a balance between FUA and NO FUA, but > > > > > filesystems seem to favor FUA. > > > > > > > > > > FUA brings a drop in random write performance. If enough tests are > > > > > done, NO FUA is acceptable. > > > > > > > > Testing this isn't entirely easy. It requires you to hook up > > > > electrical switches to allow you to automate the powering on/off of > > > > the platform(s). Then at each cycle, really make sure to stress test > > > > the data integrity of the flash memory. Is that what the tests did - > > > > or can you elaborate a bit on what was really tested? > > > > > > > > In any case, the performance impact boils down to how each eMMC/SD > > > > card internally manages reliable writes vs regular writes. Some > > > > vendors may treat them very similarly, while others do not. > > > > > > > > That said, trying to disable REQ_FUA from an mmc host driver is the > > > > wrong approach, as also pointed out by Adrian above. These types of > > > > decisions belong solely in the mmc core layer. > > > > > > > > Instead of what the $subject series proposes, I would rather suggest > > > > we discuss (and test) whether it could make sense to disable REQ_FUA - > > > > *if* the eMMC/SD card supports a write-back-cache (REQ_OP_FLUSH) too. > > > > Hence, the mmc core could then announce only REQ_OP_FLUSH. > > > > > > > > > > Below is a simple patch that does the above. We may not want to enable > > > this for *all* eMMC/SD cards, but it works fine for testing and to > > > continue the discussions here. > > > > > > > > > From: Ulf Hansson <ulf.hansson@linaro.org> > > > Date: Fri, 11 Nov 2022 12:48:02 +0100 > > > Subject: [PATCH] mmc: core: Disable REQ_FUA if the card supports an internal > > > cache > > > > > > !!!! This is not for merge, but only for test and discussions!!! > > > > > > It has been reported that REQ_FUA can be costly on some eMMC devices. A > > > potential option that could mitigate this problem, is to rely solely on > > > REQ_OP_FLUSH instead, but that requires the eMMC/SD to support an internal > > > cache. This is an attempt to try this out to see how it behaves. > > > > > > Signed-off-by: Ulf Hansson <ulf.hansson@linaro.org> > > > --- > > > drivers/mmc/core/block.c | 10 +++++----- > > > 1 file changed, 5 insertions(+), 5 deletions(-) > > > > > > diff --git a/drivers/mmc/core/block.c b/drivers/mmc/core/block.c > > > index db6d8a099910..197e9f6cdaad 100644 > > > --- a/drivers/mmc/core/block.c > > > +++ b/drivers/mmc/core/block.c > > > @@ -2494,15 +2494,15 @@ static struct mmc_blk_data > > > *mmc_blk_alloc_req(struct mmc_card *card, > > > md->flags |= MMC_BLK_CMD23; > > > } > > > > > > - if (md->flags & MMC_BLK_CMD23 && > > > - ((card->ext_csd.rel_param & EXT_CSD_WR_REL_PARAM_EN) || > > > - card->ext_csd.rel_sectors)) { > > > + if (mmc_cache_enabled(card->host)) { > > > + cache_enabled = true; > > > + } else if (md->flags & MMC_BLK_CMD23 && > > > + (card->ext_csd.rel_param & EXT_CSD_WR_REL_PARAM_EN || > > > + card->ext_csd.rel_sectors)) { > > > md->flags |= MMC_BLK_REL_WR; > > > fua_enabled = true; > > > cache_enabled = true; > > > } > > > - if (mmc_cache_enabled(card->host)) > > > - cache_enabled = true; > > > > > > blk_queue_write_cache(md->queue.queue, cache_enabled, fua_enabled); > > > > > > -- > > > 2.34.1 > > > > Wenchao, > > > > Did you manage to try the above patch to see if that could improve the > > situation? > > > > Hi Uffe, > Yes, it can solve my problem. Thank you very much. Okay, that's very interesting news. I will prepare a formal patch and make a new submission soon. Let's continue the discussion then. Kind regards Uffe