Message ID | 20231027000525.1278806-6-tina.zhang@intel.com |
---|---|
State | New |
Headers |
Return-Path: <linux-kernel-owner@vger.kernel.org> Delivered-To: ouuuleilei@gmail.com Received: by 2002:a59:d641:0:b0:403:3b70:6f57 with SMTP id cy1csp251399vqb; Thu, 26 Oct 2023 17:06:44 -0700 (PDT) X-Google-Smtp-Source: AGHT+IEOtYAk+5valqnfV99yN3DaS/4p5fnshfWUph00kE/dyxyc8nfAhKAWUb0i904CRG1e15mI X-Received: by 2002:a25:2611:0:b0:da0:275a:25fb with SMTP id m17-20020a252611000000b00da0275a25fbmr2439613ybm.9.1698365204603; Thu, 26 Oct 2023 17:06:44 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1698365204; cv=none; d=google.com; s=arc-20160816; b=i21OeRgwJ/k3s/ihSHjv2WXEp98yRAeGNeC60mYZqM2/B/fB3Jfab0UdWzygzAmokT uAS8feiYAu7t5raAHSWEqJMUE5auG/aSmsB5se1oOEVZfCUCky/SFzv0ADMrU/o1yWME KPbBOZGvn5hA9kE9hP6YVvoyhui1Ldx+lNAS5DEVB/dpYsDETcLY6BoMR/XshaQLPtqU h7/5V07dkXhyNIHkQYOIfFq2TnZQfUXc9xz55qtANSdZoXZJbrpRaFfdPQBZlAYWUJXq 9e7dw/02tfabpMwYkMITiTje26Ijk8lihv5mgRqO3HHnWsk1hXKO4h0sTSU38y+AmWeu gcRA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from :dkim-signature; bh=hlwnI1mCiN+GQ+6nPHmyY7UWujIyg/zExdVh9v4Q4+o=; fh=dJBxdn+RXcUABM1aUIOqfrsndRxIHilofn8/gNUoPfs=; b=0A0v0HUKot7HvR+1QQ71Es9tfe8gSR8qeqZzymEuKlQ/KYsJZkQA0My4928m+nuvYa 4jfAZOMxVttZ3qMWlx0lDDxUEjnlfmkFE4HWHNTDj64OOCfUYRINH/ltR0UaphpNqDmI Lq8btDQEW5z0bXAKsBK3eQ4pDG/wfcA4vsFyCuoNVqnxicScA9UK7hSEpcqoTjUPTeC0 xzmSCgilRmYan99zFFb5AKU943lmQiYAOR2RBNWwcFKgMk2O3d4ojtVaT4oFeRXP9Xv+ keMgMK8IfV2TGK+dj8oFqCy97cDmUNGeZe4VmvApoljRSyyVE0T3FO3/C9Jvf4L2aWsZ Qstg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@intel.com header.s=Intel header.b=b0fRLztR; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.32 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: from agentk.vger.email (agentk.vger.email. [23.128.96.32]) by mx.google.com with ESMTPS id r17-20020a5b0191000000b00d3cb701f4adsi689731ybl.231.2023.10.26.17.06.44 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 26 Oct 2023 17:06:44 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.32 as permitted sender) client-ip=23.128.96.32; Authentication-Results: mx.google.com; dkim=pass header.i=@intel.com header.s=Intel header.b=b0fRLztR; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.32 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: from out1.vger.email (depot.vger.email [IPv6:2620:137:e000::3:0]) by agentk.vger.email (Postfix) with ESMTP id A838081DEEC6; Thu, 26 Oct 2023 17:06:41 -0700 (PDT) X-Virus-Status: Clean X-Virus-Scanned: clamav-milter 0.103.10 at agentk.vger.email Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S235074AbjJ0AGe (ORCPT <rfc822;aposhian.dev@gmail.com> + 26 others); Thu, 26 Oct 2023 20:06:34 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:38252 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1345093AbjJ0AGZ (ORCPT <rfc822;linux-kernel@vger.kernel.org>); Thu, 26 Oct 2023 20:06:25 -0400 Received: from mgamail.intel.com (mgamail.intel.com [134.134.136.24]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 940C110FD for <linux-kernel@vger.kernel.org>; Thu, 26 Oct 2023 17:06:18 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1698365178; x=1729901178; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=L9eyFj95pKnzjt3LxhfE3TFhO0038AM7EKiLfkWLEes=; b=b0fRLztRCeTgUN8Fm8Xz0YgqDdMT4dnFBIv42iX00CBmUd1t9dQhjEFX qm7qUqz5AdUdIhcyiX+KIZeyD3k6hbr7QoQ3kjDD5eEu69+9cluUOgAEQ utZ1YmaFTpHo7OX0oUptCbolWriG8IewfuM43ioBkGqhgXDUnKXJnjMsf NTEGXGmA5N52m3Qwby0V36kpvQnDWqCebRn5cug8vIed5Eby8kEaXLZ7J 6oXqrAXZ8d86jJlflLvez3JBoKdjQwlUseK7UEAMiUM99bS+Nv6NNgjpe 8Ar/dfs2+R8oZmusmeNRRB8uHL8oP+dGwp+qtfBvYObGvxUk6QofbAUc1 g==; X-IronPort-AV: E=McAfee;i="6600,9927,10875"; a="390536030" X-IronPort-AV: E=Sophos;i="6.03,255,1694761200"; d="scan'208";a="390536030" Received: from orsmga004.jf.intel.com ([10.7.209.38]) by orsmga102.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 26 Oct 2023 17:06:18 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6600,9927,10875"; a="883017517" X-IronPort-AV: E=Sophos;i="6.03,255,1694761200"; d="scan'208";a="883017517" Received: from sqa-gate.sh.intel.com (HELO spr-2s5.tsp.org) ([10.239.48.212]) by orsmga004.jf.intel.com with ESMTP; 26 Oct 2023 17:06:14 -0700 From: Tina Zhang <tina.zhang@intel.com> To: iommu@lists.linux.dev, linux-kernel@vger.kernel.org Cc: David Woodhouse <dwmw2@infradead.org>, Lu Baolu <baolu.lu@linux.intel.com>, Joerg Roedel <joro@8bytes.org>, Will Deacon <will@kernel.org>, Robin Murphy <robin.murphy@arm.com>, Jason Gunthorpe <jgg@ziepe.ca>, Kevin Tian <kevin.tian@intel.com>, Nicolin Chen <nicolinc@nvidia.com>, Michael Shavit <mshavit@google.com>, Vasant Hegde <vasant.hegde@amd.com>, Tina Zhang <tina.zhang@intel.com>, Jason Gunthorpe <jgg@nvidia.com> Subject: [PATCH v10 5/6] iommu: Support mm PASID 1:n with sva domains Date: Fri, 27 Oct 2023 08:05:24 +0800 Message-Id: <20231027000525.1278806-6-tina.zhang@intel.com> X-Mailer: git-send-email 2.39.3 In-Reply-To: <20231027000525.1278806-1-tina.zhang@intel.com> References: <20231027000525.1278806-1-tina.zhang@intel.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Spam-Status: No, score=-0.9 required=5.0 tests=DKIMWL_WL_HIGH,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on agentk.vger.email Precedence: bulk List-ID: <linux-kernel.vger.kernel.org> X-Mailing-List: linux-kernel@vger.kernel.org X-Greylist: Sender passed SPF test, not delayed by milter-greylist-4.6.4 (agentk.vger.email [0.0.0.0]); Thu, 26 Oct 2023 17:06:41 -0700 (PDT) X-getmail-retrieved-from-mailbox: INBOX X-GMAIL-THRID: 1780864992887764868 X-GMAIL-MSGID: 1780864992887764868 |
Series |
Share sva domains with all devices bound to a mm
|
|
Commit Message
Zhang, Tina
Oct. 27, 2023, 12:05 a.m. UTC
Each mm bound to devices gets a PASID and corresponding sva domains allocated in iommu_sva_bind_device(), which are referenced by iommu_mm field of the mm. The PASID is released in __mmdrop(), while a sva domain is released when no one is using it (the reference count is decremented in iommu_sva_unbind_device()). However, although sva domains and their PASID are separate objects such that their own life cycles could be handled independently, an enqcmd use case may require releasing the PASID in releasing the mm (i.e., once a PASID is allocated for a mm, it will be permanently used by the mm and won't be released until the end of mm) and only allows to drop the PASID after the sva domains are released. To this end, mmgrab() is called in iommu_sva_domain_alloc() to increment the mm reference count and mmdrop() is invoked in iommu_domain_free() to decrement the mm reference count. Since the required info of PASID and sva domains is kept in struct iommu_mm_data of a mm, use mm->iommu_mm field instead of the old pasid field in mm struct. The sva domain list is protected by iommu_sva_lock. Besides, this patch removes mm_pasid_init(), as with the introduced iommu_mm structure, initializing mm pasid in mm_init() is unnecessary. Reviewed-by: Lu Baolu <baolu.lu@linux.intel.com> Reviewed-by: Vasant Hegde <vasant.hegde@amd.com> Reviewed-by: Jason Gunthorpe <jgg@nvidia.com> Tested-by: Nicolin Chen <nicolinc@nvidia.com> Signed-off-by: Tina Zhang <tina.zhang@intel.com> Signed-off-by: Jason Gunthorpe <jgg@nvidia.com> --- drivers/iommu/iommu-sva.c | 92 +++++++++++++++++++++++---------------- include/linux/iommu.h | 23 ++++++++-- 2 files changed, 74 insertions(+), 41 deletions(-)
Comments
Hi, Tina On Fri, 27 Oct 2023 at 08:06, Tina Zhang <tina.zhang@intel.com> wrote: > > Each mm bound to devices gets a PASID and corresponding sva domains > allocated in iommu_sva_bind_device(), which are referenced by iommu_mm > field of the mm. The PASID is released in __mmdrop(), while a sva domain > is released when no one is using it (the reference count is decremented > in iommu_sva_unbind_device()). However, although sva domains and their > PASID are separate objects such that their own life cycles could be > handled independently, an enqcmd use case may require releasing the > PASID in releasing the mm (i.e., once a PASID is allocated for a mm, it > will be permanently used by the mm and won't be released until the end > of mm) and only allows to drop the PASID after the sva domains are > released. To this end, mmgrab() is called in iommu_sva_domain_alloc() to > increment the mm reference count and mmdrop() is invoked in > iommu_domain_free() to decrement the mm reference count. > > Since the required info of PASID and sva domains is kept in struct > iommu_mm_data of a mm, use mm->iommu_mm field instead of the old pasid > field in mm struct. The sva domain list is protected by iommu_sva_lock. > > Besides, this patch removes mm_pasid_init(), as with the introduced > iommu_mm structure, initializing mm pasid in mm_init() is unnecessary. > > Reviewed-by: Lu Baolu <baolu.lu@linux.intel.com> > Reviewed-by: Vasant Hegde <vasant.hegde@amd.com> > Reviewed-by: Jason Gunthorpe <jgg@nvidia.com> > Tested-by: Nicolin Chen <nicolinc@nvidia.com> > Signed-off-by: Tina Zhang <tina.zhang@intel.com> > Signed-off-by: Jason Gunthorpe <jgg@nvidia.com> > --- > drivers/iommu/iommu-sva.c | 92 +++++++++++++++++++++++---------------- > include/linux/iommu.h | 23 ++++++++-- > 2 files changed, 74 insertions(+), 41 deletions(-) > > diff --git a/drivers/iommu/iommu-sva.c b/drivers/iommu/iommu-sva.c > index 4a2f5699747f..5175e8d85247 100644 > --- a/drivers/iommu/iommu-sva.c > +++ b/drivers/iommu/iommu-sva.c > @@ -12,32 +12,42 @@ > static DEFINE_MUTEX(iommu_sva_lock); > > /* Allocate a PASID for the mm within range (inclusive) */ > -static int iommu_sva_alloc_pasid(struct mm_struct *mm, struct device *dev) > +static struct iommu_mm_data *iommu_alloc_mm_data(struct mm_struct *mm, struct device *dev) > { > + struct iommu_mm_data *iommu_mm; > ioasid_t pasid; > - int ret = 0; > + > + lockdep_assert_held(&iommu_sva_lock); > > if (!arch_pgtable_dma_compat(mm)) > - return -EBUSY; > + return ERR_PTR(-EBUSY); > > - mutex_lock(&iommu_sva_lock); > + iommu_mm = mm->iommu_mm; > /* Is a PASID already associated with this mm? */ > - if (mm_valid_pasid(mm)) { > - if (mm->pasid >= dev->iommu->max_pasids) > - ret = -EOVERFLOW; > - goto out; > + if (iommu_mm) { > + if (iommu_mm->pasid >= dev->iommu->max_pasids) > + return ERR_PTR(-EOVERFLOW); > + return iommu_mm; > } > > + iommu_mm = kzalloc(sizeof(struct iommu_mm_data), GFP_KERNEL); > + if (!iommu_mm) > + return ERR_PTR(-ENOMEM); > + > pasid = iommu_alloc_global_pasid(dev); > if (pasid == IOMMU_PASID_INVALID) { > - ret = -ENOSPC; > - goto out; > + kfree(iommu_mm); > + return ERR_PTR(-ENOSPC); > } > - mm->pasid = pasid; > - ret = 0; > -out: > - mutex_unlock(&iommu_sva_lock); > - return ret; > + iommu_mm->pasid = pasid; > + INIT_LIST_HEAD(&iommu_mm->sva_domains); > + /* > + * Make sure the write to mm->iommu_mm is not reordered in front of > + * initialization to iommu_mm fields. If it does, readers may see a > + * valid iommu_mm with uninitialized values. > + */ > + smp_store_release(&mm->iommu_mm, iommu_mm); > + return iommu_mm; > } > > /** > @@ -58,31 +68,33 @@ static int iommu_sva_alloc_pasid(struct mm_struct *mm, struct device *dev) > */ > struct iommu_sva *iommu_sva_bind_device(struct device *dev, struct mm_struct *mm) > { > + struct iommu_mm_data *iommu_mm; > struct iommu_domain *domain; > struct iommu_sva *handle; > int ret; > > + mutex_lock(&iommu_sva_lock); > + > /* Allocate mm->pasid if necessary. */ > - ret = iommu_sva_alloc_pasid(mm, dev); > - if (ret) > - return ERR_PTR(ret); > + iommu_mm = iommu_alloc_mm_data(mm, dev); > + if (IS_ERR(iommu_mm)) { > + ret = PTR_ERR(iommu_mm); > + goto out_unlock; > + } > > handle = kzalloc(sizeof(*handle), GFP_KERNEL); > - if (!handle) > - return ERR_PTR(-ENOMEM); > - > - mutex_lock(&iommu_sva_lock); > - /* Search for an existing domain. */ > - domain = iommu_get_domain_for_dev_pasid(dev, mm->pasid, > - IOMMU_DOMAIN_SVA); > - if (IS_ERR(domain)) { > - ret = PTR_ERR(domain); > + if (!handle) { > + ret = -ENOMEM; > goto out_unlock; > } > > - if (domain) { > - domain->users++; > - goto out; Our multi bind test case broke since 6.8-rc1. The test case can use same domain & pasid, return different handle, 6.7 simply domain->users ++ and return. > + /* Search for an existing domain. */ > + list_for_each_entry(domain, &mm->iommu_mm->sva_domains, next) { > + ret = iommu_attach_device_pasid(domain, dev, iommu_mm->pasid); Now iommu_attach_device_pasid return BUSY since the same pasid. And then iommu_sva_bind_device attach ret=-16 > + if (!ret) { Simply tried if (!ret || ret == -EBUSY) The test passes, but report waring WARNING: CPU: 12 PID: 2992 at drivers/iommu/iommu.c:3591 iommu_detach_device_pasid+0xa4/0xd0 Will check more tomorrow. > + domain->users++; > + goto out; > + } > } > Thanks
Hi Zhangfei, > -----Original Message----- > From: Zhangfei Gao <zhangfei.gao@linaro.org> > Sent: Wednesday, February 21, 2024 12:27 AM > To: Zhang, Tina <tina.zhang@intel.com> > Cc: iommu@lists.linux.dev; linux-kernel@vger.kernel.org; David Woodhouse > <dwmw2@infradead.org>; Lu Baolu <baolu.lu@linux.intel.com>; Joerg > Roedel <joro@8bytes.org>; Will Deacon <will@kernel.org>; Robin Murphy > <robin.murphy@arm.com>; Jason Gunthorpe <jgg@ziepe.ca>; Tian, Kevin > <kevin.tian@intel.com>; Nicolin Chen <nicolinc@nvidia.com>; Michael Shavit > <mshavit@google.com>; Vasant Hegde <vasant.hegde@amd.com>; Jason > Gunthorpe <jgg@nvidia.com> > Subject: Re: [PATCH v10 5/6] iommu: Support mm PASID 1:n with sva > domains > > Hi, Tina > > On Fri, 27 Oct 2023 at 08:06, Tina Zhang <tina.zhang@intel.com> wrote: > > > > Each mm bound to devices gets a PASID and corresponding sva domains > > allocated in iommu_sva_bind_device(), which are referenced by > iommu_mm > > field of the mm. The PASID is released in __mmdrop(), while a sva > > domain is released when no one is using it (the reference count is > > decremented in iommu_sva_unbind_device()). However, although sva > > domains and their PASID are separate objects such that their own life > > cycles could be handled independently, an enqcmd use case may require > > releasing the PASID in releasing the mm (i.e., once a PASID is > > allocated for a mm, it will be permanently used by the mm and won't be > > released until the end of mm) and only allows to drop the PASID after > > the sva domains are released. To this end, mmgrab() is called in > > iommu_sva_domain_alloc() to increment the mm reference count and > > mmdrop() is invoked in > > iommu_domain_free() to decrement the mm reference count. > > > > Since the required info of PASID and sva domains is kept in struct > > iommu_mm_data of a mm, use mm->iommu_mm field instead of the old > pasid > > field in mm struct. The sva domain list is protected by iommu_sva_lock. > > > > Besides, this patch removes mm_pasid_init(), as with the introduced > > iommu_mm structure, initializing mm pasid in mm_init() is unnecessary. > > > > Reviewed-by: Lu Baolu <baolu.lu@linux.intel.com> > > Reviewed-by: Vasant Hegde <vasant.hegde@amd.com> > > Reviewed-by: Jason Gunthorpe <jgg@nvidia.com> > > Tested-by: Nicolin Chen <nicolinc@nvidia.com> > > Signed-off-by: Tina Zhang <tina.zhang@intel.com> > > Signed-off-by: Jason Gunthorpe <jgg@nvidia.com> > > --- > > drivers/iommu/iommu-sva.c | 92 +++++++++++++++++++++++---------------- > > include/linux/iommu.h | 23 ++++++++-- > > 2 files changed, 74 insertions(+), 41 deletions(-) > > > > diff --git a/drivers/iommu/iommu-sva.c b/drivers/iommu/iommu-sva.c > > index 4a2f5699747f..5175e8d85247 100644 > > --- a/drivers/iommu/iommu-sva.c > > +++ b/drivers/iommu/iommu-sva.c > > @@ -12,32 +12,42 @@ > > static DEFINE_MUTEX(iommu_sva_lock); > > > > /* Allocate a PASID for the mm within range (inclusive) */ -static > > int iommu_sva_alloc_pasid(struct mm_struct *mm, struct device *dev) > > +static struct iommu_mm_data *iommu_alloc_mm_data(struct mm_struct > > +*mm, struct device *dev) > > { > > + struct iommu_mm_data *iommu_mm; > > ioasid_t pasid; > > - int ret = 0; > > + > > + lockdep_assert_held(&iommu_sva_lock); > > > > if (!arch_pgtable_dma_compat(mm)) > > - return -EBUSY; > > + return ERR_PTR(-EBUSY); > > > > - mutex_lock(&iommu_sva_lock); > > + iommu_mm = mm->iommu_mm; > > /* Is a PASID already associated with this mm? */ > > - if (mm_valid_pasid(mm)) { > > - if (mm->pasid >= dev->iommu->max_pasids) > > - ret = -EOVERFLOW; > > - goto out; > > + if (iommu_mm) { > > + if (iommu_mm->pasid >= dev->iommu->max_pasids) > > + return ERR_PTR(-EOVERFLOW); > > + return iommu_mm; > > } > > > > + iommu_mm = kzalloc(sizeof(struct iommu_mm_data), GFP_KERNEL); > > + if (!iommu_mm) > > + return ERR_PTR(-ENOMEM); > > + > > pasid = iommu_alloc_global_pasid(dev); > > if (pasid == IOMMU_PASID_INVALID) { > > - ret = -ENOSPC; > > - goto out; > > + kfree(iommu_mm); > > + return ERR_PTR(-ENOSPC); > > } > > - mm->pasid = pasid; > > - ret = 0; > > -out: > > - mutex_unlock(&iommu_sva_lock); > > - return ret; > > + iommu_mm->pasid = pasid; > > + INIT_LIST_HEAD(&iommu_mm->sva_domains); > > + /* > > + * Make sure the write to mm->iommu_mm is not reordered in front > of > > + * initialization to iommu_mm fields. If it does, readers may see a > > + * valid iommu_mm with uninitialized values. > > + */ > > + smp_store_release(&mm->iommu_mm, iommu_mm); > > + return iommu_mm; > > } > > > > /** > > @@ -58,31 +68,33 @@ static int iommu_sva_alloc_pasid(struct mm_struct > *mm, struct device *dev) > > */ > > struct iommu_sva *iommu_sva_bind_device(struct device *dev, struct > > mm_struct *mm) { > > + struct iommu_mm_data *iommu_mm; > > struct iommu_domain *domain; > > struct iommu_sva *handle; > > int ret; > > > > + mutex_lock(&iommu_sva_lock); > > + > > /* Allocate mm->pasid if necessary. */ > > - ret = iommu_sva_alloc_pasid(mm, dev); > > - if (ret) > > - return ERR_PTR(ret); > > + iommu_mm = iommu_alloc_mm_data(mm, dev); > > + if (IS_ERR(iommu_mm)) { > > + ret = PTR_ERR(iommu_mm); > > + goto out_unlock; > > + } > > > > handle = kzalloc(sizeof(*handle), GFP_KERNEL); > > - if (!handle) > > - return ERR_PTR(-ENOMEM); > > - > > - mutex_lock(&iommu_sva_lock); > > - /* Search for an existing domain. */ > > - domain = iommu_get_domain_for_dev_pasid(dev, mm->pasid, > > - IOMMU_DOMAIN_SVA); > > - if (IS_ERR(domain)) { > > - ret = PTR_ERR(domain); > > + if (!handle) { > > + ret = -ENOMEM; > > goto out_unlock; > > } > > > > - if (domain) { > > - domain->users++; > > - goto out; > > Our multi bind test case broke since 6.8-rc1. > The test case can use same domain & pasid, return different handle, > 6.7 simply domain->users ++ and return. > > > + /* Search for an existing domain. */ > > + list_for_each_entry(domain, &mm->iommu_mm->sva_domains, next) > { > > + ret = iommu_attach_device_pasid(domain, dev, > > + iommu_mm->pasid); > > Now iommu_attach_device_pasid return BUSY since the same pasid. > And then iommu_sva_bind_device attach ret=-16 Sounds like the test case tries to bind a device to a same mm multiple times without unbinding the device and the expectation is that it can always return a valid handle to pass the test. Right? Regards, -Tina > > > + if (!ret) { > > Simply tried if (!ret || ret == -EBUSY) > The test passes, but report waring > WARNING: CPU: 12 PID: 2992 at drivers/iommu/iommu.c:3591 > iommu_detach_device_pasid+0xa4/0xd0 > > Will check more tomorrow. > > > + domain->users++; > > + goto out; > > + } > > } > > > > Thanks
Hi, Tina On Wed, 21 Feb 2024 at 07:58, Zhang, Tina <tina.zhang@intel.com> wrote: > > > struct iommu_sva *iommu_sva_bind_device(struct device *dev, struct > > > mm_struct *mm) { > > > + struct iommu_mm_data *iommu_mm; > > > struct iommu_domain *domain; > > > struct iommu_sva *handle; > > > int ret; > > > > > > + mutex_lock(&iommu_sva_lock); > > > + > > > /* Allocate mm->pasid if necessary. */ > > > - ret = iommu_sva_alloc_pasid(mm, dev); > > > - if (ret) > > > - return ERR_PTR(ret); > > > + iommu_mm = iommu_alloc_mm_data(mm, dev); > > > + if (IS_ERR(iommu_mm)) { > > > + ret = PTR_ERR(iommu_mm); > > > + goto out_unlock; > > > + } > > > > > > handle = kzalloc(sizeof(*handle), GFP_KERNEL); > > > - if (!handle) > > > - return ERR_PTR(-ENOMEM); > > > - > > > - mutex_lock(&iommu_sva_lock); > > > - /* Search for an existing domain. */ > > > - domain = iommu_get_domain_for_dev_pasid(dev, mm->pasid, > > > - IOMMU_DOMAIN_SVA); > > > - if (IS_ERR(domain)) { > > > - ret = PTR_ERR(domain); > > > + if (!handle) { > > > + ret = -ENOMEM; > > > goto out_unlock; > > > } > > > > > > - if (domain) { > > > - domain->users++; > > > - goto out; > > > > Our multi bind test case broke since 6.8-rc1. > > The test case can use same domain & pasid, return different handle, > > 6.7 simply domain->users ++ and return. > > > > > + /* Search for an existing domain. */ > > > + list_for_each_entry(domain, &mm->iommu_mm->sva_domains, next) > > { > > > + ret = iommu_attach_device_pasid(domain, dev, > > > + iommu_mm->pasid); > > > > Now iommu_attach_device_pasid return BUSY since the same pasid. > > And then iommu_sva_bind_device attach ret=-16 > Sounds like the test case tries to bind a device to a same mm multiple times without unbinding the device and the expectation is that it can always return a valid handle to pass the test. Right? Yes The device can bind to the same mm multi-times and return different handle, Since the refcount, no need to unbind and bind sequently, The unbind can happen later with the handle. Thanks
On Wed, 21 Feb 2024 at 09:28, Zhangfei Gao <zhangfei.gao@linaro.org> wrote: > > Hi, Tina > > On Wed, 21 Feb 2024 at 07:58, Zhang, Tina <tina.zhang@intel.com> wrote: > > > > > struct iommu_sva *iommu_sva_bind_device(struct device *dev, struct > > > > mm_struct *mm) { > > > > + struct iommu_mm_data *iommu_mm; > > > > struct iommu_domain *domain; > > > > struct iommu_sva *handle; > > > > int ret; > > > > > > > > + mutex_lock(&iommu_sva_lock); > > > > + > > > > /* Allocate mm->pasid if necessary. */ > > > > - ret = iommu_sva_alloc_pasid(mm, dev); > > > > - if (ret) > > > > - return ERR_PTR(ret); > > > > + iommu_mm = iommu_alloc_mm_data(mm, dev); > > > > + if (IS_ERR(iommu_mm)) { > > > > + ret = PTR_ERR(iommu_mm); > > > > + goto out_unlock; > > > > + } > > > > > > > > handle = kzalloc(sizeof(*handle), GFP_KERNEL); > > > > - if (!handle) > > > > - return ERR_PTR(-ENOMEM); > > > > - > > > > - mutex_lock(&iommu_sva_lock); > > > > - /* Search for an existing domain. */ > > > > - domain = iommu_get_domain_for_dev_pasid(dev, mm->pasid, > > > > - IOMMU_DOMAIN_SVA); > > > > - if (IS_ERR(domain)) { > > > > - ret = PTR_ERR(domain); > > > > + if (!handle) { > > > > + ret = -ENOMEM; > > > > goto out_unlock; > > > > } > > > > > > > > - if (domain) { > > > > - domain->users++; > > > > - goto out; > > > > > > Our multi bind test case broke since 6.8-rc1. > > > The test case can use same domain & pasid, return different handle, > > > 6.7 simply domain->users ++ and return. > > > > > > > + /* Search for an existing domain. */ > > > > + list_for_each_entry(domain, &mm->iommu_mm->sva_domains, next) > > > { > > > > + ret = iommu_attach_device_pasid(domain, dev, > > > > + iommu_mm->pasid); > > > > > > Now iommu_attach_device_pasid return BUSY since the same pasid. > > > And then iommu_sva_bind_device attach ret=-16 > > Sounds like the test case tries to bind a device to a same mm multiple times without unbinding the device and the expectation is that it can always return a valid handle to pass the test. Right? > > Yes > > The device can bind to the same mm multi-times and return different handle, > Since the refcount, no need to unbind and bind sequently, > The unbind can happen later with the handle. With this diff can solve the issue, what's your suggestion? @@ -88,10 +94,12 @@ struct iommu_sva *iommu_sva_bind_device(struct device *dev, struct mm_struct *mm /* Search for an existing domain. */ list_for_each_entry(domain, &mm->iommu_mm->sva_domains, next) { ret = iommu_attach_device_pasid(domain, dev, iommu_mm->pasid); - if (!ret) { + if (!ret || ret == -EBUSY) { domain->users++; goto out; } @@ -141,8 +151,8 @@ void iommu_sva_unbind_device(struct iommu_sva *handle) struct device *dev = handle->dev; mutex_lock(&iommu_sva_lock); - iommu_detach_device_pasid(domain, dev, iommu_mm->pasid); if (--domain->users == 0) { + iommu_detach_device_pasid(domain, dev, iommu_mm->pasid); list_del(&domain->next); iommu_domain_free(domain); } Thanks
On 2024/2/21 9:28, Zhangfei Gao wrote: > On Wed, 21 Feb 2024 at 07:58, Zhang, Tina<tina.zhang@intel.com> wrote: > >>>> struct iommu_sva *iommu_sva_bind_device(struct device *dev, struct >>>> mm_struct *mm) { >>>> + struct iommu_mm_data *iommu_mm; >>>> struct iommu_domain *domain; >>>> struct iommu_sva *handle; >>>> int ret; >>>> >>>> + mutex_lock(&iommu_sva_lock); >>>> + >>>> /* Allocate mm->pasid if necessary. */ >>>> - ret = iommu_sva_alloc_pasid(mm, dev); >>>> - if (ret) >>>> - return ERR_PTR(ret); >>>> + iommu_mm = iommu_alloc_mm_data(mm, dev); >>>> + if (IS_ERR(iommu_mm)) { >>>> + ret = PTR_ERR(iommu_mm); >>>> + goto out_unlock; >>>> + } >>>> >>>> handle = kzalloc(sizeof(*handle), GFP_KERNEL); >>>> - if (!handle) >>>> - return ERR_PTR(-ENOMEM); >>>> - >>>> - mutex_lock(&iommu_sva_lock); >>>> - /* Search for an existing domain. */ >>>> - domain = iommu_get_domain_for_dev_pasid(dev, mm->pasid, >>>> - IOMMU_DOMAIN_SVA); >>>> - if (IS_ERR(domain)) { >>>> - ret = PTR_ERR(domain); >>>> + if (!handle) { >>>> + ret = -ENOMEM; >>>> goto out_unlock; >>>> } >>>> >>>> - if (domain) { >>>> - domain->users++; >>>> - goto out; >>> Our multi bind test case broke since 6.8-rc1. >>> The test case can use same domain & pasid, return different handle, >>> 6.7 simply domain->users ++ and return. >>> >>>> + /* Search for an existing domain. */ >>>> + list_for_each_entry(domain, &mm->iommu_mm->sva_domains, next) >>> { >>>> + ret = iommu_attach_device_pasid(domain, dev, >>>> + iommu_mm->pasid); >>> Now iommu_attach_device_pasid return BUSY since the same pasid. >>> And then iommu_sva_bind_device attach ret=-16 >> Sounds like the test case tries to bind a device to a same mm multiple times without unbinding the device and the expectation is that it can always return a valid handle to pass the test. Right? > Yes > > The device can bind to the same mm multi-times and return different handle, > Since the refcount, no need to unbind and bind sequently, > The unbind can happen later with the handle. Is there any real use case to bind an mm to the pasid of a device multiple times? If there are cases, is it better to handle this in the uacce driver? From iommu core's perspective, it doesn't make sense to attach the same domain to the same device (or pasid) multiple times. Best regards, baolu
On Wed, 21 Feb 2024 at 10:06, Baolu Lu <baolu.lu@linux.intel.com> wrote: > > On 2024/2/21 9:28, Zhangfei Gao wrote: > > On Wed, 21 Feb 2024 at 07:58, Zhang, Tina<tina.zhang@intel.com> wrote: > > > >>>> struct iommu_sva *iommu_sva_bind_device(struct device *dev, struct > >>>> mm_struct *mm) { > >>>> + struct iommu_mm_data *iommu_mm; > >>>> struct iommu_domain *domain; > >>>> struct iommu_sva *handle; > >>>> int ret; > >>>> > >>>> + mutex_lock(&iommu_sva_lock); > >>>> + > >>>> /* Allocate mm->pasid if necessary. */ > >>>> - ret = iommu_sva_alloc_pasid(mm, dev); > >>>> - if (ret) > >>>> - return ERR_PTR(ret); > >>>> + iommu_mm = iommu_alloc_mm_data(mm, dev); > >>>> + if (IS_ERR(iommu_mm)) { > >>>> + ret = PTR_ERR(iommu_mm); > >>>> + goto out_unlock; > >>>> + } > >>>> > >>>> handle = kzalloc(sizeof(*handle), GFP_KERNEL); > >>>> - if (!handle) > >>>> - return ERR_PTR(-ENOMEM); > >>>> - > >>>> - mutex_lock(&iommu_sva_lock); > >>>> - /* Search for an existing domain. */ > >>>> - domain = iommu_get_domain_for_dev_pasid(dev, mm->pasid, > >>>> - IOMMU_DOMAIN_SVA); > >>>> - if (IS_ERR(domain)) { > >>>> - ret = PTR_ERR(domain); > >>>> + if (!handle) { > >>>> + ret = -ENOMEM; > >>>> goto out_unlock; > >>>> } > >>>> > >>>> - if (domain) { > >>>> - domain->users++; > >>>> - goto out; > >>> Our multi bind test case broke since 6.8-rc1. > >>> The test case can use same domain & pasid, return different handle, > >>> 6.7 simply domain->users ++ and return. > >>> > >>>> + /* Search for an existing domain. */ > >>>> + list_for_each_entry(domain, &mm->iommu_mm->sva_domains, next) > >>> { > >>>> + ret = iommu_attach_device_pasid(domain, dev, > >>>> + iommu_mm->pasid); > >>> Now iommu_attach_device_pasid return BUSY since the same pasid. > >>> And then iommu_sva_bind_device attach ret=-16 > >> Sounds like the test case tries to bind a device to a same mm multiple times without unbinding the device and the expectation is that it can always return a valid handle to pass the test. Right? > > Yes > > > > The device can bind to the same mm multi-times and return different handle, > > Since the refcount, no need to unbind and bind sequently, > > The unbind can happen later with the handle. > > Is there any real use case to bind an mm to the pasid of a device > multiple times? If there are cases, is it better to handle this in the > uacce driver? Yes, it is required for multi-thread, the device can provide multi-queue to speed up. > > From iommu core's perspective, it doesn't make sense to attach the same > domain to the same device (or pasid) multiple times. But is it the refcount domain->user++ used for? Is there any reason not doing this. Thanks
On 2024/2/21 10:45, Zhangfei Gao wrote: > On Wed, 21 Feb 2024 at 10:06, Baolu Lu<baolu.lu@linux.intel.com> > wrote: >> On 2024/2/21 9:28, Zhangfei Gao wrote: >>> On Wed, 21 Feb 2024 at 07:58, Zhang, Tina<tina.zhang@intel.com> >>> wrote: >>> >>>>>> struct iommu_sva *iommu_sva_bind_device(struct device >>>>>> *dev, struct mm_struct *mm) { + struct >>>>>> iommu_mm_data *iommu_mm; struct iommu_domain *domain; >>>>>> struct iommu_sva *handle; int ret; >>>>>> >>>>>> + mutex_lock(&iommu_sva_lock); + /* Allocate >>>>>> mm->pasid if necessary. */ - ret = >>>>>> iommu_sva_alloc_pasid(mm, dev); - if (ret) - return >>>>>> ERR_PTR(ret); + iommu_mm = iommu_alloc_mm_data(mm, >>>>>> dev); + if (IS_ERR(iommu_mm)) { + ret = >>>>>> PTR_ERR(iommu_mm); + goto out_unlock; + } >>>>>> >>>>>> handle = kzalloc(sizeof(*handle), GFP_KERNEL); - if >>>>>> (!handle) - return ERR_PTR(-ENOMEM); - - >>>>>> mutex_lock(&iommu_sva_lock); - /* Search for an >>>>>> existing domain. */ - domain = >>>>>> iommu_get_domain_for_dev_pasid(dev, mm->pasid, - >>>>>> IOMMU_DOMAIN_SVA); - if (IS_ERR(domain)) { - ret = >>>>>> PTR_ERR(domain); + if (!handle) { + ret = -ENOMEM; >>>>>> goto out_unlock; } >>>>>> >>>>>> - if (domain) { - domain->users++; - >>>>>> goto out; >>>>> Our multi bind test case broke since 6.8-rc1. The test case >>>>> can use same domain & pasid, return different handle, 6.7 >>>>> simply domain->users ++ and return. >>>>> >>>>>> + /* Search for an existing domain. */ + >>>>>> list_for_each_entry(domain, &mm->iommu_mm->sva_domains, >>>>>> next) >>>>> { >>>>>> + ret = iommu_attach_device_pasid(domain, >>>>>> dev, + iommu_mm->pasid); >>>>> Now iommu_attach_device_pasid return BUSY since the same >>>>> pasid. And then iommu_sva_bind_device attach ret=-16 >>>> Sounds like the test case tries to bind a device to a same mm >>>> multiple times without unbinding the device and the >>>> expectation is that it can always return a valid handle to pass >>>> the test. Right? >>> Yes >>> >>> The device can bind to the same mm multi-times and return >>> different handle, Since the refcount, no need to unbind and bind >>> sequently, The unbind can happen later with the handle. >> Is there any real use case to bind an mm to the pasid of a device >> multiple times? If there are cases, is it better to handle this in >> the uacce driver? > Yes, it is required for multi-thread, the device can provide > multi-queue to speed up. > >> From iommu core's perspective, it doesn't make sense to attach the >> same domain to the same device (or pasid) multiple times. > But is it the refcount domain->user++ used for? Is there any reason > not doing this. I was just thinking about whether to do this in the iommu core, or in the upper layers, like uacce or iommufd. It seems that there is no need to attach a domain to a device or pasid again if it has already been attached. Best regards, baolu
On Wed, 21 Feb 2024 at 11:52, Baolu Lu <baolu.lu@linux.intel.com> wrote: > > On 2024/2/21 10:45, Zhangfei Gao wrote: > > On Wed, 21 Feb 2024 at 10:06, Baolu Lu<baolu.lu@linux.intel.com> > > wrote: > >> On 2024/2/21 9:28, Zhangfei Gao wrote: > >>> On Wed, 21 Feb 2024 at 07:58, Zhang, Tina<tina.zhang@intel.com> > >>> wrote: > >>> > >>>>>> struct iommu_sva *iommu_sva_bind_device(struct device > >>>>>> *dev, struct mm_struct *mm) { + struct > >>>>>> iommu_mm_data *iommu_mm; struct iommu_domain *domain; > >>>>>> struct iommu_sva *handle; int ret; > >>>>>> > >>>>>> + mutex_lock(&iommu_sva_lock); + /* Allocate > >>>>>> mm->pasid if necessary. */ - ret = > >>>>>> iommu_sva_alloc_pasid(mm, dev); - if (ret) - return > >>>>>> ERR_PTR(ret); + iommu_mm = iommu_alloc_mm_data(mm, > >>>>>> dev); + if (IS_ERR(iommu_mm)) { + ret = > >>>>>> PTR_ERR(iommu_mm); + goto out_unlock; + } > >>>>>> > >>>>>> handle = kzalloc(sizeof(*handle), GFP_KERNEL); - if > >>>>>> (!handle) - return ERR_PTR(-ENOMEM); - - > >>>>>> mutex_lock(&iommu_sva_lock); - /* Search for an > >>>>>> existing domain. */ - domain = > >>>>>> iommu_get_domain_for_dev_pasid(dev, mm->pasid, - > >>>>>> IOMMU_DOMAIN_SVA); - if (IS_ERR(domain)) { - ret = > >>>>>> PTR_ERR(domain); + if (!handle) { + ret = -ENOMEM; > >>>>>> goto out_unlock; } > >>>>>> > >>>>>> - if (domain) { - domain->users++; - > >>>>>> goto out; > >>>>> Our multi bind test case broke since 6.8-rc1. The test case > >>>>> can use same domain & pasid, return different handle, 6.7 > >>>>> simply domain->users ++ and return. > >>>>> > >>>>>> + /* Search for an existing domain. */ + > >>>>>> list_for_each_entry(domain, &mm->iommu_mm->sva_domains, > >>>>>> next) > >>>>> { > >>>>>> + ret = iommu_attach_device_pasid(domain, > >>>>>> dev, + iommu_mm->pasid); > >>>>> Now iommu_attach_device_pasid return BUSY since the same > >>>>> pasid. And then iommu_sva_bind_device attach ret=-16 > >>>> Sounds like the test case tries to bind a device to a same mm > >>>> multiple times without unbinding the device and the > >>>> expectation is that it can always return a valid handle to pass > >>>> the test. Right? > >>> Yes > >>> > >>> The device can bind to the same mm multi-times and return > >>> different handle, Since the refcount, no need to unbind and bind > >>> sequently, The unbind can happen later with the handle. > >> Is there any real use case to bind an mm to the pasid of a device > >> multiple times? If there are cases, is it better to handle this in > >> the uacce driver? > > Yes, it is required for multi-thread, the device can provide > > multi-queue to speed up. > > > >> From iommu core's perspective, it doesn't make sense to attach the > >> same domain to the same device (or pasid) multiple times. > > But is it the refcount domain->user++ used for? Is there any reason > > not doing this. > > I was just thinking about whether to do this in the iommu core, or in > the upper layers, like uacce or iommufd. It seems that there is no need > to attach a domain to a device or pasid again if it has already been > attached. It would be more complicated since the return handle can be used to distinguish different queues of the device. I think domain->user should handle this case as before. Anyway, I have sent a patch to get more feedback. Thanks
Hi, > -----Original Message----- > From: Zhangfei Gao <zhangfei.gao@linaro.org> > Sent: Wednesday, February 21, 2024 10:45 AM > To: Baolu Lu <baolu.lu@linux.intel.com> > Cc: Zhang, Tina <tina.zhang@intel.com>; iommu@lists.linux.dev; linux- > kernel@vger.kernel.org; David Woodhouse <dwmw2@infradead.org>; Joerg > Roedel <joro@8bytes.org>; Will Deacon <will@kernel.org>; Robin Murphy > <robin.murphy@arm.com>; Jason Gunthorpe <jgg@ziepe.ca>; Tian, Kevin > <kevin.tian@intel.com>; Nicolin Chen <nicolinc@nvidia.com>; Michael Shavit > <mshavit@google.com>; Vasant Hegde <vasant.hegde@amd.com>; Jason > Gunthorpe <jgg@nvidia.com>; Jean-Philippe Brucker <jean- > philippe@linaro.org>; Hao Fang <fanghao11@huawei.com> > Subject: Re: [PATCH v10 5/6] iommu: Support mm PASID 1:n with sva > domains > > On Wed, 21 Feb 2024 at 10:06, Baolu Lu <baolu.lu@linux.intel.com> wrote: > > > > On 2024/2/21 9:28, Zhangfei Gao wrote: > > > On Wed, 21 Feb 2024 at 07:58, Zhang, Tina<tina.zhang@intel.com> > wrote: > > > > > >>>> struct iommu_sva *iommu_sva_bind_device(struct device *dev, > > >>>> struct mm_struct *mm) { > > >>>> + struct iommu_mm_data *iommu_mm; > > >>>> struct iommu_domain *domain; > > >>>> struct iommu_sva *handle; > > >>>> int ret; > > >>>> > > >>>> + mutex_lock(&iommu_sva_lock); > > >>>> + > > >>>> /* Allocate mm->pasid if necessary. */ > > >>>> - ret = iommu_sva_alloc_pasid(mm, dev); > > >>>> - if (ret) > > >>>> - return ERR_PTR(ret); > > >>>> + iommu_mm = iommu_alloc_mm_data(mm, dev); > > >>>> + if (IS_ERR(iommu_mm)) { > > >>>> + ret = PTR_ERR(iommu_mm); > > >>>> + goto out_unlock; > > >>>> + } > > >>>> > > >>>> handle = kzalloc(sizeof(*handle), GFP_KERNEL); > > >>>> - if (!handle) > > >>>> - return ERR_PTR(-ENOMEM); > > >>>> - > > >>>> - mutex_lock(&iommu_sva_lock); > > >>>> - /* Search for an existing domain. */ > > >>>> - domain = iommu_get_domain_for_dev_pasid(dev, mm->pasid, > > >>>> - IOMMU_DOMAIN_SVA); > > >>>> - if (IS_ERR(domain)) { > > >>>> - ret = PTR_ERR(domain); > > >>>> + if (!handle) { > > >>>> + ret = -ENOMEM; > > >>>> goto out_unlock; > > >>>> } > > >>>> > > >>>> - if (domain) { > > >>>> - domain->users++; > > >>>> - goto out; > > >>> Our multi bind test case broke since 6.8-rc1. > > >>> The test case can use same domain & pasid, return different > > >>> handle, > > >>> 6.7 simply domain->users ++ and return. > > >>> > > >>>> + /* Search for an existing domain. */ > > >>>> + list_for_each_entry(domain, &mm->iommu_mm->sva_domains, > > >>>> + next) > > >>> { > > >>>> + ret = iommu_attach_device_pasid(domain, dev, > > >>>> + iommu_mm->pasid); > > >>> Now iommu_attach_device_pasid return BUSY since the same pasid. > > >>> And then iommu_sva_bind_device attach ret=-16 > > >> Sounds like the test case tries to bind a device to a same mm multiple > times without unbinding the device and the expectation is that it can always > return a valid handle to pass the test. Right? > > > Yes > > > > > > The device can bind to the same mm multi-times and return different > > > handle, Since the refcount, no need to unbind and bind sequently, > > > The unbind can happen later with the handle. > > > > Is there any real use case to bind an mm to the pasid of a device > > multiple times? If there are cases, is it better to handle this in the > > uacce driver? > > Yes, it is required for multi-thread, the device can provide multi-queue to > speed up. > > > > > From iommu core's perspective, it doesn't make sense to attach the > > same domain to the same device (or pasid) multiple times. > > But is it the refcount domain->user++ used for? > Is there any reason not doing this. The domain->user is a refcount of the devices (or iommu group) attached to the domain. IOMMU core needs to keep this refcount to ensure that a sva domain will be released when no device uses it. Regards, -Tina > > Thanks
On Wed, 21 Feb 2024 at 15:41, Zhang, Tina <tina.zhang@intel.com> wrote: > > Hi, > > > -----Original Message----- > > From: Zhangfei Gao <zhangfei.gao@linaro.org> > > Sent: Wednesday, February 21, 2024 10:45 AM > > To: Baolu Lu <baolu.lu@linux.intel.com> > > Cc: Zhang, Tina <tina.zhang@intel.com>; iommu@lists.linux.dev; linux- > > kernel@vger.kernel.org; David Woodhouse <dwmw2@infradead.org>; Joerg > > Roedel <joro@8bytes.org>; Will Deacon <will@kernel.org>; Robin Murphy > > <robin.murphy@arm.com>; Jason Gunthorpe <jgg@ziepe.ca>; Tian, Kevin > > <kevin.tian@intel.com>; Nicolin Chen <nicolinc@nvidia.com>; Michael Shavit > > <mshavit@google.com>; Vasant Hegde <vasant.hegde@amd.com>; Jason > > Gunthorpe <jgg@nvidia.com>; Jean-Philippe Brucker <jean- > > philippe@linaro.org>; Hao Fang <fanghao11@huawei.com> > > Subject: Re: [PATCH v10 5/6] iommu: Support mm PASID 1:n with sva > > domains > > > > On Wed, 21 Feb 2024 at 10:06, Baolu Lu <baolu.lu@linux.intel.com> wrote: > > > > > > On 2024/2/21 9:28, Zhangfei Gao wrote: > > > > On Wed, 21 Feb 2024 at 07:58, Zhang, Tina<tina.zhang@intel.com> > > wrote: > > > > > > > >>>> struct iommu_sva *iommu_sva_bind_device(struct device *dev, > > > >>>> struct mm_struct *mm) { > > > >>>> + struct iommu_mm_data *iommu_mm; > > > >>>> struct iommu_domain *domain; > > > >>>> struct iommu_sva *handle; > > > >>>> int ret; > > > >>>> > > > >>>> + mutex_lock(&iommu_sva_lock); > > > >>>> + > > > >>>> /* Allocate mm->pasid if necessary. */ > > > >>>> - ret = iommu_sva_alloc_pasid(mm, dev); > > > >>>> - if (ret) > > > >>>> - return ERR_PTR(ret); > > > >>>> + iommu_mm = iommu_alloc_mm_data(mm, dev); > > > >>>> + if (IS_ERR(iommu_mm)) { > > > >>>> + ret = PTR_ERR(iommu_mm); > > > >>>> + goto out_unlock; > > > >>>> + } > > > >>>> > > > >>>> handle = kzalloc(sizeof(*handle), GFP_KERNEL); > > > >>>> - if (!handle) > > > >>>> - return ERR_PTR(-ENOMEM); > > > >>>> - > > > >>>> - mutex_lock(&iommu_sva_lock); > > > >>>> - /* Search for an existing domain. */ > > > >>>> - domain = iommu_get_domain_for_dev_pasid(dev, mm->pasid, > > > >>>> - IOMMU_DOMAIN_SVA); > > > >>>> - if (IS_ERR(domain)) { > > > >>>> - ret = PTR_ERR(domain); > > > >>>> + if (!handle) { > > > >>>> + ret = -ENOMEM; > > > >>>> goto out_unlock; > > > >>>> } > > > >>>> > > > >>>> - if (domain) { > > > >>>> - domain->users++; > > > >>>> - goto out; > > > >>> Our multi bind test case broke since 6.8-rc1. > > > >>> The test case can use same domain & pasid, return different > > > >>> handle, > > > >>> 6.7 simply domain->users ++ and return. > > > >>> > > > >>>> + /* Search for an existing domain. */ > > > >>>> + list_for_each_entry(domain, &mm->iommu_mm->sva_domains, > > > >>>> + next) > > > >>> { > > > >>>> + ret = iommu_attach_device_pasid(domain, dev, > > > >>>> + iommu_mm->pasid); > > > >>> Now iommu_attach_device_pasid return BUSY since the same pasid. > > > >>> And then iommu_sva_bind_device attach ret=-16 > > > >> Sounds like the test case tries to bind a device to a same mm multiple > > times without unbinding the device and the expectation is that it can always > > return a valid handle to pass the test. Right? > > > > Yes > > > > > > > > The device can bind to the same mm multi-times and return different > > > > handle, Since the refcount, no need to unbind and bind sequently, > > > > The unbind can happen later with the handle. > > > > > > Is there any real use case to bind an mm to the pasid of a device > > > multiple times? If there are cases, is it better to handle this in the > > > uacce driver? > > > > Yes, it is required for multi-thread, the device can provide multi-queue to > > speed up. > > > > > > > > From iommu core's perspective, it doesn't make sense to attach the > > > same domain to the same device (or pasid) multiple times. > > > > But is it the refcount domain->user++ used for? > > Is there any reason not doing this. > The domain->user is a refcount of the devices (or iommu group) attached to the domain. IOMMU core needs to keep this refcount to ensure that a sva domain will be released when no device uses it. I think the limitation of one user only attach one domain one time does not make sense. Just like one file can only be opened one time by a user, then refcount is meanless. Thanks > > Regards, > -Tina > > > > > Thanks
diff --git a/drivers/iommu/iommu-sva.c b/drivers/iommu/iommu-sva.c index 4a2f5699747f..5175e8d85247 100644 --- a/drivers/iommu/iommu-sva.c +++ b/drivers/iommu/iommu-sva.c @@ -12,32 +12,42 @@ static DEFINE_MUTEX(iommu_sva_lock); /* Allocate a PASID for the mm within range (inclusive) */ -static int iommu_sva_alloc_pasid(struct mm_struct *mm, struct device *dev) +static struct iommu_mm_data *iommu_alloc_mm_data(struct mm_struct *mm, struct device *dev) { + struct iommu_mm_data *iommu_mm; ioasid_t pasid; - int ret = 0; + + lockdep_assert_held(&iommu_sva_lock); if (!arch_pgtable_dma_compat(mm)) - return -EBUSY; + return ERR_PTR(-EBUSY); - mutex_lock(&iommu_sva_lock); + iommu_mm = mm->iommu_mm; /* Is a PASID already associated with this mm? */ - if (mm_valid_pasid(mm)) { - if (mm->pasid >= dev->iommu->max_pasids) - ret = -EOVERFLOW; - goto out; + if (iommu_mm) { + if (iommu_mm->pasid >= dev->iommu->max_pasids) + return ERR_PTR(-EOVERFLOW); + return iommu_mm; } + iommu_mm = kzalloc(sizeof(struct iommu_mm_data), GFP_KERNEL); + if (!iommu_mm) + return ERR_PTR(-ENOMEM); + pasid = iommu_alloc_global_pasid(dev); if (pasid == IOMMU_PASID_INVALID) { - ret = -ENOSPC; - goto out; + kfree(iommu_mm); + return ERR_PTR(-ENOSPC); } - mm->pasid = pasid; - ret = 0; -out: - mutex_unlock(&iommu_sva_lock); - return ret; + iommu_mm->pasid = pasid; + INIT_LIST_HEAD(&iommu_mm->sva_domains); + /* + * Make sure the write to mm->iommu_mm is not reordered in front of + * initialization to iommu_mm fields. If it does, readers may see a + * valid iommu_mm with uninitialized values. + */ + smp_store_release(&mm->iommu_mm, iommu_mm); + return iommu_mm; } /** @@ -58,31 +68,33 @@ static int iommu_sva_alloc_pasid(struct mm_struct *mm, struct device *dev) */ struct iommu_sva *iommu_sva_bind_device(struct device *dev, struct mm_struct *mm) { + struct iommu_mm_data *iommu_mm; struct iommu_domain *domain; struct iommu_sva *handle; int ret; + mutex_lock(&iommu_sva_lock); + /* Allocate mm->pasid if necessary. */ - ret = iommu_sva_alloc_pasid(mm, dev); - if (ret) - return ERR_PTR(ret); + iommu_mm = iommu_alloc_mm_data(mm, dev); + if (IS_ERR(iommu_mm)) { + ret = PTR_ERR(iommu_mm); + goto out_unlock; + } handle = kzalloc(sizeof(*handle), GFP_KERNEL); - if (!handle) - return ERR_PTR(-ENOMEM); - - mutex_lock(&iommu_sva_lock); - /* Search for an existing domain. */ - domain = iommu_get_domain_for_dev_pasid(dev, mm->pasid, - IOMMU_DOMAIN_SVA); - if (IS_ERR(domain)) { - ret = PTR_ERR(domain); + if (!handle) { + ret = -ENOMEM; goto out_unlock; } - if (domain) { - domain->users++; - goto out; + /* Search for an existing domain. */ + list_for_each_entry(domain, &mm->iommu_mm->sva_domains, next) { + ret = iommu_attach_device_pasid(domain, dev, iommu_mm->pasid); + if (!ret) { + domain->users++; + goto out; + } } /* Allocate a new domain and set it on device pasid. */ @@ -92,23 +104,23 @@ struct iommu_sva *iommu_sva_bind_device(struct device *dev, struct mm_struct *mm goto out_unlock; } - ret = iommu_attach_device_pasid(domain, dev, mm->pasid); + ret = iommu_attach_device_pasid(domain, dev, iommu_mm->pasid); if (ret) goto out_free_domain; domain->users = 1; + list_add(&domain->next, &mm->iommu_mm->sva_domains); + out: mutex_unlock(&iommu_sva_lock); handle->dev = dev; handle->domain = domain; - return handle; out_free_domain: iommu_domain_free(domain); + kfree(handle); out_unlock: mutex_unlock(&iommu_sva_lock); - kfree(handle); - return ERR_PTR(ret); } EXPORT_SYMBOL_GPL(iommu_sva_bind_device); @@ -124,12 +136,13 @@ EXPORT_SYMBOL_GPL(iommu_sva_bind_device); void iommu_sva_unbind_device(struct iommu_sva *handle) { struct iommu_domain *domain = handle->domain; - ioasid_t pasid = domain->mm->pasid; + struct iommu_mm_data *iommu_mm = domain->mm->iommu_mm; struct device *dev = handle->dev; mutex_lock(&iommu_sva_lock); + iommu_detach_device_pasid(domain, dev, iommu_mm->pasid); if (--domain->users == 0) { - iommu_detach_device_pasid(domain, dev, pasid); + list_del(&domain->next); iommu_domain_free(domain); } mutex_unlock(&iommu_sva_lock); @@ -205,8 +218,11 @@ iommu_sva_handle_iopf(struct iommu_fault *fault, void *data) void mm_pasid_drop(struct mm_struct *mm) { - if (likely(!mm_valid_pasid(mm))) + struct iommu_mm_data *iommu_mm = mm->iommu_mm; + + if (!iommu_mm) return; - iommu_free_global_pasid(mm->pasid); + iommu_free_global_pasid(iommu_mm->pasid); + kfree(iommu_mm); } diff --git a/include/linux/iommu.h b/include/linux/iommu.h index a807182c3d2e..98b199603588 100644 --- a/include/linux/iommu.h +++ b/include/linux/iommu.h @@ -113,6 +113,11 @@ struct iommu_domain { struct { /* IOMMU_DOMAIN_SVA */ struct mm_struct *mm; int users; + /* + * Next iommu_domain in mm->iommu_mm->sva-domains list + * protected by iommu_sva_lock. + */ + struct list_head next; }; }; }; @@ -1197,16 +1202,28 @@ static inline bool tegra_dev_iommu_get_stream_id(struct device *dev, u32 *stream #ifdef CONFIG_IOMMU_MM_DATA static inline void mm_pasid_init(struct mm_struct *mm) { - mm->pasid = IOMMU_PASID_INVALID; + /* + * During dup_mm(), a new mm will be memcpy'd from an old one and that makes + * the new mm and the old one point to a same iommu_mm instance. When either + * one of the two mms gets released, the iommu_mm instance is freed, leaving + * the other mm running into a use-after-free/double-free problem. To avoid + * the problem, zeroing the iommu_mm pointer of a new mm is needed here. + */ + mm->iommu_mm = NULL; } + static inline bool mm_valid_pasid(struct mm_struct *mm) { - return mm->pasid != IOMMU_PASID_INVALID; + return READ_ONCE(mm->iommu_mm); } static inline u32 mm_get_enqcmd_pasid(struct mm_struct *mm) { - return mm->pasid; + struct iommu_mm_data *iommu_mm = READ_ONCE(mm->iommu_mm); + + if (!iommu_mm) + return IOMMU_PASID_INVALID; + return iommu_mm->pasid; } void mm_pasid_drop(struct mm_struct *mm);