For supporting post MSI-X enable allocations and for the upcoming PCI/IMS
support a seperate interface is required which allows not only the
allocation of a specific index, but also the allocation of any, i.e. the
next free index. The latter is especially required for IMS because IMS
completely does away with index to functionality mappings which are
often found in MSI/MSI-X implementation.
But even with MSI-X there are devices where only the first few indices have
a fixed functionality and the rest is freely assignable by software,
e.g. to queues.
msi_domain_alloc_irq_at() is also different from the range based interfaces
as it always enforces that the MSI descriptor is allocated by the core code
and not preallocated by the caller like the PCI/MSI[-X] enable code path
does.
msi_domain_alloc_irq_at() can be invoked with the index argument set to
MSI_ANY_INDEX which makes the core code pick the next free index. The irq
domain can provide a prepare_desc() operation callback in its
msi_domain_ops to do domain specific post allocation initialization before
the actual Linux interrupt and the associated interrupt descriptor and
hierarchy alloccations are conducted.
The function also takes an optional @cookie argument which is of type union
msi_dev_cookie. This cookie is not used by the core code and is stored in
the allocated msi_desc::data::cookie. The meaning of the cookie is
completely implementation defined. In case of IMS this might be a PASID or
a pointer to a device queue, but for the MSI core it's opaque and not used
in any way.
The function returns a struct msi_map which on success contains the
allocated index number and the Linux interrupt number so the caller can
spare the index to Linux interrupt number lookup.
On failure map::index contains the error code and map::virq is 0.
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
---
V2: Fix the recursive allocation issue and the index calculation (Reinette)
---
include/linux/msi.h | 4 +
include/linux/msi_api.h | 7 ++
kernel/irq/msi.c | 113 +++++++++++++++++++++++++++++++++++++++++++-----
3 files changed, 113 insertions(+), 11 deletions(-)
@@ -80,6 +80,7 @@ struct pci_dev;
struct platform_msi_priv_data;
struct device_attribute;
struct irq_domain;
+struct irq_affinity_desc;
void __get_cached_msi_msg(struct msi_desc *entry, struct msi_msg *msg);
#ifdef CONFIG_GENERIC_MSI_IRQ
@@ -591,6 +592,9 @@ int msi_domain_alloc_irqs_range(struct d
unsigned int first, unsigned int last);
int msi_domain_alloc_irqs_all_locked(struct device *dev, unsigned int domid, int nirqs);
+struct msi_map msi_domain_alloc_irq_at(struct device *dev, unsigned int domid, unsigned int index,
+ const struct irq_affinity_desc *affdesc,
+ union msi_instance_cookie *cookie);
void msi_domain_free_irqs_range_locked(struct device *dev, unsigned int domid,
unsigned int first, unsigned int last);
@@ -48,6 +48,13 @@ struct msi_map {
int virq;
};
+/*
+ * Constant to be used for dynamic allocations when the allocation
+ * is any free MSI index (entry in the MSI-X table or a software
+ * managed index.
+ */
+#define MSI_ANY_INDEX UINT_MAX
+
unsigned int msi_domain_get_virq(struct device *dev, unsigned int domid, unsigned int index);
/**
@@ -40,6 +40,7 @@ static inline int msi_sysfs_create_group
/* Invalid XA index which is outside of any searchable range */
#define MSI_XA_MAX_INDEX (ULONG_MAX - 1)
#define MSI_XA_DOMAIN_SIZE (MSI_MAX_INDEX + 1)
+#define MSI_ANY_INDEX UINT_MAX
static inline void msi_setup_default_irqdomain(struct device *dev, struct msi_device_data *md)
{
@@ -127,18 +128,34 @@ static int msi_insert_desc(struct device
}
hwsize = msi_domain_get_hwsize(dev, domid);
- if (index >= hwsize) {
- ret = -ERANGE;
- goto fail;
- }
- desc->msi_index = index;
- index += baseidx;
- ret = xa_insert(&md->__store, index, desc, GFP_KERNEL);
- if (ret)
- goto fail;
- return 0;
+ if (index == MSI_ANY_INDEX) {
+ struct xa_limit limit;
+ unsigned int index;
+
+ limit.min = baseidx;
+ limit.max = baseidx + hwsize - 1;
+
+ /* Let the xarray allocate a free index within the limits */
+ ret = xa_alloc(&md->__store, &index, desc, limit, GFP_KERNEL);
+ if (ret)
+ goto fail;
+ desc->msi_index = index - baseidx;
+ return 0;
+ } else {
+ if (index >= hwsize) {
+ ret = -ERANGE;
+ goto fail;
+ }
+
+ desc->msi_index = index;
+ index += baseidx;
+ ret = xa_insert(&md->__store, index, desc, GFP_KERNEL);
+ if (ret)
+ goto fail;
+ return 0;
+ }
fail:
msi_free_desc(desc);
return ret;
@@ -333,7 +350,7 @@ int msi_setup_device_data(struct device
msi_setup_default_irqdomain(dev, md);
- xa_init(&md->__store);
+ xa_init_flags(&md->__store, XA_FLAGS_ALLOC);
mutex_init(&md->mutex);
md->__iter_idx = MSI_XA_MAX_INDEX;
dev->msi.data = md;
@@ -1433,6 +1450,80 @@ int msi_domain_alloc_irqs_all_locked(str
return msi_domain_alloc_locked(dev, &ctrl);
}
+/**
+ * msi_domain_alloc_irq_at - Allocate an interrupt from a MSI interrupt domain at
+ * a given index - or at the next free index
+ *
+ * @dev: Pointer to device struct of the device for which the interrupts
+ * are allocated
+ * @domid: Id of the interrupt domain to operate on
+ * @index: Index for allocation. If @index == %MSI_ANY_INDEX the allocation
+ * uses the next free index.
+ * @affdesc: Optional pointer to an interrupt affinity descriptor structure
+ * @icookie: Optional pointer to a domain specific per instance cookie. If
+ * non-NULL the content of the cookie is stored in msi_desc::data.
+ * Must be NULL for MSI-X allocations
+ *
+ * This requires a MSI interrupt domain which lets the core code manage the
+ * MSI descriptors.
+ *
+ * Return: struct msi_map
+ *
+ * On success msi_map::index contains the allocated index number and
+ * msi_map::virq the corresponding Linux interrupt number
+ *
+ * On failure msi_map::index contains the error code and msi_map::virq
+ * is %0.
+ */
+struct msi_map msi_domain_alloc_irq_at(struct device *dev, unsigned int domid, unsigned int index,
+ const struct irq_affinity_desc *affdesc,
+ union msi_instance_cookie *icookie)
+{
+ struct msi_ctrl ctrl = { .domid = domid, .nirqs = 1, };
+ struct msi_domain_info *info;
+ struct irq_domain *domain;
+ struct msi_map map = { };
+ struct msi_desc *desc;
+ int ret;
+
+ msi_lock_descs(dev);
+ domain = msi_get_device_domain(dev, domid);
+ if (!domain) {
+ map.index = -ENODEV;
+ goto unlock;
+ }
+
+ desc = msi_alloc_desc(dev, 1, affdesc);
+ if (!desc) {
+ map.index = -ENOMEM;
+ goto unlock;
+ }
+
+ if (icookie)
+ desc->data.icookie = *icookie;
+
+ ret = msi_insert_desc(dev, desc, domid, index);
+ if (ret) {
+ map.index = ret;
+ goto unlock;
+ }
+
+ ctrl.first = ctrl.last = desc->msi_index;
+ info = domain->host_data;
+
+ ret = __msi_domain_alloc_irqs(dev, domain, &ctrl);
+ if (ret) {
+ map.index = ret;
+ msi_domain_free_locked(dev, &ctrl);
+ } else {
+ map.index = desc->msi_index;
+ map.virq = desc->irq;
+ }
+unlock:
+ msi_unlock_descs(dev);
+ return map;
+}
+
static void __msi_domain_free_irqs(struct device *dev, struct irq_domain *domain,
struct msi_ctrl *ctrl)
{