Re: [PATCH v4 02/09] iommu/ipmmu-vmsa: Add optional root device feature

From: Robin Murphy
Date: Mon Jun 19 2017 - 13:19:29 EST


On 19/06/17 10:14, Magnus Damm wrote:
> From: Magnus Damm <damm+renesas@xxxxxxxxxxxxx>
>
> Add root device handling to the IPMMU driver by allowing certain
> DT compat strings to enable has_cache_leaf_nodes that in turn will
> support both root devices with interrupts and leaf devices that
> face the actual IPMMU consumer devices.
>
> Signed-off-by: Magnus Damm <damm+renesas@xxxxxxxxxxxxx>
> ---
>
> Changes since V3:
> - Reworked root finding code to make it easier to follow, thanks Geert!
>
> Changes since V2:
> - Fixed a bug in ipmmu_find_root() when only leaf devices are present
> - Broke out __ipmmu_find_root() to allow ->xlate() check for root devices
>
> Changes since V1:
> - Moved patch to earlier in the series
> - Updated code to work with recent changes in:
> [PATCH v3 00/06] iommu/ipmmu-vmsa: IPMMU multi-arch update V3
>
> drivers/iommu/ipmmu-vmsa.c | 95 ++++++++++++++++++++++++++++++++++++--------
> 1 file changed, 78 insertions(+), 17 deletions(-)
>
> --- 0015/drivers/iommu/ipmmu-vmsa.c
> +++ work/drivers/iommu/ipmmu-vmsa.c 2017-06-19 13:59:41.050607110 +0900
> @@ -36,6 +36,7 @@
>
> struct ipmmu_features {
> bool use_ns_alias_offset;
> + bool has_cache_leaf_nodes;
> };
>
> struct ipmmu_vmsa_device {
> @@ -44,6 +45,7 @@ struct ipmmu_vmsa_device {
> struct iommu_device iommu;
> struct list_head list;
> const struct ipmmu_features *features;
> + bool is_leaf;
> unsigned int num_utlbs;
> spinlock_t lock; /* Protects ctx and domains[] */
> DECLARE_BITMAP(ctx, IPMMU_CTX_MAX);
> @@ -54,6 +56,7 @@ struct ipmmu_vmsa_device {
>
> struct ipmmu_vmsa_domain {
> struct ipmmu_vmsa_device *mmu;
> + struct ipmmu_vmsa_device *root;

Would it not make more sense for this to be a property of the
ipmmu_device itself, rather than per ipmmu_domain? I may of course have
got the wrong idea of the topology here, but it seems as if mmu->is_leaf
could be expressed as mmu->root == mmu vs. mmu->root == some_other_mmu,
at which point there's one less thing to worry about in the domain.

> struct iommu_domain io_domain;
>
> struct io_pgtable_cfg cfg;
> @@ -203,6 +206,44 @@ static struct ipmmu_vmsa_iommu_priv *to_
> #define IMUASID_ASID0_SHIFT 0
>
> /* -----------------------------------------------------------------------------
> + * Root device handling
> + */
> +
> +static bool ipmmu_is_root(struct ipmmu_vmsa_device *mmu)
> +{
> + if (mmu->features->has_cache_leaf_nodes)
> + return mmu->is_leaf ? false : true> + else
> + return true; /* older IPMMU hardware treated as single root */
> +}
> +
> +static struct ipmmu_vmsa_device *__ipmmu_find_root(void)
> +{
> + struct ipmmu_vmsa_device *mmu;
> + struct ipmmu_vmsa_device *root = NULL;
> +
> + spin_lock(&ipmmu_devices_lock);
> +
> + list_for_each_entry(mmu, &ipmmu_devices, list) {
> + if (ipmmu_is_root(mmu)) {
> + root = mmu;
> + break;
> + }
> + }
> +
> + spin_unlock(&ipmmu_devices_lock);
> + return root;
> +}

I wonder if it might be tidier to use driver_for_each_device() for this,
and remove the local list at the same time as its previous user.

Either way, what happens if things end up hapening in this order:

1: probe leaf IPMMU B
2: probe device X behind IPMMU B
3: create and attach default domain for device X
4: probe root IPMMU A

We know X will defer if B isn't ready, but it doesn't seem (at a glance,
admittedly) that there's anything to enforce the expected probe ordering
between A and B. This seems like another argument for moving the
root/leaf association up to the device level, such that B can look up A
once in its own probe routine, and defer itself if necessary.

> +
> +static struct ipmmu_vmsa_device *ipmmu_find_root(struct ipmmu_vmsa_device *leaf)
> +{
> + if (ipmmu_is_root(leaf))
> + return leaf;
> + else
> + return __ipmmu_find_root();

Well actually, looking at that, I think what we have here is a very
long-winded way of implementing this:

static ipmmu_vmsa_device *root;

Yuck.

> +}
> +
> +/* -----------------------------------------------------------------------------
> * Read/Write Access
> */
>
> @@ -219,13 +260,13 @@ static void ipmmu_write(struct ipmmu_vms
>
> static u32 ipmmu_ctx_read(struct ipmmu_vmsa_domain *domain, unsigned int reg)
> {
> - return ipmmu_read(domain->mmu, domain->context_id * IM_CTX_SIZE + reg);
> + return ipmmu_read(domain->root, domain->context_id * IM_CTX_SIZE + reg);
> }
>
> static void ipmmu_ctx_write(struct ipmmu_vmsa_domain *domain, unsigned int reg,
> u32 data)
> {
> - ipmmu_write(domain->mmu, domain->context_id * IM_CTX_SIZE + reg, data);
> + ipmmu_write(domain->root, domain->context_id * IM_CTX_SIZE + reg, data);
> }
>
> /* -----------------------------------------------------------------------------
> @@ -360,7 +401,7 @@ static int ipmmu_domain_init_context(str
> * TODO: Add support for coherent walk through CCI with DVM and remove
> * cache handling. For now, delegate it to the io-pgtable code.
> */
> - domain->cfg.iommu_dev = domain->mmu->dev;
> + domain->cfg.iommu_dev = domain->root->dev;
>
> domain->iop = alloc_io_pgtable_ops(ARM_32_LPAE_S1, &domain->cfg,
> domain);
> @@ -370,7 +411,7 @@ static int ipmmu_domain_init_context(str
> /*
> * Find an unused context.
> */
> - ret = ipmmu_domain_allocate_context(domain->mmu, domain);
> + ret = ipmmu_domain_allocate_context(domain->root, domain);
> if (ret == IPMMU_CTX_MAX) {
> free_io_pgtable_ops(domain->iop);
> return -EBUSY;
> @@ -441,7 +482,7 @@ static void ipmmu_domain_destroy_context
> */
> ipmmu_ctx_write(domain, IMCTR, IMCTR_FLUSH);
> ipmmu_tlb_sync(domain);
> - ipmmu_domain_free_context(domain->mmu, domain->context_id);
> + ipmmu_domain_free_context(domain->root, domain->context_id);
> }
>
> /* -----------------------------------------------------------------------------
> @@ -555,7 +596,7 @@ static int ipmmu_attach_device(struct io
> {
> struct ipmmu_vmsa_iommu_priv *priv = to_priv(dev);
> struct iommu_fwspec *fwspec = dev->iommu_fwspec;
> - struct ipmmu_vmsa_device *mmu = priv->mmu;
> + struct ipmmu_vmsa_device *root, *mmu = priv->mmu;
> struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain);
> unsigned long flags;
> unsigned int i;
> @@ -566,11 +607,18 @@ static int ipmmu_attach_device(struct io
> return -ENXIO;
> }
>
> + root = ipmmu_find_root(priv->mmu);
> + if (!root) {
> + dev_err(dev, "Unable to locate root IPMMU\n");
> + return -EAGAIN;
> + }
> +
> spin_lock_irqsave(&domain->lock, flags);
>
> if (!domain->mmu) {
> /* The domain hasn't been used yet, initialize it. */
> domain->mmu = mmu;
> + domain->root = root;
> ret = ipmmu_domain_init_context(domain);
> } else if (domain->mmu != mmu) {
> /*
> @@ -911,6 +959,7 @@ static void ipmmu_device_reset(struct ip
>
> static const struct ipmmu_features ipmmu_features_default = {
> .use_ns_alias_offset = true,
> + .has_cache_leaf_nodes = false,
> };
>
> static const struct of_device_id ipmmu_of_ids[] = {
> @@ -965,19 +1014,31 @@ static int ipmmu_probe(struct platform_d
> mmu->base += IM_NS_ALIAS_OFFSET;
>
> irq = platform_get_irq(pdev, 0);
> - if (irq < 0) {
> - dev_err(&pdev->dev, "no IRQ found\n");
> - return irq;
> - }
>
> - ret = devm_request_irq(&pdev->dev, irq, ipmmu_irq, 0,
> - dev_name(&pdev->dev), mmu);
> - if (ret < 0) {
> - dev_err(&pdev->dev, "failed to request IRQ %d\n", irq);
> - return ret;
> - }
> + /*
> + * Determine if this IPMMU instance is a leaf device by checking
> + * if the renesas,ipmmu-main property exists or not.
> + */
> + if (mmu->features->has_cache_leaf_nodes &&
> + of_find_property(pdev->dev.of_node, "renesas,ipmmu-main", NULL))
> + mmu->is_leaf = true;

Given my previous comments:

if (mmu->features->has_cache_leaf_nodes) {
root_np = of_parse_phandle(pdev->dev.of_node,
"renesas,ipmmu-main", 0)
if (root_np) {
root_pdev = of_find_device_by_phandle(root_np);
mmu->root = platform_get_drvdata(root_pdev);
} else {
mmu->root = mmu;
}
}

or something along those lines?

Robin.

> +
> + /* Root devices have mandatory IRQs */
> + if (ipmmu_is_root(mmu)) {
> + if (irq < 0) {
> + dev_err(&pdev->dev, "no IRQ found\n");
> + return irq;
> + }
>
> - ipmmu_device_reset(mmu);
> + ret = devm_request_irq(&pdev->dev, irq, ipmmu_irq, 0,
> + dev_name(&pdev->dev), mmu);
> + if (ret < 0) {
> + dev_err(&pdev->dev, "failed to request IRQ %d\n", irq);
> + return ret;
> + }
> +
> + ipmmu_device_reset(mmu);
> + }
>
> ret = iommu_device_register(&mmu->iommu);
> if (ret)
>