The other way around. mapping/unmapping IOVAs doesn't comes for free.> I've also managed to trigger some of them now that I have access to
> a decent box with nvme storage.
I only have 2x NVMe SSDs when this occurs - I should not be hitting this...
Out of curiosity, have you tried
> with the SMMU disabled? I'm wondering whether we hit some livelock
> condition on unmapping buffers...
No, but I can give it a try. Doing that should lower the CPU usage, though,
so maybe masks the issue - probably not.
Lots of CPU lockup can is performance issue if there isn't obvious bug.
I am wondering if you may explain it a bit why enabling SMMU may save
CPU a it?
I'm trying to find out whether the NVMe map/unmap patterns trigger
something unexpected in the SMMU driver, but that's a very long shot.
So I tested v5.5-rc3 with and without the SMMU enabled, and without
the SMMU enabled I don't get the lockup.
Obviously this is not conclusive, especially with such limited
testing - 5 minute runs each. The CPU load goes up when disabling the
SMMU, but that could be attributed to extra throughput (1183K ->
1539K) loading.
I do notice that since we complete the NVMe request in irq context,
we also do the DMA unmap, i.e. talk to the SMMU, in the same context,
which is less than ideal.
I need to finish for the Christmas break today, so can't check this
much further ATM.