Re: [RFC PATCH v3] pci: Concurrency issue during pci enable bridge
From: Bjorn Helgaas
Date: Wed Aug 16 2017 - 09:44:01 EST
On Fri, Aug 04, 2017 at 08:27:28PM +0530, Srinath Mannam wrote:
> Concurrency issue is observed during pci enable bridge called
> for multiple pci devices initialization in SMP system.
>
> Setup details:
> - SMP system with 8 ARMv8 cores running Linux kernel(4.11).
> - Two EPs are connected to PCIe RC through bridge as shown
> in the below figure.
>
> [RC]
> |
> [BRIDGE]
> |
> -----------
> | |
> [EP] [EP]
>
> Issue description:
> After PCIe enumeration completed EP driver probe function called
> for both the devices from two CPUS simultaneously.
> From EP probe function, pci_enable_device_mem called for both the EPs.
> This function called pci_enable_bridge enable for all the bridges
> recursively in the path of EP to RC.
>
> Inside pci_enable_bridge function, at two places concurrency issue is
> observed.
>
> Place 1:
> CPU 0:
> 1. Done Atomic increment dev->enable_cnt
> in pci_enable_device_flags
> 2. Inside pci_enable_resources
> 3. Completed pci_read_config_word(dev, PCI_COMMAND, &cmd)
> 4. Ready to set PCI_COMMAND_MEMORY (0x2) in
> pci_write_config_word(dev, PCI_COMMAND, cmd)
> CPU 1:
> 1. Check pci_is_enabled in function pci_enable_bridge
> and it is true
> 2. Check (!dev->is_busmaster) also true
> 3. Gone into pci_set_master
> 4. Completed pci_read_config_word(dev, PCI_COMMAND, &old_cmd)
> 5. Ready to set PCI_COMMAND_MASTER (0x4) in
> pci_write_config_word(dev, PCI_COMMAND, cmd)
>
> By the time of last point for both the CPUs are read value 0 and
> ready to write 2 and 4.
> After last point final value in PCI_COMMAND register is 4 instead of 6.
>
> Place 2:
> CPU 0:
> 1. Done Atomic increment dev->enable_cnt in
> pci_enable_device_flags
>
> Signed-off-by: Srinath Mannam <srinath.mannam@xxxxxxxxxxxx>
> ---
> drivers/pci/pci.c | 8 ++++++--
> 1 file changed, 6 insertions(+), 2 deletions(-)
>
> diff --git a/drivers/pci/pci.c b/drivers/pci/pci.c
> index af0cc34..12721df 100644
> --- a/drivers/pci/pci.c
> +++ b/drivers/pci/pci.c
> @@ -52,6 +52,7 @@ static void pci_pme_list_scan(struct work_struct *work);
> static LIST_HEAD(pci_pme_list);
> static DEFINE_MUTEX(pci_pme_list_mutex);
> static DECLARE_DELAYED_WORK(pci_pme_work, pci_pme_list_scan);
> +static DEFINE_MUTEX(pci_bridge_mutex);
>
> struct pci_pme_device {
> struct list_head list;
> @@ -1348,10 +1349,11 @@ static void pci_enable_bridge(struct pci_dev *dev)
> if (bridge)
> pci_enable_bridge(bridge);
>
> + mutex_lock(&pci_bridge_mutex);
> if (pci_is_enabled(dev)) {
> if (!dev->is_busmaster)
> pci_set_master(dev);
> - return;
> + goto end;
> }
>
> retval = pci_enable_device(dev);
> @@ -1359,6 +1361,8 @@ static void pci_enable_bridge(struct pci_dev *dev)
> dev_err(&dev->dev, "Error enabling bridge (%d), continuing\n",
> retval);
> pci_set_master(dev);
> +end:
> + mutex_unlock(&pci_bridge_mutex);
I think this will deadlock because we're holding pci_bridge_mutex
while we call pci_enable_device(), which may recursively call
pci_enable_bridge(), which would try to acquire pci_bridge_mutex
again. My original suggestion of a mutex in the host bridge would
have the same problem.
We talked about using device_lock() earlier. You found some problems
with that, and I'd like to understand them better. You said:
> But the pci_enable_bridge is called in the context of the driver
> probe function, we will have nexted lock problem.
The driver core does hold device_lock() while calling the driver probe
function, in this path:
device_initial_probe
__device_attach
device_lock(dev) # <-- lock
__device_attach_driver
...
pci_device_probe
...
->probe # driver probe function
device_unlock(dev) # <-- unlock
I didn't see your patch using device_lock(), but what I had in mind
was something like the patch below, where pci_enable_bridge() acquires
the device_lock() of the bridge.
For the sake of argument, assume a hierarchy:
bridge A -> bridge B -> endpoint C
Here's what I think will happen:
device_lock(C) # driver core
...
->probe(C) # driver probe function
pci_enable_device_flags(C)
pci_enable_bridge(B) # enable C's upstream bridge
device_lock(B)
pci_enable_bridge(A) # enable B's upstream bridge
device_lock(A) # A has no upstream bridge
pci_enable_device(A)
do_pci_enable_device(A) # update A PCI_COMMAND
pci_set_master(A) # update A PCI_COMMAND
device_unlock(A)
pci_enable_device(B) # update B PCI_COMMAND
pci_set_master(B) # update B PCI_COMMAND
device_unlock(B)
do_pci_enable_device(C) # update C PCI_COMMAND
device_unlock(C)
I don't see a nested lock problem here. What am I missing?
Bjorn
diff --git a/drivers/pci/pci.c b/drivers/pci/pci.c
index e8e40dea2842..38154ba628a9 100644
--- a/drivers/pci/pci.c
+++ b/drivers/pci/pci.c
@@ -1344,6 +1344,7 @@ static void pci_enable_bridge(struct pci_dev *dev)
struct pci_dev *bridge;
int retval;
+ device_lock(&dev->dev);
bridge = pci_upstream_bridge(dev);
if (bridge)
pci_enable_bridge(bridge);
@@ -1351,7 +1352,7 @@ static void pci_enable_bridge(struct pci_dev *dev)
if (pci_is_enabled(dev)) {
if (!dev->is_busmaster)
pci_set_master(dev);
- return;
+ goto out;
}
retval = pci_enable_device(dev);
@@ -1359,6 +1360,9 @@ static void pci_enable_bridge(struct pci_dev *dev)
dev_err(&dev->dev, "Error enabling bridge (%d), continuing\n",
retval);
pci_set_master(dev);
+
+out:
+ device_unlock(&dev->dev);
}
static int pci_enable_device_flags(struct pci_dev *dev, unsigned long flags)