Re: Please add to stable: module: don't unlink the module untilwe've removed all exposure.
From: Ben Greear
Date: Tue Jun 04 2013 - 13:46:16 EST
On 06/04/2013 09:53 AM, Ben Greear wrote:
On 06/04/2013 07:07 AM, Joe Lawrence wrote:
On Tue, 04 Jun 2013 15:26:28 +0930
Rusty Russell <rusty@xxxxxxxxxxxxxxx> wrote:
Do you have a backtrace of the 3.9.4 crash? You can add "CFLAGS_module.o
= -O0" to get a clearer backtrace if you want...
Hi Rusty,
See my 3.9 stack traces below, which may or may not be what Ben had
been seeing. If you like, I can try a similar loop as the one you were
testing in the other email.
My stack traces are similar. I had better luck reproducing the problem
once I enabled lots of debugging (slub memory poisoning, lockdep,
object debugging, etc).
I'm using Fedora 17 on 2-core core-i7 (4 CPU threads total) for most of this
testing. We reproduced on dual-core Atom system as well
(32-bit Fedora 14 and Fedora 17). Relatively standard hardware as far
as I know.
I'll run the insmod/rmmod stress test on my patched systems
and see if I can reproduce with the patch in the title applied.
Rusty: I'm also seeing lockups related to migration on stock 3.9.4+
(with and without the 'don't unlink the module...' patch. Much harder
to reproduce. But, that code appears to be mostly called during
module load/unload, so it's possible it is related. The first
traces are from a system with local patches, applied, but a later
post by me has traces from clean upstream kernel.
Further debugging showed that this could be a race, because it seems
that all migration/ threads think they are done with their state machine,
but the atomic thread counter sits at 1, so no progress is ever made.
http://www.mail-archive.com/linux-kernel@xxxxxxxxxxxxxxx/msg443471.html
I reproduced the migration deadlock after a while (loading and unloading
the macvlan module with this command:
for i in `seq 10000`; do modprobe macvlan; rmmod macvlan; done
I did not see the kobj crash, but this kernel was running your patch
that makes the problem go away for me, for whatever reason.
I have some printk debugging in (see bottom of email) and was using a serial console, so things
were probably running a bit slower than on most systems. Here is trace
from my kernel with local patches and not so much debugging enabled
(this is NOT a clean upstream kernel, though I reproduced the same thing
with a clean upstream 3.9.4 kernel plus your module unlink patch yesterday).
__stop_machine, num-threads: 4, fn: __try_stop_module data: ffff8801c6ae7f28
cpu: 0 loops: 1 jiffies: 4299011449 timeout: 4299011448 curstate: 0 smdata->state: 1 thread_ack: 4
cpu: 1 loops: 1 jiffies: 4299011449 timeout: 4299011448 curstate: 0 smdata->state: 1 thread_ack: 4
cpu: 2 loops: 1 jiffies: 4299011449 timeout: 4299011448 curstate: 0 smdata->state: 1 thread_ack: 3
cpu: 3 loops: 1 jiffies: 4299011449 timeout: 4299011448 curstate: 0 smdata->state: 1 thread_ack: 2
__stop_machine, num-threads: 4, fn: __unlink_module data: ffffffffa0aeeab0
cpu: 0 loops: 1 jiffies: 4299011501 timeout: 4299011500 curstate: 0 smdata->state: 1 thread_ack: 4
cpu: 1 loops: 1 jiffies: 4299011501 timeout: 4299011500 curstate: 0 smdata->state: 1 thread_ack: 4
cpu: 3 loops: 1 jiffies: 4299011501 timeout: 4299011500 curstate: 0 smdata->state: 1 thread_ack: 3
ath: wiphy0: Failed to stop TX DMA, queues=0x005!
cpu: 2 loops: 1 jiffies: 4299011501 timeout: 4299011500 curstate: 0 smdata->state: 1 thread_ack: 2
BUG: soft lockup - CPU#3 stuck for 23s! [migration/3:29]
Modules linked in: nf_nat_ipv4 nf_nat veth 8021q garp stp mrp llc wanlink(O) pktgen lockd sunrpc coretemp hwmon mperf intel_po]
CPU 3
Pid: 29, comm: migration/3 Tainted: G C O 3.9.4+ #60 To be filled by O.E.M. To be filled by O.E.M./To be filled by O.E.
RIP: 0010:[<ffffffff8109d69c>] [<ffffffff8109d69c>] tasklet_action+0x46/0xcc
RSP: 0000:ffff88022bd83ed8 EFLAGS: 00000282
RAX: ffff88022bd8e080 RBX: ffff8802222a4000 RCX: ffffffff81a90f06
RDX: ffff88021f48afa8 RSI: 0000000000000000 RDI: ffffffff81a050b0
RBP: ffff88022bd83ee8 R08: 0000000000000000 R09: 0000000000000000
R10: 00000000000005f2 R11: 00000000fd010018 R12: ffff88022bd83e48
R13: ffffffff815d145d R14: ffff88022bd83ee8 R15: ffff880222282000
FS: 0000000000000000(0000) GS:ffff88022bd80000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00007fa8a06d5000 CR3: 00000002007ea000 CR4: 00000000000007e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process migration/3 (pid: 29, threadinfo ffff880222282000, task ffff880222279770)
Stack:
ffffffff81a050b0 ffff880222282000 ffff88022bd83f78 ffffffff8109db1f
ffff88022bd83f08 ffff880222282010 ffff880222283fd8 04208040810b79ef
00000001003db5d8 000000032bd8e150 ffff880222282000 0000000000000030
Call Trace:
<IRQ>
[<ffffffff8109db1f>] __do_softirq+0x107/0x23c
[<ffffffff8109dce6>] irq_exit+0x4b/0xa8
[<ffffffff815d257f>] smp_apic_timer_interrupt+0x8b/0x99
[<ffffffff815d145d>] apic_timer_interrupt+0x6d/0x80
<EOI>
[<ffffffff810f9984>] ? stop_machine_cpu_stop+0xc7/0x145
[<ffffffff810f9974>] ? stop_machine_cpu_stop+0xb7/0x145
[<ffffffff81016445>] ? sched_clock+0x9/0xd
[<ffffffff810f98bd>] ? stop_one_cpu_nowait+0x30/0x30
[<ffffffff810f9624>] cpu_stopper_thread+0xbd/0x176
[<ffffffff815c9065>] ? __schedule+0x59f/0x5e7
[<ffffffff810bb434>] smpboot_thread_fn+0x217/0x21f
[<ffffffff810bb21d>] ? test_ti_thread_flag.clone.0+0x11/0x11
[<ffffffff810b4a09>] kthread+0xb5/0xbd
[<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
[<ffffffff815d07ac>] ret_from_fork+0x7c/0xb0
[<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
Code: 48 8b 14 25 80 e0 00 00 65 48 c7 04 25 80 e0 00 00 00 00 00 00 65 48 03 04 25 d8 da 00 00 65 48 89 04 25 88 e0 00 00 fb
------------[ cut here ]------------
WARNING: at /home/greearb/git/linux-3.9.dev.y/kernel/watchdog.c:245 watchdog_overflow_callback+0x9b/0xa6()
Hardware name: To be filled by O.E.M.
Watchdog detected hard LOCKUP on cpu 1
Modules linked in: nf_nat_ipv4 nf_nat veth 8021q garp stp mrp llc wanlink(O) pktgen lockd sunrpc coretemp hwmon mperf intel_po]
Pid: 17, comm: migration/1 Tainted: G C O 3.9.4+ #60
Call Trace:
<NMI> [<ffffffff810963e1>] warn_slowpath_common+0x85/0x9f
[<ffffffff8109649e>] warn_slowpath_fmt+0x46/0x48
[<ffffffff810c4dfb>] ? sched_clock_cpu+0x44/0xce
[<ffffffff81103b70>] watchdog_overflow_callback+0x9b/0xa6
[<ffffffff8113354f>] __perf_event_overflow+0x137/0x1cb
[<ffffffff8101db3f>] ? x86_perf_event_set_period+0x107/0x113
[<ffffffff81133a9e>] perf_event_overflow+0x14/0x16
[<ffffffff810230dc>] intel_pmu_handle_irq+0x2b0/0x32d
[<ffffffff815cbc51>] perf_event_nmi_handler+0x19/0x1b
[<ffffffff815cb4ca>] nmi_handle+0x55/0x7e
[<ffffffff815cb59b>] do_nmi+0xa8/0x2db
[<ffffffff815cac31>] end_repeat_nmi+0x1e/0x2e
[<ffffffff810f9942>] ? stop_machine_cpu_stop+0x85/0x145
[<ffffffff810f9942>] ? stop_machine_cpu_stop+0x85/0x145
[<ffffffff810f9942>] ? stop_machine_cpu_stop+0x85/0x145
<<EOE>> [<ffffffff810c6b1c>] ? set_next_entity+0x28/0x7e
[<ffffffff810f98bd>] ? stop_one_cpu_nowait+0x30/0x30
[<ffffffff810f9624>] cpu_stopper_thread+0xbd/0x176
[<ffffffff815c9065>] ? __schedule+0x59f/0x5e7
[<ffffffff810bb434>] smpboot_thread_fn+0x217/0x21f
[<ffffffff810bb21d>] ? test_ti_thread_flag.clone.0+0x11/0x11
[<ffffffff810b4a09>] kthread+0xb5/0xbd
[<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
[<ffffffff815d07ac>] ret_from_fork+0x7c/0xb0
[<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
---[ end trace dcd772d6fdf499cf ]---
...
SysRq : Show backtrace of all active CPUs
sending NMI to all CPUs:
NMI backtrace for cpu 2
CPU 2
Pid: 23, comm: migration/2 Tainted: G WC O 3.9.4+ #60 To be filled by O.E.M. To be filled by O.E.M./To be filled by O.E.
RIP: 0010:[<ffffffff810f9987>] [<ffffffff810f9987>] stop_machine_cpu_stop+0xca/0x145
RSP: 0018:ffff880222219cf8 EFLAGS: 00000012
RAX: 00000001003db5d7 RBX: ffff8801c6ae7e18 RCX: 0000000000000097
RDX: 0000000000000002 RSI: 0000000000000006 RDI: 0000000000000246
RBP: ffff880222219d68 R08: 00000001003dc935 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000000 R12: ffff8801c6ae7e3c
R13: 0000000000000000 R14: 0000000000000002 R15: 0000000000000002
FS: 0000000000000000(0000) GS:ffff88022bd00000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00000000027e2178 CR3: 000000021955d000 CR4: 00000000000007e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process migration/2 (pid: 23, threadinfo ffff880222218000, task ffff880222210000)
Stack:
ffff880200000001 ffff880200000002 ffff880222219d28 ffffffff810c6b1c
ffff880222219d88 00ffffff8100f8a4 000000051058262e 0000000000000292
ffff880222219d58 ffff88022bd0e400 ffff8801c6ae7d08 ffff880222218000
Call Trace:
[<ffffffff810c6b1c>] ? set_next_entity+0x28/0x7e
[<ffffffff810f98bd>] ? stop_one_cpu_nowait+0x30/0x30
[<ffffffff810f9624>] cpu_stopper_thread+0xbd/0x176
[<ffffffff815c9065>] ? __schedule+0x59f/0x5e7
[<ffffffff810bb434>] smpboot_thread_fn+0x217/0x21f
[<ffffffff810bb21d>] ? test_ti_thread_flag.clone.0+0x11/0x11
[<ffffffff810b4a09>] kthread+0xb5/0xbd
[<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
[<ffffffff815d07ac>] ret_from_fork+0x7c/0xb0
[<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
Code: 48 c7 c7 24 fa 80 81 89 44 24 08 8b 43 20 89 04 24 31 c0 e8 7a e1 4c 00 4c 8b 05 85 c6 9e 00 49 81 c0 88 13 00 00 f3 90
NMI backtrace for cpu 0
CPU 0
Pid: 8, comm: migration/0 Tainted: G WC O 3.9.4+ #60 To be filled by O.E.M. To be filled by O.E.M./To be filled by O.E..
RIP: 0010:[<ffffffff810f9984>] [<ffffffff810f9984>] stop_machine_cpu_stop+0xc7/0x145
RSP: 0000:ffff880222145cf8 EFLAGS: 00000006
RAX: 00000001003db5d7 RBX: ffff8801c6ae7e18 RCX: 0000000000000004
RDX: 0000000000000002 RSI: ffff88022bc0de68 RDI: 0000000000000246
RBP: ffff880222145d68 R08: 00000001003dc95f R09: 0000000000000001
R10: ffff880222145bf8 R11: 0000000000000000 R12: ffff8801c6ae7e3c
R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000002
FS: 0000000000000000(0000) GS:ffff88022bc00000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00000000006fb9a8 CR3: 000000021af43000 CR4: 00000000000007f0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process migration/0 (pid: 8, threadinfo ffff880222144000, task ffff88022213aee0)
Stack:
ffff880200000001 ffff880200000004 ffff880222145d28 ffffffff810c6b1c
ffff880222145d88 01ffffff8100f8a4 000000051023ce2c 0000000000000292
ffff880222145d98 ffff88022bc0e400 ffff8801c6ae7d08 ffff880222144000
Call Trace:
[<ffffffff810c6b1c>] ? set_next_entity+0x28/0x7e
[<ffffffff810f98bd>] ? stop_one_cpu_nowait+0x30/0x30
[<ffffffff810f9624>] cpu_stopper_thread+0xbd/0x176
[<ffffffff815c9065>] ? __schedule+0x59f/0x5e7
[<ffffffff810bb434>] smpboot_thread_fn+0x217/0x21f
[<ffffffff810bb21d>] ? test_ti_thread_flag.clone.0+0x11/0x11
[<ffffffff810b4a09>] kthread+0xb5/0xbd
[<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
[<ffffffff815d07ac>] ret_from_fork+0x7c/0xb0
[<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
Code: 44 89 f6 48 c7 c7 24 fa 80 81 89 44 24 08 8b 43 20 89 04 24 31 c0 e8 7a e1 4c 00 4c 8b 05 85 c6 9e 00 49 81 c0 88 13 00
NMI backtrace for cpu 1
CPU 1
Pid: 17, comm: migration/1 Tainted: G WC O 3.9.4+ #60 To be filled by O.E.M. To be filled by O.E.M./To be filled by O.E.
RIP: 0010:[<ffffffff810f9984>] [<ffffffff810f9984>] stop_machine_cpu_stop+0xc7/0x145
RSP: 0000:ffff88022217dcf8 EFLAGS: 00000012
RAX: 00000001003db5d7 RBX: ffff8801c6ae7e18 RCX: 0000000000000094
RDX: 0000000000000002 RSI: 0000000000000006 RDI: 0000000000000246
RBP: ffff88022217dd68 R08: 00000001003dc935 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000000 R12: ffff8801c6ae7e3c
R13: 0000000000000000 R14: 0000000000000001 R15: 0000000000000002
FS: 0000000000000000(0000) GS:ffff88022bc80000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00007fbe801fd000 CR3: 00000001be9bb000 CR4: 00000000000007e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process migration/1 (pid: 17, threadinfo ffff88022217c000, task ffff88022216ddc0)
Stack:
ffff880200000001 ffff880200000004 ffff88022217dd28 ffffffff810c6b1c
ffff88022217dd88 00ffffff8100f8a4 00000004c3cbfc0c 0000000000000292
ffff88022217dd58 ffff88022bc8e400 ffff8801c6ae7d08 ffff88022217c000
Call Trace:
[<ffffffff810c6b1c>] ? set_next_entity+0x28/0x7e
[<ffffffff810f98bd>] ? stop_one_cpu_nowait+0x30/0x30
[<ffffffff810f9624>] cpu_stopper_thread+0xbd/0x176
[<ffffffff815c9065>] ? __schedule+0x59f/0x5e7
[<ffffffff810bb434>] smpboot_thread_fn+0x217/0x21f
[<ffffffff810bb21d>] ? test_ti_thread_flag.clone.0+0x11/0x11
[<ffffffff810b4a09>] kthread+0xb5/0xbd
[<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
[<ffffffff815d07ac>] ret_from_fork+0x7c/0xb0
[<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
Code: 44 89 f6 48 c7 c7 24 fa 80 81 89 44 24 08 8b 43 20 89 04 24 31 c0 e8 7a e1 4c 00 4c 8b 05 85 c6 9e 00 49 81 c0 88 13 00
NMI backtrace for cpu 3
CPU 3
Pid: 29, comm: migration/3 Tainted: G WC O 3.9.4+ #60 To be filled by O.E.M. To be filled by O.E.M./To be filled by O.E.
RIP: 0010:[<ffffffff812f6975>] [<ffffffff812f6975>] delay_tsc+0x83/0xee
RSP: 0000:ffff88022bd83b60 EFLAGS: 00000046
RAX: 00000b04f0fdf718 RBX: ffff880222282000 RCX: ffff880222282010
RDX: 0000000000000000 RSI: 0000000000000002 RDI: 0000000000289672
RBP: ffff88022bd83bb0 R08: 0000000000000040 R09: 0000000000000001
R10: ffff88022bd83ad0 R11: 0000000000000000 R12: 00000000f0fdf6e8
R13: 0000000000000003 R14: ffff880222282000 R15: 0000000000000001
FS: 0000000000000000(0000) GS:ffff88022bd80000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00007fa8a06d5000 CR3: 00000002007ea000 CR4: 00000000000007e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process migration/3 (pid: 29, threadinfo ffff880222282000, task ffff880222279770)
Stack:
000000000000000f 0000000000000002 ffff880222282010 0028967200000082
ffff88022bd83b90 0000000000000001 000000000000006c 0000000000000007
0000000000000086 0000000000000001 ffff88022bd83bc0 ffffffff812f68c5
Call Trace:
<IRQ>
[<ffffffff812f68c5>] __const_udelay+0x28/0x2a
[<ffffffff8102ff25>] arch_trigger_all_cpu_backtrace+0x66/0x7d
[<ffffffff813ae154>] sysrq_handle_showallcpus+0xe/0x10
[<ffffffff813ae46b>] __handle_sysrq+0xbf/0x15b
[<ffffffff813ae80e>] handle_sysrq+0x2c/0x2e
[<ffffffff813c25a2>] serial8250_rx_chars+0x13c/0x1b9
[<ffffffff813c2691>] serial8250_handle_irq+0x72/0xa8
[<ffffffff813c2752>] serial8250_default_handle_irq+0x23/0x28
[<ffffffff813c148c>] serial8250_interrupt+0x4d/0xc6
[<ffffffff811046b8>] handle_irq_event_percpu+0x7a/0x1e5
[<ffffffff81104864>] handle_irq_event+0x41/0x61
[<ffffffff81107028>] handle_edge_irq+0xa6/0xcb
[<ffffffff81011d9f>] handle_irq+0x24/0x2d
[<ffffffff815d248d>] do_IRQ+0x4d/0xb4
[<ffffffff815ca5ad>] common_interrupt+0x6d/0x6d
[<ffffffff810a47ee>] ? run_timer_softirq+0x24/0x1df
[<ffffffff8109dabd>] ? __do_softirq+0xa5/0x23c
[<ffffffff8109db8a>] ? __do_softirq+0x172/0x23c
[<ffffffff8109dce6>] irq_exit+0x4b/0xa8
[<ffffffff815d257f>] smp_apic_timer_interrupt+0x8b/0x99
[<ffffffff815d145d>] apic_timer_interrupt+0x6d/0x80
<EOI>
[<ffffffff810f9984>] ? stop_machine_cpu_stop+0xc7/0x145
[<ffffffff810f9974>] ? stop_machine_cpu_stop+0xb7/0x145
[<ffffffff81016445>] ? sched_clock+0x9/0xd
[<ffffffff810f98bd>] ? stop_one_cpu_nowait+0x30/0x30
[<ffffffff810f9624>] cpu_stopper_thread+0xbd/0x176
[<ffffffff815c9065>] ? __schedule+0x59f/0x5e7
[<ffffffff810bb434>] smpboot_thread_fn+0x217/0x21f
[<ffffffff810bb21d>] ? test_ti_thread_flag.clone.0+0x11/0x11
[<ffffffff810b4a09>] kthread+0xb5/0xbd
[<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
[<ffffffff815d07ac>] ret_from_fork+0x7c/0xb0
[<ffffffff810b4954>] ? kthread_freezable_should_stop+0x60/0x60
Code: fa d1 ff 66 90 89 c2 44 29 e2 3b 55 cc 73 4a ff 4b 1c 48 8b 4d c0 48 8b 11 80 e2 08 74 0b 89 45 b8 e8 08 28 2d 00 8b 45
commit 5b783a651bb9941923cd8dbff6d0ce80d2617f97
Author: Ben Greear <greearb@xxxxxxxxxxxxxxx>
Date: Mon Jun 3 13:31:09 2013 -0700
debugging: Instrument the 'migration/[0-x]' process a bit.
Hoping to figure out the deadlock we see occassionally due
to migration processes all spinning in busy loop.
Signed-off-by: Ben Greear <greearb@xxxxxxxxxxxxxxx>
diff --git a/kernel/stop_machine.c b/kernel/stop_machine.c
index c09f295..f7a1de5 100644
--- a/kernel/stop_machine.c
+++ b/kernel/stop_machine.c
@@ -408,6 +408,9 @@ static int stop_machine_cpu_stop(void *data)
int cpu = smp_processor_id(), err = 0;
unsigned long flags;
bool is_active;
+ unsigned long loops = 0;
+ unsigned long start_at = jiffies;
+ unsigned long timeout = start_at - 1;
/*
* When called from stop_machine_from_inactive_cpu(), irq might
@@ -422,6 +425,13 @@ static int stop_machine_cpu_stop(void *data)
/* Simple state machine */
do {
+ loops++;
+ if (time_after(jiffies, timeout)) {
+ printk("cpu: %i loops: %lu jiffies: %lu timeout: %lu curstate: %i smdata->state:
+ cpu, loops, jiffies, timeout, curstate, smdata->state,
+ atomic_read(&smdata->thread_ack));
+ timeout = jiffies + 5000;
+ }
/* Chill out and ensure we re-read stopmachine_state. */
cpu_relax();
if (smdata->state != curstate) {
@@ -473,6 +483,14 @@ int __stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus)
/* Set the initial state and stop all online cpus. */
set_state(&smdata, STOPMACHINE_PREPARE);
+ {
+ char ksym_buf[KSYM_NAME_LEN];
+ printk("__stop_machine, num-threads: %i, fn: %s data: %p\n",
+ smdata.num_threads,
+ kallsyms_lookup((unsigned long)fn, NULL, NULL, NULL,
+ ksym_buf),
+ data);
+ }
return stop_cpus(cpu_online_mask, stop_machine_cpu_stop, &smdata);
}
@@ -517,10 +535,16 @@ int stop_machine_from_inactive_cpu(int (*fn)(void *), void *data,
.active_cpus = cpus };
struct cpu_stop_done done;
int ret;
+ char ksym_buf[KSYM_NAME_LEN];
/* Local CPU must be inactive and CPU hotplug in progress. */
BUG_ON(cpu_active(raw_smp_processor_id()));
smdata.num_threads = num_active_cpus() + 1; /* +1 for local */
+ printk("stop-machine-from-inactive-cpu, num-threads: %i fn: %s(%p)\n",
+ smdata.num_threads,
+ kallsyms_lookup((unsigned long)fn, NULL, NULL, NULL,
+ ksym_buf),
+ data);
/* No proper task established and can't sleep - busy wait for lock. */
while (!mutex_trylock(&stop_cpus_mutex))
Thanks,
Ben
--
Ben Greear <greearb@xxxxxxxxxxxxxxx>
Candela Technologies Inc http://www.candelatech.com
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/