2.6.25 regression: vivi - scheduling while atomic
From: Gregor Jasny
Date: Mon Apr 21 2008 - 04:56:36 EST
Hi,
during the test of our video conference application I noticed the
following kernel messages:
vivi: open called (minor=0)
vivi: open called (minor=0)
BUG: scheduling while atomic: vidconference/21556/0x00000002
Pid: 21556, comm: vidconference Tainted: P 2.6.25 #1
Call Trace:
[<ffffffff803efc9b>] schedule+0xe5/0x5c7
[<ffffffff80251c90>] __rmqueue_smallest+0x88/0x107
[<ffffffff8023e84b>] getnstimeofday+0x2f/0x83
[<ffffffff8023cf8a>] ktime_get_ts+0x17/0x48
[<ffffffff803f0424>] schedule_timeout+0x1e/0xad
[<ffffffff80220498>] enqueue_task+0x13/0x1e
[<ffffffff803efab8>] wait_for_common+0xf6/0x16b
[<ffffffff802230a0>] default_wake_function+0x0/0xe
[<ffffffff8023a270>] kthread_create+0xa3/0x108
[<ffffffff880d2471>] :vivi:vivi_thread+0x0/0x779
[<ffffffff802634cb>] remap_vmalloc_range+0xa1/0xe6
[<ffffffff80231242>] lock_timer_base+0x26/0x4c
[<ffffffff8023138e>] __mod_timer+0xb6/0xc5
[<ffffffff880d23fc>] :vivi:vivi_start_thread+0x54/0xc9
[<ffffffff88053603>] :videobuf_core:videobuf_streamon+0x6c/0xaa
[<ffffffff8809dba3>] :videodev:__video_do_ioctl+0x1327/0x2ad9
[<ffffffff80222d76>] __wake_up+0x38/0x4f
[<ffffffff80242f1f>] futex_wake+0xdb/0xfa
[<ffffffff8809f6ab>] :videodev:video_ioctl2+0x17c/0x210
[<ffffffff8025bb36>] handle_mm_fault+0x6b1/0x6cb
[<ffffffff8027b47d>] vfs_ioctl+0x55/0x6b
[<ffffffff8027b6e6>] do_vfs_ioctl+0x253/0x264
[<ffffffff8027b733>] sys_ioctl+0x3c/0x5d
[<ffffffff8020afcb>] system_call_after_swapgs+0x7b/0x80
vivi/0: [ffff8100633e1f00/1] timeout
vivi/0: [ffff8100633e1900/0] timeout
vivi: open called (minor=0)
vivi: open called (minor=0)
This happenes on a vanilla 2.6.25 with loaded nvidia graphics module.
System architecture is x86_64. If it matters I'll try to reproduce this
error on a non tainted kernel.
After the error, the vivi driver and the dvb driver of my DVB-T stick were inoperable.
Thanks,
Gregor
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/