On Wed, 2007-01-24 at 18:58 -0800, Linus Torvalds wrote: > It's been more than a week since -rc5, but I blame everybody (including > me) being away for Linux.conf.au and then me waiting for a few days > afterwards to let everybody sync up.
2.6.20-rc6-git (today) on a dual core laptop: PM: Preparing system for mem sleep Disabling non-boot CPUs ... ======================================================= [ INFO: possible circular locking dependency detected ] 2.6.20-rc6 #3 ------------------------------------------------------- pm-suspend/3601 is trying to acquire lock: (cpu_bitmask_lock){--..}, at: [<c032cd2b>] mutex_lock+0x1c/0x1f but task is already holding lock: (workqueue_mutex){--..}, at: [<c032cd2b>] mutex_lock+0x1c/0x1f which lock already depends on the new lock. the existing dependency chain (in reverse order) is: -> #3 (workqueue_mutex){--..}: [<c0140880>] __lock_acquire+0x8dd/0xa04 [<c0140c90>] lock_acquire+0x56/0x6f [<c032cb80>] __mutex_lock_slowpath+0xe5/0x274 [<c032cd2b>] mutex_lock+0x1c/0x1f [<c0136d14>] __create_workqueue+0x61/0x136 [<f8bfe62e>] cpufreq_governor_dbs+0xa1/0x30e [cpufreq_ondemand] [<c02b2c3c>] __cpufreq_governor+0x9e/0xd2 [<c02b2df7>] __cpufreq_set_policy+0x187/0x209 [<c02b3056>] store_scaling_governor+0x164/0x1b1 [<c02b24f9>] store+0x37/0x48 [<c01aeb8d>] sysfs_write_file+0xb3/0xdb [<c0175e0f>] vfs_write+0xaf/0x163 [<c017645d>] sys_write+0x3d/0x61 [<c0103f8c>] sysenter_past_esp+0x5d/0x99 [<ffffffff>] 0xffffffff -> #2 (dbs_mutex){--..}: [<c0140880>] __lock_acquire+0x8dd/0xa04 [<c0140c90>] lock_acquire+0x56/0x6f [<c032cb80>] __mutex_lock_slowpath+0xe5/0x274 [<c032cd2b>] mutex_lock+0x1c/0x1f [<f8bfe612>] cpufreq_governor_dbs+0x85/0x30e [cpufreq_ondemand] [<c02b2c3c>] __cpufreq_governor+0x9e/0xd2 [<c02b2df7>] __cpufreq_set_policy+0x187/0x209 [<c02b3056>] store_scaling_governor+0x164/0x1b1 [<c02b24f9>] store+0x37/0x48 [<c01aeb8d>] sysfs_write_file+0xb3/0xdb [<c0175e0f>] vfs_write+0xaf/0x163 [<c017645d>] sys_write+0x3d/0x61 [<c0103f8c>] sysenter_past_esp+0x5d/0x99 [<ffffffff>] 0xffffffff -> #1 (&policy->lock){--..}: [<c0140880>] __lock_acquire+0x8dd/0xa04 [<c0140c90>] lock_acquire+0x56/0x6f [<c032cb80>] __mutex_lock_slowpath+0xe5/0x274 [<c032cd2b>] mutex_lock+0x1c/0x1f [<c02b2ea2>] cpufreq_set_policy+0x29/0x79 [<c02b3804>] cpufreq_add_dev+0x342/0x48a [<c025a799>] sysdev_driver_register+0x5f/0xa9 [<c02b2ad5>] cpufreq_register_driver+0xac/0x175 [<c046fd3f>] centrino_init+0x9b/0xa2 [<c01004c4>] init+0x11b/0x2c8 [<c0104c87>] kernel_thread_helper+0x7/0x10 [<ffffffff>] 0xffffffff -> #0 (cpu_bitmask_lock){--..}: [<c0140781>] __lock_acquire+0x7de/0xa04 [<c0140c90>] lock_acquire+0x56/0x6f [<c032cb80>] __mutex_lock_slowpath+0xe5/0x274 [<c032cd2b>] mutex_lock+0x1c/0x1f [<c0144840>] lock_cpu_hotplug+0x6c/0x78 [<c02b3264>] cpufreq_driver_target+0x28/0x5e [<c02b398e>] cpufreq_cpu_callback+0x42/0x52 [<c0133cd3>] notifier_call_chain+0x20/0x31 [<c0133d00>] raw_notifier_call_chain+0x8/0xa [<c014452e>] _cpu_down+0x47/0x1fb [<c01448c7>] disable_nonboot_cpus+0x7b/0x100 [<c014853f>] enter_state+0x91/0x1bb [<c01486ef>] state_store+0x86/0x9c [<c01ae8e8>] subsys_attr_store+0x20/0x25 [<c01aeb8d>] sysfs_write_file+0xb3/0xdb [<c0175e0f>] vfs_write+0xaf/0x163 [<c017645d>] sys_write+0x3d/0x61 [<c0103f8c>] sysenter_past_esp+0x5d/0x99 [<ffffffff>] 0xffffffff other info that might help us debug this: 4 locks held by pm-suspend/3601: #0: (pm_mutex){--..}, at: [<c01484ee>] enter_state+0x40/0x1bb #1: (cpu_add_remove_lock){--..}, at: [<c032cd2b>] mutex_lock+0x1c/0x1f #2: (cache_chain_mutex){--..}, at: [<c032cd2b>] mutex_lock+0x1c/0x1f #3: (workqueue_mutex){--..}, at: [<c032cd2b>] mutex_lock+0x1c/0x1f stack backtrace: [<c0104ff6>] show_trace_log_lvl+0x1a/0x2f [<c01056af>] show_trace+0x12/0x14 [<c0105761>] dump_stack+0x16/0x18 [<c013f101>] print_circular_bug_tail+0x5f/0x68 [<c0140781>] __lock_acquire+0x7de/0xa04 [<c0140c90>] lock_acquire+0x56/0x6f [<c032cb80>] __mutex_lock_slowpath+0xe5/0x274 [<c032cd2b>] mutex_lock+0x1c/0x1f [<c0144840>] lock_cpu_hotplug+0x6c/0x78 [<c02b3264>] cpufreq_driver_target+0x28/0x5e [<c02b398e>] cpufreq_cpu_callback+0x42/0x52 [<c0133cd3>] notifier_call_chain+0x20/0x31 [<c0133d00>] raw_notifier_call_chain+0x8/0xa [<c014452e>] _cpu_down+0x47/0x1fb [<c01448c7>] disable_nonboot_cpus+0x7b/0x100 [<c014853f>] enter_state+0x91/0x1bb [<c01486ef>] state_store+0x86/0x9c [<c01ae8e8>] subsys_attr_store+0x20/0x25 [<c01aeb8d>] sysfs_write_file+0xb3/0xdb [<c0175e0f>] vfs_write+0xaf/0x163 [<c017645d>] sys_write+0x3d/0x61 [<c0103f8c>] sysenter_past_esp+0x5d/0x99 ======================= Breaking affinity for irq 1 Breaking affinity for irq 12 Breaking affinity for irq 21 Breaking affinity for irq 22 Breaking affinity for irq 219 CPU 1 is now offline - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/