INFO: task kworker/R-wg-cr:5847 blocked for more than 143 seconds.
Not tainted 6.13.0-rc1-syzkaller-00337-g7503345ac5f5 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/R-wg-cr state:D stack:29264 pid:5847 tgid:5847 ppid:2 flags:0x00004000
Call Trace:
context_switch kernel/sched/core.c:5369 [inline]
__schedule+0xe58/0x5ad0 kernel/sched/core.c:6756
__schedule_loop kernel/sched/core.c:6833 [inline]
schedule+0xe7/0x350 kernel/sched/core.c:6848
schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6905
__mutex_lock_common kernel/locking/mutex.c:665 [inline]
__mutex_lock+0x62b/0xa60 kernel/locking/mutex.c:735
worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
rescuer_thread+0x3bf/0xe20 kernel/workqueue.c:3471
kthread+0x2c1/0x3a0 kernel/kthread.c:389
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
INFO: task kworker/R-wg-cr:5849 blocked for more than 143 seconds.
Not tainted 6.13.0-rc1-syzkaller-00337-g7503345ac5f5 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/R-wg-cr state:D stack:30104 pid:5849 tgid:5849 ppid:2 flags:0x00004000
Call Trace:
context_switch kernel/sched/core.c:5369 [inline]
__schedule+0xe58/0x5ad0 kernel/sched/core.c:6756
__schedule_loop kernel/sched/core.c:6833 [inline]
schedule+0xe7/0x350 kernel/sched/core.c:6848
schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6905
__mutex_lock_common kernel/locking/mutex.c:665 [inline]
__mutex_lock+0x62b/0xa60 kernel/locking/mutex.c:735
worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
rescuer_thread+0x3bf/0xe20 kernel/workqueue.c:3471
kthread+0x2c1/0x3a0 kernel/kthread.c:389
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
INFO: task kworker/R-wg-cr:5853 blocked for more than 144 seconds.
Not tainted 6.13.0-rc1-syzkaller-00337-g7503345ac5f5 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/R-wg-cr state:D stack:29072 pid:5853 tgid:5853 ppid:2 flags:0x00004000
Call Trace:
context_switch kernel/sched/core.c:5369 [inline]
__schedule+0xe58/0x5ad0 kernel/sched/core.c:6756
__schedule_loop kernel/sched/core.c:6833 [inline]
schedule+0xe7/0x350 kernel/sched/core.c:6848
schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6905
__mutex_lock_common kernel/locking/mutex.c:665 [inline]
__mutex_lock+0x62b/0xa60 kernel/locking/mutex.c:735
worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
rescuer_thread+0x3bf/0xe20 kernel/workqueue.c:3471
kthread+0x2c1/0x3a0 kernel/kthread.c:389
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
INFO: task kworker/0:4:5865 blocked for more than 144 seconds.
Not tainted 6.13.0-rc1-syzkaller-00337-g7503345ac5f5 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/0:4 state:D stack:17856 pid:5865 tgid:5865 ppid:2 flags:0x00004000
Workqueue: 0x0 (events)
Call Trace:
context_switch kernel/sched/core.c:5369 [inline]
__schedule+0xe58/0x5ad0 kernel/sched/core.c:6756
__schedule_loop kernel/sched/core.c:6833 [inline]
schedule+0xe7/0x350 kernel/sched/core.c:6848
schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6905
__mutex_lock_common kernel/locking/mutex.c:665 [inline]
__mutex_lock+0x62b/0xa60 kernel/locking/mutex.c:735
worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
create_worker+0x2a5/0x800 kernel/workqueue.c:2811
maybe_create_worker kernel/workqueue.c:3054 [inline]
manage_workers kernel/workqueue.c:3106 [inline]
worker_thread+0x9c5/0xf00 kernel/workqueue.c:3366
kthread+0x2c1/0x3a0 kernel/kthread.c:389
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
INFO: task kworker/u8:8:5876 blocked for more than 145 seconds.
Not tainted 6.13.0-rc1-syzkaller-00337-g7503345ac5f5 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/u8:8 state:D stack:23504 pid:5876 tgid:5876 ppid:2 flags:0x00004000
Workqueue: netns cleanup_net
Call Trace:
context_switch kernel/sched/core.c:5369 [inline]
__schedule+0xe58/0x5ad0 kernel/sched/core.c:6756
__schedule_loop kernel/sched/core.c:6833 [inline]
schedule+0xe7/0x350 kernel/sched/core.c:6848
schedule_timeout+0x244/0x280 kernel/time/sleep_timeout.c:75
do_wait_for_common kernel/sched/completion.c:95 [inline]
__wait_for_common+0x3e1/0x600 kernel/sched/completion.c:116
kthread_stop+0x19a/0x660 kernel/kthread.c:712
destroy_workqueue+0x154/0xaa0 kernel/workqueue.c:5856
wg_destruct+0x24d/0x3d0 drivers/net/wireguard/device.c:257
netdev_run_todo+0x760/0x12d0 net/core/dev.c:10917
default_device_exit_batch+0x86c/0xae0 net/core/dev.c:12074
ops_exit_list+0x128/0x180 net/core/net_namespace.c:177
cleanup_net+0x5b7/0xb40 net/core/net_namespace.c:632
process_one_work+0x9c5/0x1ba0 kernel/workqueue.c:3229
process_scheduled_works kernel/workqueue.c:3310 [inline]
worker_thread+0x6c8/0xf00 kernel/workqueue.c:3391
kthread+0x2c1/0x3a0 kernel/kthread.c:389
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
INFO: task kworker/R-wg-cr:15177 blocked for more than 145 seconds.
Not tainted 6.13.0-rc1-syzkaller-00337-g7503345ac5f5 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/R-wg-cr state:D stack:30112 pid:15177 tgid:15177 ppid:2 flags:0x00004000
Call Trace:
context_switch kernel/sched/core.c:5369 [inline]
__schedule+0xe58/0x5ad0 kernel/sched/core.c:6756
__schedule_loop kernel/sched/core.c:6833 [inline]
schedule+0xe7/0x350 kernel/sched/core.c:6848
schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6905
__mutex_lock_common kernel/locking/mutex.c:665 [inline]
__mutex_lock+0x62b/0xa60 kernel/locking/mutex.c:735
worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
rescuer_thread+0x3bf/0xe20 kernel/workqueue.c:3471
kthread+0x2c1/0x3a0 kernel/kthread.c:389
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
INFO: task kworker/R-wg-cr:15199 blocked for more than 146 seconds.
Not tainted 6.13.0-rc1-syzkaller-00337-g7503345ac5f5 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/R-wg-cr state:D stack:30040 pid:15199 tgid:15199 ppid:2 flags:0x00024000
Call Trace:
context_switch kernel/sched/core.c:5369 [inline]
__schedule+0xe58/0x5ad0 kernel/sched/core.c:6756
__schedule_loop kernel/sched/core.c:6833 [inline]
schedule+0xe7/0x350 kernel/sched/core.c:6848
schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6905
__mutex_lock_common kernel/locking/mutex.c:665 [inline]
__mutex_lock+0x62b/0xa60 kernel/locking/mutex.c:735
set_pf_worker kernel/workqueue.c:3316 [inline]
rescuer_thread+0x9f7/0xe20 kernel/workqueue.c:3535
kthread+0x2c1/0x3a0 kernel/kthread.c:389
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
INFO: task kworker/R-wg-cr:15366 blocked for more than 146 seconds.
Not tainted 6.13.0-rc1-syzkaller-00337-g7503345ac5f5 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/R-wg-cr state:D stack:28968 pid:15366 tgid:15366 ppid:2 flags:0x00004000
Workqueue: 0x0 (wg-crypt-wg1)
Call Trace:
context_switch kernel/sched/core.c:5369 [inline]
__schedule+0xe58/0x5ad0 kernel/sched/core.c:6756
__schedule_loop kernel/sched/core.c:6833 [inline]
schedule+0xe7/0x350 kernel/sched/core.c:6848
schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6905
__mutex_lock_common kernel/locking/mutex.c:665 [inline]
__mutex_lock+0x62b/0xa60 kernel/locking/mutex.c:735
worker_detach_from_pool kernel/workqueue.c:2727 [inline]
rescuer_thread+0x856/0xe20 kernel/workqueue.c:3526
kthread+0x2c1/0x3a0 kernel/kthread.c:389
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
INFO: task kworker/R-wg-cr:15399 blocked for more than 147 seconds.
Not tainted 6.13.0-rc1-syzkaller-00337-g7503345ac5f5 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/R-wg-cr state:D stack:30112 pid:15399 tgid:15399 ppid:2 flags:0x00004000
Call Trace:
context_switch kernel/sched/core.c:5369 [inline]
__schedule+0xe58/0x5ad0 kernel/sched/core.c:6756
__schedule_loop kernel/sched/core.c:6833 [inline]
schedule+0xe7/0x350 kernel/sched/core.c:6848
schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6905
__mutex_lock_common kernel/locking/mutex.c:665 [inline]
__mutex_lock+0x62b/0xa60 kernel/locking/mutex.c:735
worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
rescuer_thread+0x3bf/0xe20 kernel/workqueue.c:3471
kthread+0x2c1/0x3a0 kernel/kthread.c:389
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
INFO: task kworker/0:0:15608 blocked for more than 147 seconds.
Not tainted 6.13.0-rc1-syzkaller-00337-g7503345ac5f5 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/0:0 state:D stack:30832 pid:15608 tgid:15608 ppid:2 flags:0x00004000
Call Trace:
context_switch kernel/sched/core.c:5369 [inline]
__schedule+0xe58/0x5ad0 kernel/sched/core.c:6756
__schedule_loop kernel/sched/core.c:6833 [inline]
schedule+0xe7/0x350 kernel/sched/core.c:6848
schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6905
kthread+0x1fe/0x3a0 kernel/kthread.c:382
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
Future hung task reports are suppressed, see sysctl kernel.hung_task_warnings
Showing all locks held in the system:
1 lock held by kworker/R-rcu_g/4:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
3 locks held by kworker/u8:1/12:
#0: ffff88801b081148 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_one_work+0x1293/0x1ba0 kernel/workqueue.c:3204
#1: ffffc90000117d80 ((linkwatch_work).work){+.+.}-{0:0}, at: process_one_work+0x921/0x1ba0 kernel/workqueue.c:3205
#2: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: linkwatch_event+0x51/0xc0 net/core/link_watch.c:281
1 lock held by kworker/R-mm_pe/13:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0x856/0xe20 kernel/workqueue.c:3526
3 locks held by ksoftirqd/0/16:
1 lock held by khungtaskd/30:
#0: ffffffff8e1bb500 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire include/linux/rcupdate.h:337 [inline]
#0: ffffffff8e1bb500 (rcu_read_lock){....}-{1:3}, at: rcu_read_lock include/linux/rcupdate.h:849 [inline]
#0: ffffffff8e1bb500 (rcu_read_lock){....}-{1:3}, at: debug_show_all_locks+0x7f/0x390 kernel/locking/lockdep.c:6744
1 lock held by kworker/R-write/32:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
2 locks held by ksmd/36:
#0: ffffffff8e374a48 (ksm_thread_mutex){+.+.}-{4:4}, at: ksm_scan_thread+0xdf/0x50e0 mm/ksm.c:2673
#1: ffffffff8e2ebca8 (lock#3){+.+.}-{4:4}, at: __lru_add_drain_all+0x69/0x6f0 mm/swap.c:798
1 lock held by kworker/R-krdsd/3381:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-bat_e/3392:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
2 locks held by getty/5572:
#0: ffff888031b750a0 (&tty->ldisc_sem){++++}-{0:0}, at: tty_ldisc_ref_wait+0x24/0x80 drivers/tty/tty_ldisc.c:243
#1: ffffc90002fde2f0 (&ldata->atomic_read_lock){+.+.}-{4:4}, at: n_tty_read+0xfba/0x1480 drivers/tty/n_tty.c:2211
1 lock held by kworker/R-bond0/5844:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/5847:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/5849:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/5853:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/5854:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0x856/0xe20 kernel/workqueue.c:3526
1 lock held by kworker/0:4/5865:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
3 locks held by kworker/0:5/5866:
4 locks held by kworker/u8:8/5876:
#0: ffff88801beeb148 ((wq_completion)netns){+.+.}-{0:0}, at: process_one_work+0x1293/0x1ba0 kernel/workqueue.c:3204
#1: ffffc90003317d80 (net_cleanup_work){+.+.}-{0:0}, at: process_one_work+0x921/0x1ba0 kernel/workqueue.c:3205
#2: ffffffff8fec4610 (pernet_ops_rwsem){++++}-{4:4}, at: cleanup_net+0xbb/0xb40 net/core/net_namespace.c:586
#3: ffff888045e8d4e8 (&wg->device_update_lock){+.+.}-{4:4}, at: wg_destruct+0x151/0x3d0 drivers/net/wireguard/device.c:249
1 lock held by kworker/u8:22/8358:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-bond1/8438:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
3 locks held by kworker/u8:25/11936:
#0: ffff88801b081148 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_one_work+0x1293/0x1ba0 kernel/workqueue.c:3204
#1: ffffc90003247d80 ((work_completion)(&pool->idle_cull_work)){+.+.}-{0:0}, at: process_one_work+0x921/0x1ba0 kernel/workqueue.c:3205
#2: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: idle_cull_fn+0x9a/0x450 kernel/workqueue.c:2951
3 locks held by kworker/u8:28/11939:
#0: ffff88814d334148 ((wq_completion)ipv6_addrconf){+.+.}-{0:0}, at: process_one_work+0x1293/0x1ba0 kernel/workqueue.c:3204
#1: ffffc90003687d80 ((work_completion)(&(&net->ipv6.addr_chk_work)->work)){+.+.}-{0:0}, at: process_one_work+0x921/0x1ba0 kernel/workqueue.c:3205
#2: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: addrconf_verify_work+0x12/0x30 net/ipv6/addrconf.c:4755
1 lock held by kworker/R-wg-cr/15175:
1 lock held by kworker/R-wg-cr/15177:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/15178:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/15199:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0x9f7/0xe20 kernel/workqueue.c:3535
1 lock held by kworker/R-wg-cr/15258:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0x856/0xe20 kernel/workqueue.c:3526
1 lock held by kworker/R-wg-cr/15259:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/15260:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/15365:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0x856/0xe20 kernel/workqueue.c:3526
1 lock held by kworker/R-wg-cr/15366:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0x856/0xe20 kernel/workqueue.c:3526
1 lock held by kworker/R-wg-cr/15367:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/15370:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0x856/0xe20 kernel/workqueue.c:3526
1 lock held by kworker/R-wg-cr/15371:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/15372:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/15389:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/15390:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/15391:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/15399:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/15401:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_attach_to_pool+0x27/0x420 kernel/workqueue.c:2669
1 lock held by kworker/R-wg-cr/15405:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0x856/0xe20 kernel/workqueue.c:3526
1 lock held by syz.9.2383/15651:
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: netdev_wait_allrefs_any net/core/dev.c:10786 [inline]
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: netdev_run_todo+0xdfc/0x12d0 net/core/dev.c:10905
1 lock held by kworker/R-bond0/15683:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-bond0/15685:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-bond0/15690:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15691:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15693:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-bond0/15694:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15696:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15698:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15699:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15700:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15701:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15702:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15703:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15704:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15705:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-bond0/15706:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15707:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15709:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15710:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15711:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
7 locks held by syz-executor/15768:
#0: ffff888035e96420 (sb_writers#9){.+.+}-{0:0}, at: ksys_write+0x12b/0x250 fs/read_write.c:731
#1: ffff888044f95488 (&of->mutex){+.+.}-{4:4}, at: kernfs_fop_write_iter+0x27b/0x500 fs/kernfs/file.c:325
#2: ffff888027eda0f8 (kn->active#49){.+.+}-{0:0}, at: kernfs_fop_write_iter+0x29e/0x500 fs/kernfs/file.c:326
#3: ffffffff8f483bc8 (nsim_bus_dev_list_lock){+.+.}-{4:4}, at: del_device_store+0xd2/0x4b0 drivers/net/netdevsim/bus.c:216
#4: ffff888143ecf0e8 (&dev->mutex){....}-{4:4}, at: device_lock include/linux/device.h:1014 [inline]
#4: ffff888143ecf0e8 (&dev->mutex){....}-{4:4}, at: __device_driver_lock drivers/base/dd.c:1095 [inline]
#4: ffff888143ecf0e8 (&dev->mutex){....}-{4:4}, at: device_release_driver_internal+0xa4/0x610 drivers/base/dd.c:1293
#5: ffff888144ab6250 (&devlink->lock_key#46){+.+.}-{4:4}, at: nsim_drv_remove+0x4a/0x1d0 drivers/net/netdevsim/dev.c:1675
#6: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: nsim_destroy+0x6f/0x6b0 drivers/net/netdevsim/netdev.c:816
1 lock held by kworker/R-bond0/15781:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15787:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15788:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by kworker/R-wg-cr/15789:
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: set_pf_worker kernel/workqueue.c:3316 [inline]
#0: ffffffff8e074e88 (wq_pool_attach_mutex){+.+.}-{4:4}, at: rescuer_thread+0xc6/0xe20 kernel/workqueue.c:3443
1 lock held by syz.7.2397/15816:
#0: ffffc900035290a8 (&kvm->slots_lock){+.+.}-{4:4}, at: kvm_set_memory_region virt/kvm/kvm_main.c:2066 [inline]
#0: ffffc900035290a8 (&kvm->slots_lock){+.+.}-{4:4}, at: kvm_vm_ioctl_set_memory_region virt/kvm/kvm_main.c:2079 [inline]
#0: ffffc900035290a8 (&kvm->slots_lock){+.+.}-{4:4}, at: kvm_vm_ioctl+0x1523/0x3df0 virt/kvm/kvm_main.c:5062
5 locks held by syz.8.2399/15821:
#0: ffff8880697289b8 (&f->f_pos_lock){+.+.}-{4:4}, at: fdget_pos+0x267/0x390 fs/file.c:1191
#1: ffff888035e96420 (sb_writers#9){.+.+}-{0:0}, at: ksys_write+0x12b/0x250 fs/read_write.c:731
#2: ffff88804b750888 (&of->mutex){+.+.}-{4:4}, at: kernfs_fop_write_iter+0x27b/0x500 fs/kernfs/file.c:325
#3: ffff88801fee2d28 (kn->active#60){.+.+}-{0:0}, at: kernfs_fop_write_iter+0x29e/0x500 fs/kernfs/file.c:326
#4: ffffffff8e374a48 (ksm_thread_mutex){+.+.}-{4:4}, at: run_store+0xe4/0x9b0 mm/ksm.c:3357
2 locks held by syz.6.2401/15831:
#0: ffff888051ec0e08 (&sb->s_type->i_mutex_key#10){+.+.}-{4:4}, at: inode_lock include/linux/fs.h:818 [inline]
#0: ffff888051ec0e08 (&sb->s_type->i_mutex_key#10){+.+.}-{4:4}, at: __sock_release+0x86/0x270 net/socket.c:639
#1: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: raw_release+0x1c6/0xad0 net/can/raw.c:415
1 lock held by syz.6.2401/15832:
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: rtnl_lock net/core/rtnetlink.c:79 [inline]
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: rtnetlink_rcv_msg+0x372/0xea0 net/core/rtnetlink.c:6925
1 lock held by syz.6.2401/15833:
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: dev_ioctl+0x213/0x10c0 net/core/dev_ioctl.c:782
1 lock held by syz.6.2401/15834:
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: rtnl_lock net/core/rtnetlink.c:79 [inline]
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: rtnl_net_lock include/linux/rtnetlink.h:128 [inline]
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: rtnl_dellink+0x281/0xaa0 net/core/rtnetlink.c:3510
2 locks held by syz.2.2403/15841:
#0: ffffffff8f3366c8 (ppp_mutex){+.+.}-{4:4}, at: ppp_ioctl+0xc3/0x2590 drivers/net/ppp/ppp_generic.c:729
#1: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: ppp_create_interface drivers/net/ppp/ppp_generic.c:3346 [inline]
#1: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: ppp_unattached_ioctl drivers/net/ppp/ppp_generic.c:1060 [inline]
#1: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: ppp_ioctl+0x17d8/0x2590 drivers/net/ppp/ppp_generic.c:733
1 lock held by syz.2.2403/15842:
#0: ffffffff8f3366c8 (ppp_mutex){+.+.}-{4:4}, at: ppp_ioctl+0xc3/0x2590 drivers/net/ppp/ppp_generic.c:729
1 lock held by syz.2.2403/15843:
#0: ffffffff8f3366c8 (ppp_mutex){+.+.}-{4:4}, at: ppp_ioctl+0xc3/0x2590 drivers/net/ppp/ppp_generic.c:729
1 lock held by syz-executor/15848:
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: rtnl_net_lock include/linux/rtnetlink.h:128 [inline]
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: inet_rtm_newaddr+0x316/0x1560 net/ipv4/devinet.c:987
1 lock held by syz-executor/15849:
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: rtnl_net_lock include/linux/rtnetlink.h:128 [inline]
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: inet_rtm_newaddr+0x316/0x1560 net/ipv4/devinet.c:987
1 lock held by syz-executor/15853:
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: rtnl_net_lock include/linux/rtnetlink.h:128 [inline]
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: inet_rtm_newaddr+0x316/0x1560 net/ipv4/devinet.c:987
1 lock held by syz-executor/15856:
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: rtnl_net_lock include/linux/rtnetlink.h:128 [inline]
#0: ffffffff8feda348 (rtnl_mutex){+.+.}-{4:4}, at: inet_rtm_newaddr+0x316/0x1560 net/ipv4/devinet.c:987
=============================================
NMI backtrace for cpu 1
CPU: 1 UID: 0 PID: 30 Comm: khungtaskd Not tainted 6.13.0-rc1-syzkaller-00337-g7503345ac5f5 #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 09/13/2024
Call Trace:
__dump_stack lib/dump_stack.c:94 [inline]
dump_stack_lvl+0x116/0x1f0 lib/dump_stack.c:120
nmi_cpu_backtrace+0x27b/0x390 lib/nmi_backtrace.c:113
nmi_trigger_cpumask_backtrace+0x29c/0x300 lib/nmi_backtrace.c:62
trigger_all_cpu_backtrace include/linux/nmi.h:162 [inline]
check_hung_uninterruptible_tasks kernel/hung_task.c:234 [inline]
watchdog+0xf14/0x1240 kernel/hung_task.c:397
kthread+0x2c1/0x3a0 kernel/kthread.c:389
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
Sending NMI from CPU 1 to CPUs 0:
NMI backtrace for cpu 0
CPU: 0 UID: 0 PID: 16 Comm: ksoftirqd/0 Not tainted 6.13.0-rc1-syzkaller-00337-g7503345ac5f5 #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 09/13/2024
RIP: 0010:asm_sysvec_apic_timer_interrupt+0x0/0x20 arch/x86/include/asm/idtentry.h:702
Code: a6 c8 e9 ff e9 71 06 00 00 90 f3 0f 1e fa 0f 01 ca fc 6a ff e8 21 05 00 00 48 89 c4 48 89 e7 e8 b6 c7 e9 ff e9 51 06 00 00 90 0f 1e fa 0f 01 ca fc 6a ff e8 01 05 00 00 48 89 c4 48 89 e7 e8
RSP: 0018:ffffc90000157b68 EFLAGS: 00000046
RAX: 0000000000000006 RBX: ffff8880b862ca00 RCX: 1ffffffff2dde358
RDX: 0000000000000000 RSI: ffffffff8b6cd840 RDI: ffffffff8bd1b560
RBP: 0000000000000246 R08: 0000000000000001 R09: fffffbfff2dca5ca
R10: ffffffff96e52e57 R11: 0000000000000003 R12: 0000000000000002
R13: ffff88803351f328 R14: 0000000000000000 R15: ffff8880b862ca80
FS: 0000000000000000(0000) GS:ffff8880b8600000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 0000001b2f817ff8 CR3: 000000000df7e000 CR4: 00000000003526f0
Call Trace:
RIP: 0010:__raw_spin_unlock_irqrestore include/linux/spinlock_api_smp.h:152 [inline]
RIP: 0010:_raw_spin_unlock_irqrestore+0x31/0x80 kernel/locking/spinlock.c:194
Code: f5 53 48 8b 74 24 10 48 89 fb 48 83 c7 18 e8 06 34 4a f6 48 89 df e8 4e b3 4a f6 f7 c5 00 02 00 00 75 23 9c 58 f6 c4 02 75 37 01 00 00 00 e8 d5 a7 3b f6 65 8b 05 c6 61 d7 74 85 c0 74 16 5b
RSP: 0018:ffffc90000157b90 EFLAGS: 00000246