INFO: task syz.2.1243:10013 blocked for more than 143 seconds.
Not tainted 6.16.0-rc5-next-20250711-syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:syz.2.1243 state:D stack:25096 pid:10013 tgid:10013 ppid:5848 task_flags:0x400040 flags:0x00004006
Call Trace:
context_switch kernel/sched/core.c:5314 [inline]
__schedule+0x16f5/0x4d00 kernel/sched/core.c:6697
__schedule_loop kernel/sched/core.c:6775 [inline]
schedule+0x165/0x360 kernel/sched/core.c:6790
request_wait_answer fs/fuse/dev.c:566 [inline]
__fuse_request_send fs/fuse/dev.c:580 [inline]
__fuse_simple_request+0xf15/0x18d0 fs/fuse/dev.c:674
fuse_simple_request fs/fuse/fuse_i.h:1230 [inline]
fuse_flush+0x5d4/0x800 fs/fuse/file.c:475
filp_flush+0xbd/0x190 fs/open.c:1547
filp_close+0x1d/0x40 fs/open.c:1560
__range_close fs/file.c:767 [inline]
__do_sys_close_range fs/file.c:826 [inline]
__se_sys_close_range+0x359/0x650 fs/file.c:790
do_syscall_x64 arch/x86/entry/syscall_64.c:63 [inline]
do_syscall_64+0xfa/0x3b0 arch/x86/entry/syscall_64.c:94
entry_SYSCALL_64_after_hwframe+0x77/0x7f
RIP: 0033:0x7f18fdf8e929
RSP: 002b:00007ffe682d51b8 EFLAGS: 00000246 ORIG_RAX: 00000000000001b4
RAX: ffffffffffffffda RBX: 00007f18fe1b7ba0 RCX: 00007f18fdf8e929
RDX: 0000000000000000 RSI: 000000000000001e RDI: 0000000000000003
RBP: 00007f18fe1b7ba0 R08: 00000000000000d0 R09: 0000000d682d54af
R10: 00007f18fe1b7ac0 R11: 0000000000000246 R12: 00000000000462e4
R13: 00007f18fe1b6080 R14: ffffffffffffffff R15: 00007ffe682d52d0
INFO: task syz.2.1243:10014 blocked for more than 143 seconds.
Not tainted 6.16.0-rc5-next-20250711-syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:syz.2.1243 state:D stack:25864 pid:10014 tgid:10013 ppid:5848 task_flags:0x400140 flags:0x00004004
Call Trace:
context_switch kernel/sched/core.c:5314 [inline]
__schedule+0x16f5/0x4d00 kernel/sched/core.c:6697
__schedule_loop kernel/sched/core.c:6775 [inline]
schedule+0x165/0x360 kernel/sched/core.c:6790
schedule_timeout+0x9a/0x270 kernel/time/sleep_timeout.c:75
do_wait_for_common kernel/sched/completion.c:100 [inline]
__wait_for_common kernel/sched/completion.c:121 [inline]
wait_for_common kernel/sched/completion.c:132 [inline]
wait_for_completion+0x2bf/0x5d0 kernel/sched/completion.c:153
fuse_direct_IO+0x9d6/0x1210 fs/fuse/file.c:2866
fuse_direct_write_iter fs/fuse/file.c:1687 [inline]
fuse_file_write_iter+0x7b6/0x10a0 fs/fuse/file.c:1736
aio_write+0x535/0x7a0 fs/aio.c:1634
__io_submit_one fs/aio.c:-1 [inline]
io_submit_one+0x78b/0x1310 fs/aio.c:2053
__do_sys_io_submit fs/aio.c:2112 [inline]
__se_sys_io_submit+0x185/0x2f0 fs/aio.c:2082
do_syscall_x64 arch/x86/entry/syscall_64.c:63 [inline]
do_syscall_64+0xfa/0x3b0 arch/x86/entry/syscall_64.c:94
entry_SYSCALL_64_after_hwframe+0x77/0x7f
RIP: 0033:0x7f18fdf8e929
RSP: 002b:00007f18fed7f038 EFLAGS: 00000246 ORIG_RAX: 00000000000000d1
RAX: ffffffffffffffda RBX: 00007f18fe1b5fa0 RCX: 00007f18fdf8e929
RDX: 0000200000000780 RSI: 0000000000000001 RDI: 00007f18fed35000
RBP: 00007f18fe010b39 R08: 0000000000000000 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000000
R13: 0000000000000000 R14: 00007f18fe1b5fa0 R15: 00007ffe682d5058
Showing all locks held in the system:
3 locks held by kworker/1:0/24:
#0: ffff88801a880d48 ((wq_completion)events){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3214 [inline]
#0: ffff88801a880d48 ((wq_completion)events){+.+.}-{0:0}, at: process_scheduled_works+0x9b4/0x17b0 kernel/workqueue.c:3322
#1: ffffc900001e7bc0 (free_ipc_work){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3215 [inline]
#1: ffffc900001e7bc0 (free_ipc_work){+.+.}-{0:0}, at: process_scheduled_works+0x9ef/0x17b0 kernel/workqueue.c:3322
#2: ffffffff8e5420f8 (rcu_state.exp_mutex){+.+.}-{4:4}, at: exp_funnel_lock kernel/rcu/tree_exp.h:343 [inline]
#2: ffffffff8e5420f8 (rcu_state.exp_mutex){+.+.}-{4:4}, at: synchronize_rcu_expedited+0x3b9/0x730 kernel/rcu/tree_exp.h:967
1 lock held by khungtaskd/31:
#0: ffffffff8e53c5a0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire include/linux/rcupdate.h:331 [inline]
#0: ffffffff8e53c5a0 (rcu_read_lock){....}-{1:3}, at: rcu_read_lock include/linux/rcupdate.h:841 [inline]
#0: ffffffff8e53c5a0 (rcu_read_lock){....}-{1:3}, at: debug_show_all_locks+0x2e/0x180 kernel/locking/lockdep.c:6770
3 locks held by kworker/1:2/984:
2 locks held by getty/5609:
#0: ffff88814d3150a0 (&tty->ldisc_sem){++++}-{0:0}, at: tty_ldisc_ref_wait+0x25/0x70 drivers/tty/tty_ldisc.c:243
#1: ffffc9000332e2f0 (&ldata->atomic_read_lock){+.+.}-{4:4}, at: n_tty_read+0x43e/0x1400 drivers/tty/n_tty.c:2222
6 locks held by kworker/u8:8/7345:
#0: ffff88801b6fe948 ((wq_completion)netns){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3214 [inline]
#0: ffff88801b6fe948 ((wq_completion)netns){+.+.}-{0:0}, at: process_scheduled_works+0x9b4/0x17b0 kernel/workqueue.c:3322
#1: ffffc9001866fbc0 (net_cleanup_work){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3215 [inline]
#1: ffffc9001866fbc0 (net_cleanup_work){+.+.}-{0:0}, at: process_scheduled_works+0x9ef/0x17b0 kernel/workqueue.c:3322
#2: ffffffff8f93ce90 (pernet_ops_rwsem){++++}-{4:4}, at: cleanup_net+0xf7/0x800 net/core/net_namespace.c:658
#3: ffff888061ac80e8 (&dev->mutex){....}-{4:4}, at: device_lock include/linux/device.h:911 [inline]
#3: ffff888061ac80e8 (&dev->mutex){....}-{4:4}, at: devl_dev_lock net/devlink/devl_internal.h:108 [inline]
#3: ffff888061ac80e8 (&dev->mutex){....}-{4:4}, at: devlink_pernet_pre_exit+0x10a/0x3d0 net/devlink/core.c:506
#4: ffff888056025250 (&devlink->lock_key#4){+.+.}-{4:4}, at: devl_lock net/devlink/core.c:276 [inline]
#4: ffff888056025250 (&devlink->lock_key#4){+.+.}-{4:4}, at: devl_dev_lock net/devlink/devl_internal.h:109 [inline]
#4: ffff888056025250 (&devlink->lock_key#4){+.+.}-{4:4}, at: devlink_pernet_pre_exit+0x11c/0x3d0 net/devlink/core.c:506
#5: ffffffff8e5420f8 (rcu_state.exp_mutex){+.+.}-{4:4}, at: exp_funnel_lock kernel/rcu/tree_exp.h:343 [inline]
#5: ffffffff8e5420f8 (rcu_state.exp_mutex){+.+.}-{4:4}, at: synchronize_rcu_expedited+0x3b9/0x730 kernel/rcu/tree_exp.h:967
1 lock held by syz.2.1243/10014:
#0: ffff88805e0cea48 (&sb->s_type->i_mutex_key#20){+.+.}-{4:4}, at: inode_lock include/linux/fs.h:869 [inline]
#0: ffff88805e0cea48 (&sb->s_type->i_mutex_key#20){+.+.}-{4:4}, at: fuse_dio_lock fs/fuse/file.c:-1 [inline]
#0: ffff88805e0cea48 (&sb->s_type->i_mutex_key#20){+.+.}-{4:4}, at: fuse_direct_write_iter fs/fuse/file.c:1682 [inline]
#0: ffff88805e0cea48 (&sb->s_type->i_mutex_key#20){+.+.}-{4:4}, at: fuse_file_write_iter+0x72a/0x10a0 fs/fuse/file.c:1736
1 lock held by syz-executor/13134:
1 lock held by syz.4.2223/13587:
1 lock held by dhcpcd-run-hook/13625:
=============================================
NMI backtrace for cpu 0
CPU: 0 UID: 0 PID: 31 Comm: khungtaskd Not tainted 6.16.0-rc5-next-20250711-syzkaller #0 PREEMPT(full)
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 05/07/2025
Call Trace:
dump_stack_lvl+0x189/0x250 lib/dump_stack.c:120
nmi_cpu_backtrace+0x39e/0x3d0 lib/nmi_backtrace.c:113
nmi_trigger_cpumask_backtrace+0x17a/0x300 lib/nmi_backtrace.c:62
trigger_all_cpu_backtrace include/linux/nmi.h:160 [inline]
check_hung_uninterruptible_tasks kernel/hung_task.c:328 [inline]
watchdog+0xfee/0x1030 kernel/hung_task.c:491
kthread+0x711/0x8a0 kernel/kthread.c:463
ret_from_fork+0x3f9/0x770 arch/x86/kernel/process.c:148
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:245
Sending NMI from CPU 0 to CPUs 1:
NMI backtrace for cpu 1
CPU: 1 UID: 0 PID: 24 Comm: kworker/1:0 Not tainted 6.16.0-rc5-next-20250711-syzkaller #0 PREEMPT(full)
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 05/07/2025
Workqueue: events drain_vmap_area_work
RIP: 0010:unwind_next_frame+0x1721/0x2390 arch/x86/kernel/unwind_orc.c:648
Code: 18 02 00 00 80 3d 1e 5c 7c 11 00 0f 85 0b 02 00 00 c6 05 11 5c 7c 11 01 4c 89 f7 e8 89 0c 00 00 e9 f7 01 00 00 48 8b 54 24 08 <48> 8d 5a 02 48 83 c2 03 48 89 d8 48 c1 e8 03 0f b6 04 28 84 c0 0f
RSP: 0018:ffffc900001e7298 EFLAGS: 00000246
RAX: 0000000000000000 RBX: ffffc900001e7400 RCX: 0000000000000000
RDX: ffffffff907583fc RSI: 0000000000000001 RDI: ffffc900001e7428
RBP: dffffc0000000000 R08: ffffc900001e7427 R09: 0000000000000000
R10: ffffc900001e7418 R11: fffff5200003ce85 R12: ffffc900001e73c8
R13: ffffc900001e7418 R14: ffffc900001e73c8 R15: 1ffffffff20eb080
FS: 0000000000000000(0000) GS:ffff88812588f000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 00007fa3f3343286 CR3: 0000000031e6a000 CR4: 00000000003526f0
Call Trace:
__unwind_start+0x5b9/0x760 arch/x86/kernel/unwind_orc.c:758
unwind_start arch/x86/include/asm/unwind.h:64 [inline]
arch_stack_walk+0xe4/0x150 arch/x86/kernel/stacktrace.c:24
stack_trace_save+0x9c/0xe0 kernel/stacktrace.c:122
save_stack+0xf5/0x1f0 mm/page_owner.c:156
__reset_page_owner+0x71/0x1f0 mm/page_owner.c:308
reset_page_owner include/linux/page_owner.h:25 [inline]
free_pages_prepare mm/page_alloc.c:1395 [inline]
__free_frozen_pages+0xbc4/0xd30 mm/page_alloc.c:2895
kasan_depopulate_vmalloc_pte+0x74/0xa0 mm/kasan/shadow.c:472
apply_to_pte_range mm/memory.c:3028 [inline]
apply_to_pmd_range mm/memory.c:3072 [inline]
apply_to_pud_range mm/memory.c:3108 [inline]
apply_to_p4d_range mm/memory.c:3144 [inline]
__apply_to_page_range+0xb8f/0x1380 mm/memory.c:3180
kasan_release_vmalloc+0xa2/0xd0 mm/kasan/shadow.c:593
kasan_release_vmalloc_node mm/vmalloc.c:2249 [inline]
purge_vmap_node+0x214/0x8f0 mm/vmalloc.c:2266
__purge_vmap_area_lazy+0x7a4/0xb40 mm/vmalloc.c:2356
drain_vmap_area_work+0x27/0x40 mm/vmalloc.c:2390
process_one_work kernel/workqueue.c:3239 [inline]
process_scheduled_works+0xade/0x17b0 kernel/workqueue.c:3322
worker_thread+0x8a0/0xda0 kernel/workqueue.c:3403
kthread+0x711/0x8a0 kernel/kthread.c:463
ret_from_fork+0x3f9/0x770 arch/x86/kernel/process.c:148
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:245