INFO: task syz.2.1243:10013 blocked for more than 143 seconds. Not tainted 6.16.0-rc5-next-20250711-syzkaller #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:syz.2.1243 state:D stack:25096 pid:10013 tgid:10013 ppid:5848 task_flags:0x400040 flags:0x00004006 Call Trace: context_switch kernel/sched/core.c:5314 [inline] __schedule+0x16f5/0x4d00 kernel/sched/core.c:6697 __schedule_loop kernel/sched/core.c:6775 [inline] schedule+0x165/0x360 kernel/sched/core.c:6790 request_wait_answer fs/fuse/dev.c:566 [inline] __fuse_request_send fs/fuse/dev.c:580 [inline] __fuse_simple_request+0xf15/0x18d0 fs/fuse/dev.c:674 fuse_simple_request fs/fuse/fuse_i.h:1230 [inline] fuse_flush+0x5d4/0x800 fs/fuse/file.c:475 filp_flush+0xbd/0x190 fs/open.c:1547 filp_close+0x1d/0x40 fs/open.c:1560 __range_close fs/file.c:767 [inline] __do_sys_close_range fs/file.c:826 [inline] __se_sys_close_range+0x359/0x650 fs/file.c:790 do_syscall_x64 arch/x86/entry/syscall_64.c:63 [inline] do_syscall_64+0xfa/0x3b0 arch/x86/entry/syscall_64.c:94 entry_SYSCALL_64_after_hwframe+0x77/0x7f RIP: 0033:0x7f18fdf8e929 RSP: 002b:00007ffe682d51b8 EFLAGS: 00000246 ORIG_RAX: 00000000000001b4 RAX: ffffffffffffffda RBX: 00007f18fe1b7ba0 RCX: 00007f18fdf8e929 RDX: 0000000000000000 RSI: 000000000000001e RDI: 0000000000000003 RBP: 00007f18fe1b7ba0 R08: 00000000000000d0 R09: 0000000d682d54af R10: 00007f18fe1b7ac0 R11: 0000000000000246 R12: 00000000000462e4 R13: 00007f18fe1b6080 R14: ffffffffffffffff R15: 00007ffe682d52d0 INFO: task syz.2.1243:10014 blocked for more than 143 seconds. Not tainted 6.16.0-rc5-next-20250711-syzkaller #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:syz.2.1243 state:D stack:25864 pid:10014 tgid:10013 ppid:5848 task_flags:0x400140 flags:0x00004004 Call Trace: context_switch kernel/sched/core.c:5314 [inline] __schedule+0x16f5/0x4d00 kernel/sched/core.c:6697 __schedule_loop kernel/sched/core.c:6775 [inline] schedule+0x165/0x360 kernel/sched/core.c:6790 schedule_timeout+0x9a/0x270 kernel/time/sleep_timeout.c:75 do_wait_for_common kernel/sched/completion.c:100 [inline] __wait_for_common kernel/sched/completion.c:121 [inline] wait_for_common kernel/sched/completion.c:132 [inline] wait_for_completion+0x2bf/0x5d0 kernel/sched/completion.c:153 fuse_direct_IO+0x9d6/0x1210 fs/fuse/file.c:2866 fuse_direct_write_iter fs/fuse/file.c:1687 [inline] fuse_file_write_iter+0x7b6/0x10a0 fs/fuse/file.c:1736 aio_write+0x535/0x7a0 fs/aio.c:1634 __io_submit_one fs/aio.c:-1 [inline] io_submit_one+0x78b/0x1310 fs/aio.c:2053 __do_sys_io_submit fs/aio.c:2112 [inline] __se_sys_io_submit+0x185/0x2f0 fs/aio.c:2082 do_syscall_x64 arch/x86/entry/syscall_64.c:63 [inline] do_syscall_64+0xfa/0x3b0 arch/x86/entry/syscall_64.c:94 entry_SYSCALL_64_after_hwframe+0x77/0x7f RIP: 0033:0x7f18fdf8e929 RSP: 002b:00007f18fed7f038 EFLAGS: 00000246 ORIG_RAX: 00000000000000d1 RAX: ffffffffffffffda RBX: 00007f18fe1b5fa0 RCX: 00007f18fdf8e929 RDX: 0000200000000780 RSI: 0000000000000001 RDI: 00007f18fed35000 RBP: 00007f18fe010b39 R08: 0000000000000000 R09: 0000000000000000 R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000000 R13: 0000000000000000 R14: 00007f18fe1b5fa0 R15: 00007ffe682d5058 Showing all locks held in the system: 3 locks held by kworker/1:0/24: #0: ffff88801a880d48 ((wq_completion)events){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3214 [inline] #0: ffff88801a880d48 ((wq_completion)events){+.+.}-{0:0}, at: process_scheduled_works+0x9b4/0x17b0 kernel/workqueue.c:3322 #1: ffffc900001e7bc0 (free_ipc_work){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3215 [inline] #1: ffffc900001e7bc0 (free_ipc_work){+.+.}-{0:0}, at: process_scheduled_works+0x9ef/0x17b0 kernel/workqueue.c:3322 #2: ffffffff8e5420f8 (rcu_state.exp_mutex){+.+.}-{4:4}, at: exp_funnel_lock kernel/rcu/tree_exp.h:343 [inline] #2: ffffffff8e5420f8 (rcu_state.exp_mutex){+.+.}-{4:4}, at: synchronize_rcu_expedited+0x3b9/0x730 kernel/rcu/tree_exp.h:967 1 lock held by khungtaskd/31: #0: ffffffff8e53c5a0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire include/linux/rcupdate.h:331 [inline] #0: ffffffff8e53c5a0 (rcu_read_lock){....}-{1:3}, at: rcu_read_lock include/linux/rcupdate.h:841 [inline] #0: ffffffff8e53c5a0 (rcu_read_lock){....}-{1:3}, at: debug_show_all_locks+0x2e/0x180 kernel/locking/lockdep.c:6770 3 locks held by kworker/1:2/984: 2 locks held by getty/5609: #0: ffff88814d3150a0 (&tty->ldisc_sem){++++}-{0:0}, at: tty_ldisc_ref_wait+0x25/0x70 drivers/tty/tty_ldisc.c:243 #1: ffffc9000332e2f0 (&ldata->atomic_read_lock){+.+.}-{4:4}, at: n_tty_read+0x43e/0x1400 drivers/tty/n_tty.c:2222 6 locks held by kworker/u8:8/7345: #0: ffff88801b6fe948 ((wq_completion)netns){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3214 [inline] #0: ffff88801b6fe948 ((wq_completion)netns){+.+.}-{0:0}, at: process_scheduled_works+0x9b4/0x17b0 kernel/workqueue.c:3322 #1: ffffc9001866fbc0 (net_cleanup_work){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3215 [inline] #1: ffffc9001866fbc0 (net_cleanup_work){+.+.}-{0:0}, at: process_scheduled_works+0x9ef/0x17b0 kernel/workqueue.c:3322 #2: ffffffff8f93ce90 (pernet_ops_rwsem){++++}-{4:4}, at: cleanup_net+0xf7/0x800 net/core/net_namespace.c:658 #3: ffff888061ac80e8 (&dev->mutex){....}-{4:4}, at: device_lock include/linux/device.h:911 [inline] #3: ffff888061ac80e8 (&dev->mutex){....}-{4:4}, at: devl_dev_lock net/devlink/devl_internal.h:108 [inline] #3: ffff888061ac80e8 (&dev->mutex){....}-{4:4}, at: devlink_pernet_pre_exit+0x10a/0x3d0 net/devlink/core.c:506 #4: ffff888056025250 (&devlink->lock_key#4){+.+.}-{4:4}, at: devl_lock net/devlink/core.c:276 [inline] #4: ffff888056025250 (&devlink->lock_key#4){+.+.}-{4:4}, at: devl_dev_lock net/devlink/devl_internal.h:109 [inline] #4: ffff888056025250 (&devlink->lock_key#4){+.+.}-{4:4}, at: devlink_pernet_pre_exit+0x11c/0x3d0 net/devlink/core.c:506 #5: ffffffff8e5420f8 (rcu_state.exp_mutex){+.+.}-{4:4}, at: exp_funnel_lock kernel/rcu/tree_exp.h:343 [inline] #5: ffffffff8e5420f8 (rcu_state.exp_mutex){+.+.}-{4:4}, at: synchronize_rcu_expedited+0x3b9/0x730 kernel/rcu/tree_exp.h:967 1 lock held by syz.2.1243/10014: #0: ffff88805e0cea48 (&sb->s_type->i_mutex_key#20){+.+.}-{4:4}, at: inode_lock include/linux/fs.h:869 [inline] #0: ffff88805e0cea48 (&sb->s_type->i_mutex_key#20){+.+.}-{4:4}, at: fuse_dio_lock fs/fuse/file.c:-1 [inline] #0: ffff88805e0cea48 (&sb->s_type->i_mutex_key#20){+.+.}-{4:4}, at: fuse_direct_write_iter fs/fuse/file.c:1682 [inline] #0: ffff88805e0cea48 (&sb->s_type->i_mutex_key#20){+.+.}-{4:4}, at: fuse_file_write_iter+0x72a/0x10a0 fs/fuse/file.c:1736 1 lock held by syz-executor/13134: 1 lock held by syz.4.2223/13587: 1 lock held by dhcpcd-run-hook/13625: ============================================= NMI backtrace for cpu 0 CPU: 0 UID: 0 PID: 31 Comm: khungtaskd Not tainted 6.16.0-rc5-next-20250711-syzkaller #0 PREEMPT(full) Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 05/07/2025 Call Trace: dump_stack_lvl+0x189/0x250 lib/dump_stack.c:120 nmi_cpu_backtrace+0x39e/0x3d0 lib/nmi_backtrace.c:113 nmi_trigger_cpumask_backtrace+0x17a/0x300 lib/nmi_backtrace.c:62 trigger_all_cpu_backtrace include/linux/nmi.h:160 [inline] check_hung_uninterruptible_tasks kernel/hung_task.c:328 [inline] watchdog+0xfee/0x1030 kernel/hung_task.c:491 kthread+0x711/0x8a0 kernel/kthread.c:463 ret_from_fork+0x3f9/0x770 arch/x86/kernel/process.c:148 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:245 Sending NMI from CPU 0 to CPUs 1: NMI backtrace for cpu 1 CPU: 1 UID: 0 PID: 24 Comm: kworker/1:0 Not tainted 6.16.0-rc5-next-20250711-syzkaller #0 PREEMPT(full) Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 05/07/2025 Workqueue: events drain_vmap_area_work RIP: 0010:unwind_next_frame+0x1721/0x2390 arch/x86/kernel/unwind_orc.c:648 Code: 18 02 00 00 80 3d 1e 5c 7c 11 00 0f 85 0b 02 00 00 c6 05 11 5c 7c 11 01 4c 89 f7 e8 89 0c 00 00 e9 f7 01 00 00 48 8b 54 24 08 <48> 8d 5a 02 48 83 c2 03 48 89 d8 48 c1 e8 03 0f b6 04 28 84 c0 0f RSP: 0018:ffffc900001e7298 EFLAGS: 00000246 RAX: 0000000000000000 RBX: ffffc900001e7400 RCX: 0000000000000000 RDX: ffffffff907583fc RSI: 0000000000000001 RDI: ffffc900001e7428 RBP: dffffc0000000000 R08: ffffc900001e7427 R09: 0000000000000000 R10: ffffc900001e7418 R11: fffff5200003ce85 R12: ffffc900001e73c8 R13: ffffc900001e7418 R14: ffffc900001e73c8 R15: 1ffffffff20eb080 FS: 0000000000000000(0000) GS:ffff88812588f000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 00007fa3f3343286 CR3: 0000000031e6a000 CR4: 00000000003526f0 Call Trace: __unwind_start+0x5b9/0x760 arch/x86/kernel/unwind_orc.c:758 unwind_start arch/x86/include/asm/unwind.h:64 [inline] arch_stack_walk+0xe4/0x150 arch/x86/kernel/stacktrace.c:24 stack_trace_save+0x9c/0xe0 kernel/stacktrace.c:122 save_stack+0xf5/0x1f0 mm/page_owner.c:156 __reset_page_owner+0x71/0x1f0 mm/page_owner.c:308 reset_page_owner include/linux/page_owner.h:25 [inline] free_pages_prepare mm/page_alloc.c:1395 [inline] __free_frozen_pages+0xbc4/0xd30 mm/page_alloc.c:2895 kasan_depopulate_vmalloc_pte+0x74/0xa0 mm/kasan/shadow.c:472 apply_to_pte_range mm/memory.c:3028 [inline] apply_to_pmd_range mm/memory.c:3072 [inline] apply_to_pud_range mm/memory.c:3108 [inline] apply_to_p4d_range mm/memory.c:3144 [inline] __apply_to_page_range+0xb8f/0x1380 mm/memory.c:3180 kasan_release_vmalloc+0xa2/0xd0 mm/kasan/shadow.c:593 kasan_release_vmalloc_node mm/vmalloc.c:2249 [inline] purge_vmap_node+0x214/0x8f0 mm/vmalloc.c:2266 __purge_vmap_area_lazy+0x7a4/0xb40 mm/vmalloc.c:2356 drain_vmap_area_work+0x27/0x40 mm/vmalloc.c:2390 process_one_work kernel/workqueue.c:3239 [inline] process_scheduled_works+0xade/0x17b0 kernel/workqueue.c:3322 worker_thread+0x8a0/0xda0 kernel/workqueue.c:3403 kthread+0x711/0x8a0 kernel/kthread.c:463 ret_from_fork+0x3f9/0x770 arch/x86/kernel/process.c:148 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:245