INFO: task kworker/1:18:5431 blocked for more than 143 seconds.
Not tainted syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/1:18 state:D stack:24976 pid:5431 ppid:2 flags:0x00004000
Workqueue: gfs_recovery gfs2_recover_func
Call Trace:
context_switch kernel/sched/core.c:5244 [inline]
__schedule+0x10ec/0x40b0 kernel/sched/core.c:6561
schedule+0xb9/0x180 kernel/sched/core.c:6637
io_schedule+0x7c/0xd0 kernel/sched/core.c:8797
folio_wait_bit_common+0x6e1/0xf60 mm/filemap.c:1324
folio_wait_locked include/linux/pagemap.h:1057 [inline]
gfs2_jhead_process_page+0x176/0x7d0 fs/gfs2/lops.c:476
gfs2_find_jhead+0xba2/0xd00 fs/gfs2/lops.c:594
gfs2_recover_func+0x666/0x1c10 fs/gfs2/recovery.c:460
process_one_work+0x898/0x1160 kernel/workqueue.c:2292
worker_thread+0xaa2/0x1250 kernel/workqueue.c:2439
kthread+0x29d/0x330 kernel/kthread.c:376
ret_from_fork+0x1f/0x30 arch/x86/entry/entry_64.S:295
INFO: task syz.3.252:5505 blocked for more than 144 seconds.
Not tainted syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:syz.3.252 state:D stack:24736 pid:5505 ppid:4274 flags:0x00004004
Call Trace:
context_switch kernel/sched/core.c:5244 [inline]
__schedule+0x10ec/0x40b0 kernel/sched/core.c:6561
schedule+0xb9/0x180 kernel/sched/core.c:6637
bit_wait+0xd/0xc0 kernel/sched/wait_bit.c:199
__wait_on_bit+0xa8/0x2d0 kernel/sched/wait_bit.c:49
out_of_line_wait_on_bit+0x11f/0x160 kernel/sched/wait_bit.c:64
wait_on_bit include/linux/wait_bit.h:76 [inline]
gfs2_recover_journal+0xd6/0x130 fs/gfs2/recovery.c:577
init_journal+0x17ea/0x2300 fs/gfs2/ops_fstype.c:835
init_inodes+0xdb/0x320 fs/gfs2/ops_fstype.c:889
gfs2_fill_super+0x1718/0x1f80 fs/gfs2/ops_fstype.c:1246
get_tree_bdev+0x3f1/0x610 fs/super.c:1366
gfs2_get_tree+0x4d/0x1e0 fs/gfs2/ops_fstype.c:1327
vfs_get_tree+0x88/0x270 fs/super.c:1573
do_new_mount+0x24a/0xa40 fs/namespace.c:3069
do_mount fs/namespace.c:3412 [inline]
__do_sys_mount fs/namespace.c:3620 [inline]
__se_sys_mount+0x2d6/0x3c0 fs/namespace.c:3597
do_syscall_x64 arch/x86/entry/common.c:51 [inline]
do_syscall_64+0x4c/0xa0 arch/x86/entry/common.c:81
entry_SYSCALL_64_after_hwframe+0x68/0xd2
RIP: 0033:0x7f9cae19034a
RSP: 002b:00007f9caeff1e68 EFLAGS: 00000246 ORIG_RAX: 00000000000000a5
RAX: ffffffffffffffda RBX: 00007f9caeff1ef0 RCX: 00007f9cae19034a
RDX: 0000200000037f40 RSI: 0000200000037f80 RDI: 00007f9caeff1eb0
RBP: 0000200000037f40 R08: 00007f9caeff1ef0 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000246 R12: 0000200000037f80
R13: 00007f9caeff1eb0 R14: 000000000003800c R15: 0000200000000000
Showing all locks held in the system:
1 lock held by rcu_tasks_kthre/12:
#0: ffffffff8cb2b570 (rcu_tasks.tasks_gp_mutex){+.+.}-{3:3}, at: rcu_tasks_one_gp+0x33/0xf00 kernel/rcu/tasks.h:517
1 lock held by rcu_tasks_trace/13:
#0: ffffffff8cb2bd90 (rcu_tasks_trace.tasks_gp_mutex){+.+.}-{3:3}, at: rcu_tasks_one_gp+0x33/0xf00 kernel/rcu/tasks.h:517
1 lock held by khungtaskd/28:
#0: ffffffff8cb2abe0 (rcu_read_lock){....}-{1:2}, at: rcu_lock_acquire include/linux/rcupdate.h:350 [inline]
#0: ffffffff8cb2abe0 (rcu_read_lock){....}-{1:2}, at: rcu_read_lock include/linux/rcupdate.h:791 [inline]
#0: ffffffff8cb2abe0 (rcu_read_lock){....}-{1:2}, at: debug_show_all_locks+0x51/0x290 kernel/locking/lockdep.c:6513
2 locks held by getty/4027:
#0: ffff88814d37c098 (&tty->ldisc_sem){++++}-{0:0}, at: tty_ldisc_ref_wait+0x21/0x70 drivers/tty/tty_ldisc.c:244
#1: ffffc9000327b2f0 (&ldata->atomic_read_lock){+.+.}-{3:3}, at: n_tty_read+0x41b/0x1380 drivers/tty/n_tty.c:2198
1 lock held by udevd/4258:
#0: ffff8880b8e3aad8 (&rq->__lock){-.-.}-{2:2}, at: raw_spin_rq_lock_nested+0xa5/0x140 kernel/sched/core.c:545
2 locks held by kworker/0:5/4328:
#0: ffff888017472138 ((wq_completion)rcu_gp){+.+.}-{0:0}, at: process_one_work+0x7a1/0x1160 kernel/workqueue.c:2267
#1: ffffc90004137d00 ((work_completion)(&rew->rew_work)){+.+.}-{0:0}, at: process_one_work+0x7a1/0x1160 kernel/workqueue.c:2267
2 locks held by kworker/1:18/5431:
#0: ffff88801b3a3d38 ((wq_completion)gfs_recovery){+.+.}-{0:0}, at: process_one_work+0x7a1/0x1160 kernel/workqueue.c:2267
#1: ffffc90003937d00 ((work_completion)(&jd->jd_work)){+.+.}-{0:0}, at: process_one_work+0x7a1/0x1160 kernel/workqueue.c:2267
1 lock held by syz.3.252/5505:
#0: ffff8880565800e0 (&type->s_umount_key#81/1){+.+.}-{3:3}, at: alloc_super+0x1fa/0x930 fs/super.c:228
2 locks held by syz.7.599/7230:
#0: ffffffff8c9e9188 (sched_core_mutex){+.+.}-{3:3}, at: sched_core_get+0x49/0x1c0 kernel/sched/core.c:398
#1: ffffffff8cb308b8 (rcu_state.exp_mutex){+.+.}-{3:3}, at: exp_funnel_lock kernel/rcu/tree_exp.h:291 [inline]
#1: ffffffff8cb308b8 (rcu_state.exp_mutex){+.+.}-{3:3}, at: synchronize_rcu_expedited+0x455/0x830 kernel/rcu/tree_exp.h:962
3 locks held by syz.5.603/7212:
1 lock held by rm/7235:
=============================================
NMI backtrace for cpu 1
CPU: 1 PID: 28 Comm: khungtaskd Not tainted syzkaller #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 08/18/2025
Call Trace:
dump_stack_lvl+0x168/0x22e lib/dump_stack.c:106
nmi_cpu_backtrace+0x3f4/0x470 lib/nmi_backtrace.c:111
nmi_trigger_cpumask_backtrace+0x1d4/0x450 lib/nmi_backtrace.c:62
trigger_all_cpu_backtrace include/linux/nmi.h:148 [inline]
check_hung_uninterruptible_tasks kernel/hung_task.c:220 [inline]
watchdog+0xeee/0xf30 kernel/hung_task.c:377
kthread+0x29d/0x330 kernel/kthread.c:376
ret_from_fork+0x1f/0x30 arch/x86/entry/entry_64.S:295
Sending NMI from CPU 1 to CPUs 0:
NMI backtrace for cpu 0
CPU: 0 PID: 7235 Comm: rm Not tainted syzkaller #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 08/18/2025
RIP: 0010:check_preemption_disabled+0x2e/0x110 lib/smp_processor_id.c:55
Code: 56 53 48 83 ec 10 65 48 8b 04 25 28 00 00 00 48 89 44 24 08 65 8b 05 a1 c1 ea 75 65 8b 0d 42 17 eb 75 f7 c1 ff ff ff 7f 74 1f <65> 48 8b 0c 25 28 00 00 00 48 3b 4c 24 08 0f 85 c4 00 00 00 48 83
RSP: 0000:ffffc900056f77d0 EFLAGS: 00000046
RAX: 0000000000000000 RBX: ffffffff81c69177 RCX: 0000000080000000
RDX: 0000000000000000 RSI: ffffffff8a8c0520 RDI: ffffffff8adefca0
RBP: ffffc900056f78f8 R08: ffff88813fffc310 R09: ffffed1027fff598
R10: ffffed1027fff598 R11: 1ffff11027fff597 R12: dffffc0000000000
R13: 0000000000000a02 R14: dffffc0000000000 R15: 1ffff92000adef08
FS: 00007f3f7871cc80(0000) GS:ffff8880b8e00000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 00007f3f787ba1d4 CR3: 00000000426b0000 CR4: 00000000003506f0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
Call Trace:
lockdep_hardirqs_off+0xa7/0x100 kernel/locking/lockdep.c:4400
trace_hardirqs_off+0xe/0x40 kernel/trace/trace_preemptirq.c:76
seqcount_lockdep_reader_access+0xc7/0x170 include/linux/seqlock.h:101
read_seqbegin include/linux/seqlock.h:893 [inline]
zone_span_seqbegin include/linux/memory_hotplug.h:132 [inline]
page_outside_zone_boundaries mm/page_alloc.c:627 [inline]
bad_range+0x7a/0x2d0 mm/page_alloc.c:654
rmqueue mm/page_alloc.c:3887 [inline]
get_page_from_freelist+0x1a0a/0x1ac0 mm/page_alloc.c:4325
__alloc_pages+0x1df/0x4e0 mm/page_alloc.c:5614
__folio_alloc+0xe/0x30 mm/page_alloc.c:5646
vma_alloc_folio+0x4a3/0x900 mm/mempolicy.c:2243
alloc_page_vma include/linux/gfp.h:284 [inline]
do_anonymous_page mm/memory.c:4191 [inline]
handle_pte_fault mm/memory.c:5029 [inline]
__handle_mm_fault mm/memory.c:5173 [inline]
handle_mm_fault+0x2242/0x3e70 mm/memory.c:5294
do_user_addr_fault+0x51f/0xb10 arch/x86/mm/fault.c:1340
handle_page_fault arch/x86/mm/fault.c:1431 [inline]
exc_page_fault+0x60/0x100 arch/x86/mm/fault.c:1487
asm_exc_page_fault+0x22/0x30 arch/x86/include/asm/idtentry.h:608
RIP: 0033:0x7f3f78795328
Code: 53 48 83 ec 38 64 48 8b 04 25 28 00 00 00 48 89 44 24 28 31 c0 e8 f8 f8 ff ff 48 8d 15 b5 4e 02 00 48 c7 44 24 18 00 00 00 00 <89> 02 48 8b 05 07 3c 02 00 48 83 38 00 74 41 31 f6 48 8d 3d e6 bf
RSP: 002b:00007fff68a25890 EFLAGS: 00010202
RAX: 0000000000001000 RBX: 00007f3f787b84d0 RCX: 00007f3f787b8590
RDX: 00007f3f787ba1d4 RSI: 00007fff68a25978 RDI: 000000000000001e
RBP: 0000000000000003 R08: 00007f3f78abcaf8 R09: 0000000000000000
R10: 00007fff68a255a0 R11: 0000000000000246 R12: 00007fff68a25978
R13: 00007fff68a25998 R14: 00007f3f787b84d0 R15: 0000000000000000
vkms_vblank_simulate: vblank timer overrun