All of lore.kernel.org
 help / color / mirror / Atom feed
* [syzbot] possible deadlock in hfsplus_block_allocate
@ 2022-11-29 13:38 syzbot
  2024-03-13 22:37 ` [syzbot] [hfs?] " syzbot
  2024-03-14  8:20 ` syzbot
  0 siblings, 2 replies; 3+ messages in thread
From: syzbot @ 2022-11-29 13:38 UTC (permalink / raw)
  To: akpm, fmdefrancesco, ira.weiny, linux-fsdevel, linux-kernel,
	slava, syzkaller-bugs

Hello,

syzbot found the following issue on:

HEAD commit:    6d464646530f Merge branch 'for-next/core' into for-kernelci
git tree:       git://git.kernel.org/pub/scm/linux/kernel/git/arm64/linux.git for-kernelci
console output: https://syzkaller.appspot.com/x/log.txt?x=12cf5fed880000
kernel config:  https://syzkaller.appspot.com/x/.config?x=54b747d981acc7b7
dashboard link: https://syzkaller.appspot.com/bug?extid=b6ccd31787585244a855
compiler:       Debian clang version 13.0.1-++20220126092033+75e33f71c2da-1~exp1~20220126212112.63, GNU ld (GNU Binutils for Debian) 2.35.2
userspace arch: arm64

Unfortunately, I don't have any reproducer for this issue yet.

Downloadable assets:
disk image: https://storage.googleapis.com/syzbot-assets/d75f5f77b3a3/disk-6d464646.raw.xz
vmlinux: https://storage.googleapis.com/syzbot-assets/9382f86e4d95/vmlinux-6d464646.xz
kernel image: https://storage.googleapis.com/syzbot-assets/cf2b5f0d51dd/Image-6d464646.gz.xz

IMPORTANT: if you fix the issue, please add the following tag to the commit:
Reported-by: syzbot+b6ccd31787585244a855@syzkaller.appspotmail.com

======================================================
WARNING: possible circular locking dependency detected
6.1.0-rc6-syzkaller-32662-g6d464646530f #0 Not tainted
------------------------------------------------------
syz-executor.1/11346 is trying to acquire lock:
ffff0000fce298f8 (&sbi->alloc_mutex){+.+.}-{3:3}, at: hfsplus_block_allocate+0x60/0x5e4 fs/hfsplus/bitmap.c:35

but task is already holding lock:
ffff0000fd681b08 (&HFSPLUS_I(inode)->extents_lock){+.+.}-{3:3}, at: hfsplus_file_extend+0x8c/0x88c fs/hfsplus/extents.c:457

which lock already depends on the new lock.


the existing dependency chain (in reverse order) is:

-> #1 (&HFSPLUS_I(inode)->extents_lock){+.+.}-{3:3}:
       __mutex_lock_common+0xd4/0xca8 kernel/locking/mutex.c:603
       __mutex_lock kernel/locking/mutex.c:747 [inline]
       mutex_lock_nested+0x38/0x44 kernel/locking/mutex.c:799
       hfsplus_get_block+0x17c/0x8a4 fs/hfsplus/extents.c:260
       block_read_full_folio+0x188/0x8c0 fs/buffer.c:2271
       hfsplus_read_folio+0x28/0x38 fs/hfsplus/inode.c:28
       filemap_read_folio+0xc4/0x468 mm/filemap.c:2407
       do_read_cache_folio+0x1c8/0x588 mm/filemap.c:3534
       do_read_cache_page mm/filemap.c:3576 [inline]
       read_cache_page+0x40/0x174 mm/filemap.c:3585
       read_mapping_page include/linux/pagemap.h:756 [inline]
       hfsplus_block_allocate+0x80/0x5e4 fs/hfsplus/bitmap.c:37
       hfsplus_file_extend+0x560/0x88c fs/hfsplus/extents.c:468
       hfsplus_get_block+0x1c0/0x8a4 fs/hfsplus/extents.c:245
       __block_write_begin_int+0x23c/0x9d4 fs/buffer.c:1991
       __block_write_begin fs/buffer.c:2041 [inline]
       block_write_begin+0x74/0x14c fs/buffer.c:2102
       cont_write_begin+0xf4/0x11c fs/buffer.c:2456
       hfsplus_write_begin+0x64/0xac fs/hfsplus/inode.c:52
       generic_perform_write+0xf0/0x2cc mm/filemap.c:3753
       __generic_file_write_iter+0xd8/0x21c mm/filemap.c:3881
       generic_file_write_iter+0x6c/0x168 mm/filemap.c:3913
       call_write_iter include/linux/fs.h:2191 [inline]
       new_sync_write fs/read_write.c:491 [inline]
       vfs_write+0x2dc/0x46c fs/read_write.c:584
       ksys_write+0xb4/0x160 fs/read_write.c:637
       __do_sys_write fs/read_write.c:649 [inline]
       __se_sys_write fs/read_write.c:646 [inline]
       __arm64_sys_write+0x24/0x34 fs/read_write.c:646
       __invoke_syscall arch/arm64/kernel/syscall.c:38 [inline]
       invoke_syscall arch/arm64/kernel/syscall.c:52 [inline]
       el0_svc_common+0x138/0x220 arch/arm64/kernel/syscall.c:142
       do_el0_svc+0x48/0x164 arch/arm64/kernel/syscall.c:206
       el0_svc+0x58/0x150 arch/arm64/kernel/entry-common.c:637
       el0t_64_sync_handler+0x84/0xf0 arch/arm64/kernel/entry-common.c:655
       el0t_64_sync+0x190/0x194 arch/arm64/kernel/entry.S:584

-> #0 (&sbi->alloc_mutex){+.+.}-{3:3}:
       check_prev_add kernel/locking/lockdep.c:3097 [inline]
       check_prevs_add kernel/locking/lockdep.c:3216 [inline]
       validate_chain kernel/locking/lockdep.c:3831 [inline]
       __lock_acquire+0x1530/0x3084 kernel/locking/lockdep.c:5055
       lock_acquire+0x100/0x1f8 kernel/locking/lockdep.c:5668
       __mutex_lock_common+0xd4/0xca8 kernel/locking/mutex.c:603
       __mutex_lock kernel/locking/mutex.c:747 [inline]
       mutex_lock_nested+0x38/0x44 kernel/locking/mutex.c:799
       hfsplus_block_allocate+0x60/0x5e4 fs/hfsplus/bitmap.c:35
       hfsplus_file_extend+0x560/0x88c fs/hfsplus/extents.c:468
       hfsplus_bmap_reserve+0x6c/0x120 fs/hfsplus/btree.c:358
       hfsplus_create_cat+0xa4/0x38c fs/hfsplus/catalog.c:272
       hfsplus_fill_super+0x758/0x864 fs/hfsplus/super.c:560
       mount_bdev+0x1b8/0x210 fs/super.c:1401
       hfsplus_mount+0x44/0x58 fs/hfsplus/super.c:641
       legacy_get_tree+0x30/0x74 fs/fs_context.c:610
       vfs_get_tree+0x40/0x140 fs/super.c:1531
       do_new_mount+0x1dc/0x4e4 fs/namespace.c:3040
       path_mount+0x358/0x890 fs/namespace.c:3370
       do_mount fs/namespace.c:3383 [inline]
       __do_sys_mount fs/namespace.c:3591 [inline]
       __se_sys_mount fs/namespace.c:3568 [inline]
       __arm64_sys_mount+0x2c4/0x3c4 fs/namespace.c:3568
       __invoke_syscall arch/arm64/kernel/syscall.c:38 [inline]
       invoke_syscall arch/arm64/kernel/syscall.c:52 [inline]
       el0_svc_common+0x138/0x220 arch/arm64/kernel/syscall.c:142
       do_el0_svc+0x48/0x164 arch/arm64/kernel/syscall.c:206
       el0_svc+0x58/0x150 arch/arm64/kernel/entry-common.c:637
       el0t_64_sync_handler+0x84/0xf0 arch/arm64/kernel/entry-common.c:655
       el0t_64_sync+0x190/0x194 arch/arm64/kernel/entry.S:584

other info that might help us debug this:

 Possible unsafe locking scenario:

       CPU0                    CPU1
       ----                    ----
  lock(&HFSPLUS_I(inode)->extents_lock);
                               lock(&sbi->alloc_mutex);
                               lock(&HFSPLUS_I(inode)->extents_lock);
  lock(&sbi->alloc_mutex);

 *** DEADLOCK ***

4 locks held by syz-executor.1/11346:
 #0: ffff0001056de0e0 (&type->s_umount_key#53/1){+.+.}-{3:3}, at: alloc_super+0xf8/0x430 fs/super.c:228
 #1: ffff0000fce29998 (&sbi->vh_mutex){+.+.}-{3:3}, at: hfsplus_fill_super+0x72c/0x864 fs/hfsplus/super.c:553
 #2: ffff0001056dc0b0 (&tree->tree_lock){+.+.}-{3:3}, at: hfsplus_find_init+0xac/0xc8
 #3: ffff0000fd681b08 (&HFSPLUS_I(inode)->extents_lock){+.+.}-{3:3}, at: hfsplus_file_extend+0x8c/0x88c fs/hfsplus/extents.c:457

stack backtrace:
CPU: 1 PID: 11346 Comm: syz-executor.1 Not tainted 6.1.0-rc6-syzkaller-32662-g6d464646530f #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 09/30/2022
Call trace:
 dump_backtrace+0x1c4/0x1f0 arch/arm64/kernel/stacktrace.c:156
 show_stack+0x2c/0x54 arch/arm64/kernel/stacktrace.c:163
 __dump_stack lib/dump_stack.c:88 [inline]
 dump_stack_lvl+0x104/0x16c lib/dump_stack.c:106
 dump_stack+0x1c/0x58 lib/dump_stack.c:113
 print_circular_bug+0x2c4/0x2c8 kernel/locking/lockdep.c:2055
 check_noncircular+0x14c/0x154 kernel/locking/lockdep.c:2177
 check_prev_add kernel/locking/lockdep.c:3097 [inline]
 check_prevs_add kernel/locking/lockdep.c:3216 [inline]
 validate_chain kernel/locking/lockdep.c:3831 [inline]
 __lock_acquire+0x1530/0x3084 kernel/locking/lockdep.c:5055
 lock_acquire+0x100/0x1f8 kernel/locking/lockdep.c:5668
 __mutex_lock_common+0xd4/0xca8 kernel/locking/mutex.c:603
 __mutex_lock kernel/locking/mutex.c:747 [inline]
 mutex_lock_nested+0x38/0x44 kernel/locking/mutex.c:799
 hfsplus_block_allocate+0x60/0x5e4 fs/hfsplus/bitmap.c:35
 hfsplus_file_extend+0x560/0x88c fs/hfsplus/extents.c:468
 hfsplus_bmap_reserve+0x6c/0x120 fs/hfsplus/btree.c:358
 hfsplus_create_cat+0xa4/0x38c fs/hfsplus/catalog.c:272
 hfsplus_fill_super+0x758/0x864 fs/hfsplus/super.c:560
 mount_bdev+0x1b8/0x210 fs/super.c:1401
 hfsplus_mount+0x44/0x58 fs/hfsplus/super.c:641
 legacy_get_tree+0x30/0x74 fs/fs_context.c:610
 vfs_get_tree+0x40/0x140 fs/super.c:1531
 do_new_mount+0x1dc/0x4e4 fs/namespace.c:3040
 path_mount+0x358/0x890 fs/namespace.c:3370
 do_mount fs/namespace.c:3383 [inline]
 __do_sys_mount fs/namespace.c:3591 [inline]
 __se_sys_mount fs/namespace.c:3568 [inline]
 __arm64_sys_mount+0x2c4/0x3c4 fs/namespace.c:3568
 __invoke_syscall arch/arm64/kernel/syscall.c:38 [inline]
 invoke_syscall arch/arm64/kernel/syscall.c:52 [inline]
 el0_svc_common+0x138/0x220 arch/arm64/kernel/syscall.c:142
 do_el0_svc+0x48/0x164 arch/arm64/kernel/syscall.c:206
 el0_svc+0x58/0x150 arch/arm64/kernel/entry-common.c:637
 el0t_64_sync_handler+0x84/0xf0 arch/arm64/kernel/entry-common.c:655
 el0t_64_sync+0x190/0x194 arch/arm64/kernel/entry.S:584
hfsplus: b-tree write err: -5, ino 4


---
This report is generated by a bot. It may contain errors.
See https://goo.gl/tpsmEJ for more information about syzbot.
syzbot engineers can be reached at syzkaller@googlegroups.com.

syzbot will keep track of this issue. See:
https://goo.gl/tpsmEJ#status for how to communicate with syzbot.

^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: [syzbot] [hfs?] possible deadlock in hfsplus_block_allocate
  2022-11-29 13:38 [syzbot] possible deadlock in hfsplus_block_allocate syzbot
@ 2024-03-13 22:37 ` syzbot
  2024-03-14  8:20 ` syzbot
  1 sibling, 0 replies; 3+ messages in thread
From: syzbot @ 2024-03-13 22:37 UTC (permalink / raw)
  To: akpm, fmdefrancesco, ira.weiny, linux-fsdevel, linux-kernel,
	slava, syzkaller-bugs

syzbot has found a reproducer for the following issue on:

HEAD commit:    61387b8dcf1d Merge tag 'for-6.9/dm-vdo' of git://git.kerne..
git tree:       upstream
console output: https://syzkaller.appspot.com/x/log.txt?x=126d69d1180000
kernel config:  https://syzkaller.appspot.com/x/.config?x=e30ff6b515606856
dashboard link: https://syzkaller.appspot.com/bug?extid=b6ccd31787585244a855
compiler:       Debian clang version 15.0.6, GNU ld (GNU Binutils for Debian) 2.40
syz repro:      https://syzkaller.appspot.com/x/repro.syz?x=129e58c9180000
C reproducer:   https://syzkaller.appspot.com/x/repro.c?x=154b15d6180000

Downloadable assets:
disk image: https://storage.googleapis.com/syzbot-assets/cab7c2663886/disk-61387b8d.raw.xz
vmlinux: https://storage.googleapis.com/syzbot-assets/5b9dd4f348b5/vmlinux-61387b8d.xz
kernel image: https://storage.googleapis.com/syzbot-assets/222aa1f905fd/bzImage-61387b8d.xz
mounted in repro #1: https://storage.googleapis.com/syzbot-assets/fead63abbb06/mount_0.gz
mounted in repro #2: https://storage.googleapis.com/syzbot-assets/f46392d41e04/mount_3.gz

IMPORTANT: if you fix the issue, please add the following tag to the commit:
Reported-by: syzbot+b6ccd31787585244a855@syzkaller.appspotmail.com

loop0: detected capacity change from 0 to 1024
======================================================
WARNING: possible circular locking dependency detected
6.8.0-syzkaller-05562-g61387b8dcf1d #0 Not tainted
------------------------------------------------------
syz-executor170/6261 is trying to acquire lock:
ffff88807f2b10f8 (&sbi->alloc_mutex){+.+.}-{3:3}, at: hfsplus_block_allocate+0x9e/0x8c0 fs/hfsplus/bitmap.c:35

but task is already holding lock:
ffff88807f19df88 (&HFSPLUS_I(inode)->extents_lock){+.+.}-{3:3}, at: hfsplus_file_extend+0x21b/0x1b70 fs/hfsplus/extents.c:457

which lock already depends on the new lock.


the existing dependency chain (in reverse order) is:

-> #1 (&HFSPLUS_I(inode)->extents_lock){+.+.}-{3:3}:
       lock_acquire+0x1e4/0x530 kernel/locking/lockdep.c:5754
       __mutex_lock_common kernel/locking/mutex.c:608 [inline]
       __mutex_lock+0x136/0xd70 kernel/locking/mutex.c:752
       hfsplus_get_block+0x383/0x14f0 fs/hfsplus/extents.c:260
       block_read_full_folio+0x42e/0xe10 fs/buffer.c:2384
       filemap_read_folio+0x1a0/0x790 mm/filemap.c:2324
       do_read_cache_folio+0x134/0x820 mm/filemap.c:3694
       do_read_cache_page+0x30/0x200 mm/filemap.c:3760
       read_mapping_page include/linux/pagemap.h:888 [inline]
       hfsplus_block_allocate+0xee/0x8c0 fs/hfsplus/bitmap.c:37
       hfsplus_file_extend+0xade/0x1b70 fs/hfsplus/extents.c:468
       hfsplus_get_block+0x406/0x14f0 fs/hfsplus/extents.c:245
       __block_write_begin_int+0x50c/0x1a70 fs/buffer.c:2105
       __block_write_begin fs/buffer.c:2154 [inline]
       block_write_begin+0x9b/0x1e0 fs/buffer.c:2213
       cont_write_begin+0x645/0x890 fs/buffer.c:2567
       hfsplus_write_begin+0x8a/0xd0 fs/hfsplus/inode.c:47
       cont_expand_zero fs/buffer.c:2527 [inline]
       cont_write_begin+0x6ee/0x890 fs/buffer.c:2557
       hfsplus_write_begin+0x8a/0xd0 fs/hfsplus/inode.c:47
       generic_cont_expand_simple+0x18f/0x2b0 fs/buffer.c:2458
       hfsplus_setattr+0x178/0x280 fs/hfsplus/inode.c:259
       notify_change+0xb9d/0xe70 fs/attr.c:497
       do_truncate fs/open.c:65 [inline]
       do_ftruncate+0x46b/0x590 fs/open.c:181
       do_sys_ftruncate fs/open.c:199 [inline]
       __do_sys_ftruncate fs/open.c:207 [inline]
       __se_sys_ftruncate fs/open.c:205 [inline]
       __x64_sys_ftruncate+0x95/0xf0 fs/open.c:205
       do_syscall_64+0xfb/0x240
       entry_SYSCALL_64_after_hwframe+0x6d/0x75

-> #0 (&sbi->alloc_mutex){+.+.}-{3:3}:
       check_prev_add kernel/locking/lockdep.c:3134 [inline]
       check_prevs_add kernel/locking/lockdep.c:3253 [inline]
       validate_chain+0x18cb/0x58e0 kernel/locking/lockdep.c:3869
       __lock_acquire+0x1346/0x1fd0 kernel/locking/lockdep.c:5137
       lock_acquire+0x1e4/0x530 kernel/locking/lockdep.c:5754
       __mutex_lock_common kernel/locking/mutex.c:608 [inline]
       __mutex_lock+0x136/0xd70 kernel/locking/mutex.c:752
       hfsplus_block_allocate+0x9e/0x8c0 fs/hfsplus/bitmap.c:35
       hfsplus_file_extend+0xade/0x1b70 fs/hfsplus/extents.c:468
       hfsplus_get_block+0x406/0x14f0 fs/hfsplus/extents.c:245
       __block_write_begin_int+0x50c/0x1a70 fs/buffer.c:2105
       __block_write_begin fs/buffer.c:2154 [inline]
       block_write_begin+0x9b/0x1e0 fs/buffer.c:2213
       cont_write_begin+0x645/0x890 fs/buffer.c:2567
       hfsplus_write_begin+0x8a/0xd0 fs/hfsplus/inode.c:47
       cont_expand_zero fs/buffer.c:2494 [inline]
       cont_write_begin+0x319/0x890 fs/buffer.c:2557
       hfsplus_write_begin+0x8a/0xd0 fs/hfsplus/inode.c:47
       generic_perform_write+0x322/0x640 mm/filemap.c:3921
       generic_file_write_iter+0xaf/0x310 mm/filemap.c:4042
       do_iter_readv_writev+0x5a4/0x800
       vfs_writev+0x395/0xbb0 fs/read_write.c:971
       do_pwritev fs/read_write.c:1072 [inline]
       __do_sys_pwritev2 fs/read_write.c:1131 [inline]
       __se_sys_pwritev2+0x1ca/0x2d0 fs/read_write.c:1122
       do_syscall_64+0xfb/0x240
       entry_SYSCALL_64_after_hwframe+0x6d/0x75

other info that might help us debug this:

 Possible unsafe locking scenario:

       CPU0                    CPU1
       ----                    ----
  lock(&HFSPLUS_I(inode)->extents_lock);
                               lock(&sbi->alloc_mutex);
                               lock(&HFSPLUS_I(inode)->extents_lock);
  lock(&sbi->alloc_mutex);

 *** DEADLOCK ***

3 locks held by syz-executor170/6261:
 #0: ffff88806d1ea420 (sb_writers#11){.+.+}-{0:0}, at: file_start_write include/linux/fs.h:2853 [inline]
 #0: ffff88806d1ea420 (sb_writers#11){.+.+}-{0:0}, at: vfs_writev+0x2d9/0xbb0 fs/read_write.c:969
 #1: ffff88807f19e180 (&sb->s_type->i_mutex_key#17){+.+.}-{3:3}, at: inode_lock include/linux/fs.h:793 [inline]
 #1: ffff88807f19e180 (&sb->s_type->i_mutex_key#17){+.+.}-{3:3}, at: generic_file_write_iter+0x83/0x310 mm/filemap.c:4039
 #2: ffff88807f19df88 (&HFSPLUS_I(inode)->extents_lock){+.+.}-{3:3}, at: hfsplus_file_extend+0x21b/0x1b70 fs/hfsplus/extents.c:457

stack backtrace:
CPU: 1 PID: 6261 Comm: syz-executor170 Not tainted 6.8.0-syzkaller-05562-g61387b8dcf1d #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 02/29/2024
Call Trace:
 <TASK>
 __dump_stack lib/dump_stack.c:88 [inline]
 dump_stack_lvl+0x241/0x360 lib/dump_stack.c:114
 check_noncircular+0x36a/0x4a0 kernel/locking/lockdep.c:2187
 check_prev_add kernel/locking/lockdep.c:3134 [inline]
 check_prevs_add kernel/locking/lockdep.c:3253 [inline]
 validate_chain+0x18cb/0x58e0 kernel/locking/lockdep.c:3869
 __lock_acquire+0x1346/0x1fd0 kernel/locking/lockdep.c:5137
 lock_acquire+0x1e4/0x530 kernel/locking/lockdep.c:5754
 __mutex_lock_common kernel/locking/mutex.c:608 [inline]
 __mutex_lock+0x136/0xd70 kernel/locking/mutex.c:752
 hfsplus_block_allocate+0x9e/0x8c0 fs/hfsplus/bitmap.c:35
 hfsplus_file_extend+0xade/0x1b70 fs/hfsplus/extents.c:468
 hfsplus_get_block+0x406/0x14f0 fs/hfsplus/extents.c:245
 __block_write_begin_int+0x50c/0x1a70 fs/buffer.c:2105
 __block_write_begin fs/buffer.c:2154 [inline]
 block_write_begin+0x9b/0x1e0 fs/buffer.c:2213
 cont_write_begin+0x645/0x890 fs/buffer.c:2567
 hfsplus_write_begin+0x8a/0xd0 fs/hfsplus/inode.c:47
 cont_expand_zero fs/buffer.c:2494 [inline]
 cont_write_begin+0x319/0x890 fs/buffer.c:2557
 hfsplus_write_begin+0x8a/0xd0 fs/hfsplus/inode.c:47
 generic_perform_write+0x322/0x640 mm/filemap.c:3921
 generic_file_write_iter+0xaf/0x310 mm/filemap.c:4042
 do_iter_readv_writev+0x5a4/0x800
 vfs_writev+0x395/0xbb0 fs/read_write.c:971
 do_pwritev fs/read_write.c:1072 [inline]
 __do_sys_pwritev2 fs/read_write.c:1131 [inline]
 __se_sys_pwritev2+0x1ca/0x2d0 fs/read_write.c:1122
 do_syscall_64+0xfb/0x240
 entry_SYSCALL_64_after_hwframe+0x6d/0x75
RIP: 0033:0x7fe7ed3c9299
Code: 28 00 00 00 75 05 48 83 c4 28 c3 e8 b1 1f 00 00 90 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 4d 89 c8 4c 8b 4c 24 08 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 c7 c1 b8 ff ff ff f7 d8 64 89 01 48
RSP: 002b:00007ffd9249b938 EFLAGS: 00000216 ORIG_RAX: 0000000000000148
RAX: ffffffffffffffda RBX: 0000000000000003 RCX: 00007fe7ed3c9299
RDX: 0000000000000001 RSI: 00000000200001c0 RDI: 0000000000000005
RBP: 0000000000000000 R08: 0000000000000000 R09: 0000000000000000
R10: 0000000000007fff R11: 0000000000000216 R12: 00007ffd9249b978
R13: 0000000000000473 R14: 00007ffd9249b9b0 R15: 431bde82d7b634db
 </TASK>


---
If you want syzbot to run the reproducer, reply with:
#syz test: git://repo/address.git branch-or-commit-hash
If you attach or paste a git patch, syzbot will apply it before testing.

^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: [syzbot] [hfs?] possible deadlock in hfsplus_block_allocate
  2022-11-29 13:38 [syzbot] possible deadlock in hfsplus_block_allocate syzbot
  2024-03-13 22:37 ` [syzbot] [hfs?] " syzbot
@ 2024-03-14  8:20 ` syzbot
  1 sibling, 0 replies; 3+ messages in thread
From: syzbot @ 2024-03-14  8:20 UTC (permalink / raw)
  To: akpm, fmdefrancesco, ira.weiny, linux-fsdevel, linux-kernel,
	pbonzini, seanjc, slava, syzkaller-bugs, yongwei.ma

syzbot has bisected this issue to:

commit 427c76aed29ec12a57f11546a5036df3cc52855e
Author: Sean Christopherson <seanjc@google.com>
Date:   Sat Jul 29 01:35:32 2023 +0000

    KVM: x86/mmu: Bug the VM if write-tracking is used but not enabled

bisection log:  https://syzkaller.appspot.com/x/bisect.txt?x=1779a8c9180000
start commit:   61387b8dcf1d Merge tag 'for-6.9/dm-vdo' of git://git.kerne..
git tree:       upstream
final oops:     https://syzkaller.appspot.com/x/report.txt?x=14f9a8c9180000
console output: https://syzkaller.appspot.com/x/log.txt?x=10f9a8c9180000
kernel config:  https://syzkaller.appspot.com/x/.config?x=e30ff6b515606856
dashboard link: https://syzkaller.appspot.com/bug?extid=b6ccd31787585244a855
syz repro:      https://syzkaller.appspot.com/x/repro.syz?x=129e58c9180000
C reproducer:   https://syzkaller.appspot.com/x/repro.c?x=154b15d6180000

Reported-by: syzbot+b6ccd31787585244a855@syzkaller.appspotmail.com
Fixes: 427c76aed29e ("KVM: x86/mmu: Bug the VM if write-tracking is used but not enabled")

For information about bisection process see: https://goo.gl/tpsmEJ#bisection

^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2024-03-14  8:20 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-11-29 13:38 [syzbot] possible deadlock in hfsplus_block_allocate syzbot
2024-03-13 22:37 ` [syzbot] [hfs?] " syzbot
2024-03-14  8:20 ` syzbot

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.