[SRU][N/O][PATCH 0/1] raid1: Fix NULL pointer dereference in process_checks()

Matthew Ruffell matthew.ruffell at canonical.com
Mon Jun 9 05:57:01 UTC 2025


BugLink: https://bugs.launchpad.net/bugs/2112519

[Impact]

A null pointer dereference was found in raid1 during failure mode testing.
A raid1 array was set up, filled with data and a check operation started. While
the check was underway, all underlying iSCSI disks were forcefully disconnected
with --failfast set to the md array, and the following kernel oops occurs:

md/raid1:: dm-0: unrecoverable I/O read error for block 527744
md/raid1:: dm-1: unrecoverable I/O read error for block 527616
md/raid1:: dm-0: unrecoverable I/O read error for block 527744
md/raid1:: dm-1: unrecoverable I/O read error for block 527616
md/raid1:: dm-1: unrecoverable I/O read error for block 527616
md/raid1:: dm-0: unrecoverable I/O read error for block 527744
md/raid1:: dm-1: unrecoverable I/O read error for block 527616
md/raid1:: dm-0: unrecoverable I/O read error for block 527744
BUG: kernel NULL pointer dereference, address: 0000000000000040
#PF: supervisor read access in kernel mode
#PF: error_code(0x0000) - not-present page
PGD 0 P4D 0 
SMP NOPTI
CPU: 3 PID: 19372 Comm: md_1t889zmbfni_ Kdump: loaded Not tainted 6.8.0-1029-aws #31-Ubuntu
Hardware name: Amazon EC2 m6a.xlarge/, BIOS 1.0 10/16/2017
RIP: 0010:process_checks+0x25e/0x5e0 [raid1]
Code: 8e 19 01 00 00 48 8b 85 78 ff ff ff b9 08 00 00 00 48 8d 7d 90 49 8b 1c c4 49 63 c7 4d 8b 74 c4 50 31 c0 f3 48 ab 48 89 5d 88 <4c> 8b 53 40 45 0f b6 4e 18 49 8b 76 40 49 81 7e 38 a0 04 7c c0 75
RSP: 0018:ffffb39e8142bcb8 EFLAGS: 00010246
RAX: 0000000000000000 RBX: 0000000000000000 RCX: 0000000000000000
RDX: 0000000000000002 RSI: 0000000000000004 RDI: ffffb39e8142bd50
RBP: ffffb39e8142bd80 R08: ffff9a2e001ea000 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000000 R12: ffff9a2e0cd63280
R13: ffff9a2e50d1f800 R14: ffff9a2e50d1f000 R15: 0000000000000000
FS:  0000000000000000(0000) GS:ffff9a3128780000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 0000000000000040 CR3: 00000001035b2004 CR4: 00000000003706f0
Call Trace:
 <TASK>
 ? show_regs+0x6d/0x80
 ? __die+0x24/0x80
 ? page_fault_oops+0x99/0x1b0
 ? do_user_addr_fault+0x2e0/0x660
 ? exc_page_fault+0x83/0x190
 ? asm_exc_page_fault+0x27/0x30
 ? process_checks+0x25e/0x5e0 [raid1]
 ? process_checks+0x125/0x5e0 [raid1]
 ? srso_alias_return_thunk+0x5/0xfbef5
 ? ___ratelimit+0xc7/0x130
 sync_request_write+0x1c8/0x1e0 [raid1]
 raid1d+0x13a/0x3f0 [raid1]
 ? srso_alias_return_thunk+0x5/0xfbef5
 md_thread+0xae/0x190
 ? __pfx_autoremove_wake_function+0x10/0x10
 ? __pfx_md_thread+0x10/0x10
 kthread+0xda/0x100
 ? __pfx_kthread+0x10/0x10
 ret_from_fork+0x47/0x70
 ? __pfx_kthread+0x10/0x10
 ret_from_fork_asm+0x1b/0x30
 </TASK>

What happens is that process_checks() loops through all the available disks to
find a primary source with intact data, all disks are missing, and we shouldn't
move forward without having a valid primary source.

[Fix]

This was fixed in 6.15-rc3 with:

commit b7c178d9e57c8fd4238ff77263b877f6f16182ba
Author: Meir Elisha <meir.elisha at volumez.com>
Date:  Tue Apr 8 17:38:08 2025 +0300
Subject: md/raid1: Add check for missing source disk in process_checks()
Link: https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/commit/?id=b7c178d9e57c8fd4238ff77263b877f6f16182ba

This has been applied to focal, jammy and plucky already through upstream 
-stable. Currently noble and oracular are lagging behind and are not up to the
-stable release with the fix.

Bug focal:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2111448
Bug jammy:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2111606
Bug plucky:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2111268

[Testcase]

You don't need to set up a full iscsi environment, you can just make some local
VMs and then forcefully remove the underlying disks using libvirt.

Create a VM, attach two scratch disks:

$ lsblk
NAME    MAJ:MIN RM  SIZE RO TYPE MOUNTPOINTS
vda     253:0    0   10G  0 disk 
├─vda1  253:1    0    9G  0 part /
├─vda14 253:14   0    4M  0 part 
├─vda15 253:15   0  106M  0 part /boot/efi
└─vda16 259:0    0  913M  0 part /boot
vdb     253:16   0  372K  0 disk 
vdc     253:32   0    3G  0 disk 
vdd     253:48   0    3G  0 disk 
vde     253:64   0    3G  0 disk 

Create a raid1 array:

$ sudo mdadm --create --failfast --verbose /dev/md0 --level=1 --raid-devices=3 /dev/vdc /dev/vdd /dev/vde

Make a filesystem:

$ sudo mkfs.xfs /dev/md0

$ sudo mkdir /mnt/disk
$ sudo mount /dev/md0 /mnt/disk

Fill scratch disks with files:

for n in {1..1000}; do     dd if=/dev/urandom of=file$( printf %03d "$n" ).bin bs=1024 count=$(( RANDOM)); done

Start a check:

$ sudo mdadm --action=check /dev/md0

Use virt manager / libvirt to detach the disks, watch dmesg.

Test kernels are available in the following ppa:

https://launchpad.net/~mruffell/+archive/ubuntu/sf411666-test

If you install the test kernel, the null pointer dereference no longer occurs.

[Where problems can occur]

We are changing the logic such that if all the reads fail in process_check(),
and we have no valid primary source, then we disable recovery mode, mark an
error occurring, free the bio and exit out. Previously we would have just
continued onward and run into the null pointer dereference.

This really only affects situations where all backing disks are lost. This isn't
too uncommon though, particularly if all are network storage and network issues
occur, losing access to the disks. Things should remain as they are if at least
one primary source disk exists.

If a regression were to occur, it would affect raid1 arrays only, and only
during check/repair operations. 

A workaround would be to disable check or repair operations on the md array 
until the issue is fixed.

[Other info]

Upstream mailing list discussion:

V1:
https://lore.kernel.org/linux-raid/712ff6db-6b01-be95-a394-266be08a1d6e@huaweicloud.com/T/
V2:
https://lore.kernel.org/linux-raid/20250408143808.1026534-1-meir.elisha@volumez.com/T/

Meir Elisha (1):
  md/raid1: Add check for missing source disk in process_checks()

 drivers/md/raid1.c | 26 ++++++++++++++++----------
 1 file changed, 16 insertions(+), 10 deletions(-)

-- 
2.48.1




More information about the kernel-team mailing list