[Bug 2102236] Re: plucky/s390x (with root disk on dm?) does not come up after kernel update + reboot
Frank Heimes
2102236 at bugs.launchpad.net
Tue Mar 18 12:54:21 UTC 2025
I asked the kernel team already to have a look at this as well, when I
marked it as affecting 'linux (Ubuntu)', it's suspicious.
And no, the disk nodes are missing in /dev.
It's a multipath environment with two HBAs, each two paths,
so there need to be four SCSI devices (sda, sdb, sdc and sdd),
but there are none (being in busybox):
ls -la /dev/sd*
ls: /dev/sd*: No such file or directory
ls -la /sys/block
lrwxrwxrwx 1 0 loop3 -> ../devices/virtual/block/loop3
lrwxrwxrwx 1 0 loop5 -> ../devices/virtual/block/loop5
lrwxrwxrwx 1 0 loop7 -> ../devices/virtual/block/loop7
lrwxrwxrwx 1 0 loop0 -> ../devices/virtual/block/loop0
lrwxrwxrwx 1 0 loop2 -> ../devices/virtual/block/loop2
lrwxrwxrwx 1 0 loop4 -> ../devices/virtual/block/loop4
lrwxrwxrwx 1 0 loop6 -> ../devices/virtual/block/loop6
lrwxrwxrwx 1 0 loop1 -> ../devices/virtual/block/loop1
dr-xr-xr-x 12 0 ..
drwxr-xr-x 2 0 .
--
You received this bug notification because you are a member of Ubuntu
Foundations Bugs, which is subscribed to lvm2 in Ubuntu.
https://bugs.launchpad.net/bugs/2102236
Title:
plucky/s390x (with root disk on dm?) does not come up after kernel
update + reboot
Status in Ubuntu on IBM z Systems:
New
Status in linux package in Ubuntu:
New
Status in lvm2 package in Ubuntu:
New
Bug description:
While using the plucky daily from March 12 (that still comes with kernel 6.12)
(and working around LP#2101831, by forcing the installation to not apply any updates)
I get a system installed, which is at kernel level 6.12.
Since 6.14 is out (in plucky release) since yesterday (March 12th) I tried to upgrade from 6.12 to 6.14,
and the update itself seemed to be smooth (I couldn't find any errors while doing a full-upgrade in the terminal - see attached logs).
But after executing a reboot, the system (in 3 different
configurations, that use dm) does not come up again, and ends up in
busybox, complaining that the root device couldn't be found:
# s1lp15 FCP/SCSI multipath with LVM
ALERT! /dev/disk/by-id/dm-uuid-LVM-SlleSC5YA825VOM3t0KHBVFrLJcNWsnwZsObNziIB9Bk2mSVphnuTEOQ2eFiBbE1 does not exist. Dropping to a shell!
# s1lp15 2DASDs with LVM:
ALERT! /dev/disk/by-id/dm-uuid-LVM-ePTbsojYPfgMacKXwpIMNMvxk80qGzlPhRYw7DJlovmqHyla9TK6NGc70p1JN29b does not exist. Dropping to a shell!
# s1lp15 FCP/SCSI Multipath no LVM
ALERT! /dev/disk/by-id/dm-uuid-part1-mpath-36005076306ffd6b60000000000002603 does not exist. Dropping to a shell!
However, using a single disk without dm (so: no multipath, no lvm) the
system is able to come up again after the reboot (after a kernel
upgrade).
# s1lp15 single DASD no LVM
here the root device is:
root=/dev/disk/by-path/ccw-0.0.260b-part1
and it exists.
In the 3 different cases that (that repeatedly fail) "/dev/disk/by-id"
is missing.
I am not sure yet what's causing this,
it can be an issue with the device-mapper/lvm2
but also udev rules or kernel.
So maybe I need to add the kernel as affected component too (for now).
To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-z-systems/+bug/2102236/+subscriptions
More information about the foundations-bugs
mailing list