Setting up my 2nd set of Raid drives - HPE Gen10 Plus
Robert Moskowitz
rgm at htt-consult.com
Tue Jun 24 03:48:12 UTC 2025
With google's AI I have come up with the following:
sgdisk -Z /dev/md2
sgdisk -n 1:0:0 -t 1:8300 -c 1:"RAID_Storage2" /dev/md2
mkfs.ext4 /dev/md2p1
mkdir /Storage2
mount /dev/md2p1 /Storage2
blkid /dev/md2p1
UUID=<your_raid_uuid> /Storage2 ext4 defaults 0 2
I LOOKS right, but some extra eyes, please,,,,
Thanks
On 6/23/25 7:28 PM, Robert Moskowitz via ubuntu-users wrote:
> This is for my HPE Gen10 Plus.
>
> In the Ubuntu 24 install, I put the boot drive on the Internal USB
> stick, as you will see below.
>
> I set up two Raid groups, messed up and restarted Install. Only this
> time just creating one Raid group (md0) and setting it up as one big
> ext4 partition.
>
> Now I want to setup the remaining 2 drives as an ext4 partition as
> directory /storage2
>
> The only guide I have so far is:
>
> https://askubuntu.com/questions/1299978/install-ubuntu-20-04-desktop-with-raid-1-and-lvm-on-machine-with-uefi-bios
>
>
> df -h shows:
>
> Filesystem Size Used Avail Use% Mounted on
> tmpfs 780M 1.3M 778M 1% /run
> efivarfs 494K 125K 365K 26% /sys/firmware/efi/efivars
> /dev/md0 3.6T 393G 3.1T 12% /
> tmpfs 3.9G 0 3.9G 0% /dev/shm
> tmpfs 5.0M 0 5.0M 0% /run/lock
> /dev/sdf1 688M 6.2M 682M 1% /boot/efi
> tmpfs 780M 12K 780M 1% /run/user/1000
>
> So definitely md0. But I run cat /proc/mdstat:
>
> Personalities : [raid1] [raid0] [raid6] [raid5] [raid4] [raid10]
> md2 : active raid1 sdc[1] sda[0]
> 3906886464 blocks super 1.2 [2/2] [UU]
> bitmap: 0/30 pages [0KB], 65536KB chunk
>
> md0 : active raid1 sdb[0] sdd[1]
> 3906886464 blocks super 1.2 [2/2] [UU]
> bitmap: 0/30 pages [0KB], 65536KB chunk
>
> And somehow, my aborted effort to make that 2nd Raid group stayed in
> the install. Or so it seems.
>
> But to further confuse me:
>
> # lsblk
> NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS
> sda 8:0 0 3.6T 0 disk
> └─md2 9:2 0 3.6T 0 raid1
> └─vg0-lv--0 252:0 0 3.6T 0 lvm
> sdb 8:16 0 3.6T 0 disk
> └─md0 9:0 0 3.6T 0 raid1 /
> sdc 8:32 0 3.6T 0 disk
> └─md2 9:2 0 3.6T 0 raid1
> └─vg0-lv--0 252:0 0 3.6T 0 lvm
> sdd 8:48 0 3.6T 0 disk
> └─md0 9:0 0 3.6T 0 raid1 /
> sde 8:64 0 2.7T 0 disk
> └─sde1 8:65 0 2.7T 0 part
> sdf 8:80 1 14.5G 0 disk
> └─sdf1 8:81 1 689M 0 part /boot/efi
>
> I did choose lvm in that aborted install for the 2nd Raid group, and
> it seems like that is still there?
>
> I don't think I need lvm, I am open to why I might. But can someone
> point me to a guide that will get md2 usable as /storage2?
>
> My googling for such a guide has come up empty...
>
> Thanks!
>
> Oh, and I will need to figure out which physical drives are which sd_
> so if I need to, I will know which to fix.
>
> And I have rsyncd up on this box ready to start receiving backups. But
> no poweron hardware timer, it seems, in the Gen10+, so I will need a
> powerstrip with a timer to turn it on at 1am and a cron job to
> shutdown at 4am...
>
>
>
>
More information about the ubuntu-users
mailing list