Setting up my 2nd set of Raid drives - HPE Gen10 Plus

Robert Moskowitz rgm at htt-consult.com
Tue Jun 24 03:48:12 UTC 2025


With google's AI I have come up with the following:

sgdisk -Z /dev/md2
sgdisk -n 1:0:0 -t 1:8300 -c 1:"RAID_Storage2" /dev/md2
mkfs.ext4 /dev/md2p1
mkdir /Storage2
mount /dev/md2p1 /Storage2
blkid /dev/md2p1
UUID=<your_raid_uuid> /Storage2 ext4 defaults 0 2

I LOOKS right, but some extra eyes, please,,,,

Thanks

On 6/23/25 7:28 PM, Robert Moskowitz via ubuntu-users wrote:
> This is for my HPE Gen10 Plus.
>
> In the Ubuntu 24 install, I put the boot drive on the Internal USB 
> stick, as you will see below.
>
> I set up two Raid groups, messed up and restarted Install.  Only this 
> time just creating one Raid group (md0) and setting it up as one big 
> ext4 partition.
>
> Now I want to setup the remaining 2 drives as an ext4 partition as 
> directory /storage2
>
> The only guide I have so far is:
>
> https://askubuntu.com/questions/1299978/install-ubuntu-20-04-desktop-with-raid-1-and-lvm-on-machine-with-uefi-bios 
>
>
> df -h shows:
>
> Filesystem      Size  Used Avail Use% Mounted on
> tmpfs           780M  1.3M  778M   1% /run
> efivarfs        494K  125K  365K  26% /sys/firmware/efi/efivars
> /dev/md0        3.6T  393G  3.1T  12% /
> tmpfs           3.9G     0  3.9G   0% /dev/shm
> tmpfs           5.0M     0  5.0M   0% /run/lock
> /dev/sdf1       688M  6.2M  682M   1% /boot/efi
> tmpfs           780M   12K  780M   1% /run/user/1000
>
> So definitely md0.  But I run cat /proc/mdstat:
>
> Personalities : [raid1] [raid0] [raid6] [raid5] [raid4] [raid10]
> md2 : active raid1 sdc[1] sda[0]
>       3906886464 blocks super 1.2 [2/2] [UU]
>       bitmap: 0/30 pages [0KB], 65536KB chunk
>
> md0 : active raid1 sdb[0] sdd[1]
>       3906886464 blocks super 1.2 [2/2] [UU]
>       bitmap: 0/30 pages [0KB], 65536KB chunk
>
> And somehow, my aborted effort to make that 2nd Raid group stayed in 
> the install.  Or so it seems.
>
> But to further confuse me:
>
> # lsblk
> NAME          MAJ:MIN RM  SIZE RO TYPE  MOUNTPOINTS
> sda             8:0    0  3.6T  0 disk
> └─md2           9:2    0  3.6T  0 raid1
>   └─vg0-lv--0 252:0    0  3.6T  0 lvm
> sdb             8:16   0  3.6T  0 disk
> └─md0           9:0    0  3.6T  0 raid1 /
> sdc             8:32   0  3.6T  0 disk
> └─md2           9:2    0  3.6T  0 raid1
>   └─vg0-lv--0 252:0    0  3.6T  0 lvm
> sdd             8:48   0  3.6T  0 disk
> └─md0           9:0    0  3.6T  0 raid1 /
> sde             8:64   0  2.7T  0 disk
> └─sde1          8:65   0  2.7T  0 part
> sdf             8:80   1 14.5G  0 disk
> └─sdf1          8:81   1  689M  0 part  /boot/efi
>
> I did choose lvm in that aborted install for the 2nd Raid group, and 
> it seems like that is still there?
>
> I don't think I need lvm, I am open to why I might.  But can someone 
> point me to a guide that will get md2 usable as /storage2?
>
> My googling for such a guide has come up empty...
>
> Thanks!
>
> Oh, and I will need to figure out which physical drives are which sd_ 
> so if I need to, I will know which to fix.
>
> And I have rsyncd up on this box ready to start receiving backups. But 
> no poweron hardware timer, it seems, in the Gen10+, so I will need a 
> powerstrip with a timer to turn it on at 1am and a cron job to 
> shutdown at 4am...
>
>
>
>




More information about the ubuntu-users mailing list