{{out of date|grub-legacy is unsupported; 0.9 superblocks and MBR partitioning are also therefore non-useful}}

You already have a fully functional system setup on a single drive, but you would like to add some redundancy to the setup by using RAID-1 to mirror your data across 2 drives. This guide follows the following steps to make the required changes, without losing data.

You already have a fully functional system setup on a single drive, but you would like to add some redundancy to the setup by using RAID-1 to mirror your data across 2 drives. This guide follows the following steps to make the required changes, without losing data.

* Create a single-disk RAID-1 array with our new disk

* Create a single-disk RAID-1 array with our new disk

Line 7:

Line 8:

* Wipe the old disk and add it to the new RAID-1 array

* Wipe the old disk and add it to the new RAID-1 array

−

{{Warning | Make a backup first. Even though our aim is to convert to a RAID setup without losing data, there's no guarantees the process will be perfect, and there is a high risk of accidents happening.}}

+

{{Warning | Make a backup first. Even though our aim is to convert to a RAID setup without losing data, there's no guarantee the process will be perfect, and there is a high risk of accidents happening.}}

+

== Assumptions ==

−

== Assumptions ==

* I will assume for the sake of the guide that the disk currently in your system is {{ic|/dev/sda}} and your new disk is {{ic|/dev/sdb}}.

* I will assume for the sake of the guide that the disk currently in your system is {{ic|/dev/sda}} and your new disk is {{ic|/dev/sdb}}.

* We will create the following configuration:

* We will create the following configuration:

Line 22:

Line 23:

== Create new RAID array ==

== Create new RAID array ==

+

First we need to create a single-disk RAID array using the new disk.

First we need to create a single-disk RAID array using the new disk.

+

=== Partition the Disk ===

=== Partition the Disk ===

+

Use fdisk or your partitioning program of choice to setup 2 primary partitions on your new disk. Make the swap partition half the size of the total swap you want (the other half will go on the other disk).

Use fdisk or your partitioning program of choice to setup 2 primary partitions on your new disk. Make the swap partition half the size of the total swap you want (the other half will go on the other disk).

Drop to single user mode:

Drop to single user mode:

−

telinit 1

+

# telinit 1

To see the current partitions:

To see the current partitions:

−

fdisk -l

+

# fdisk -l

To partition the new disk

To partition the new disk

−

fdisk /dev/sdb

+

# fdisk /dev/sdb

Then the fdisk commands to partition the new disk. Note that everything after the "#" is an explanation of what the command is doing:

Then the fdisk commands to partition the new disk. Note that everything after the "#" is an explanation of what the command is doing:

Line 60:

Line 64:

At the end of partitioning, your partitions should look something like this:

At the end of partitioning, your partitions should look something like this:

Note: If the above command causes mdadm to say "no such device /dev/sdb2", then reboot, and run the command again.

Note: If the above command causes mdadm to say "no such device /dev/sdb2", then reboot, and run the command again.

−

If you want to use Grub 0.97 (default in the Arch Linux 2010.05 release) on RAID 1, you need to specify an older version of metadata than the default. Add the option "--metadata=0.90" to the above command. Otherwise Grub will respond with "Filesystem type unknown, partition type 0xfd" and refuse to install. This is supposedly not necessary with Grub 2.

The devices are intact, however in a degraded state. (Because it's missing half the array!)

The devices are intact, however in a degraded state. (Because it's missing half the array!)

=== Make file systems ===

=== Make file systems ===

+

Use the file system of your preference here. I'll use ext3 for this guide.

Use the file system of your preference here. I'll use ext3 for this guide.

[root@arch ~]# mkfs -t ext3 -j -L RAID-ONE /dev/md0

[root@arch ~]# mkfs -t ext3 -j -L RAID-ONE /dev/md0

Line 119:

Line 128:

Make a file system on the swap partition:

Make a file system on the swap partition:

−

[root@arch ~]# mkswap -L NEW-SWAP /dev/sdb1

+

{{hc|# mkswap -L NEW-SWAP /dev/sdb1|<nowiki>

Setting up swapspace version 1, size = 271314 kB

Setting up swapspace version 1, size = 271314 kB

LABEL=NEW-SWAP, UUID=9d746813-2d6b-4706-a56a-ecfd108f3fe9

LABEL=NEW-SWAP, UUID=9d746813-2d6b-4706-a56a-ecfd108f3fe9

+

</nowiki>}}

== Copy data ==

== Copy data ==

+

The new RAID-1 array is ready to start accepting data! So now we need to mount the array, and copy everything from the old system to the new system

The new RAID-1 array is ready to start accepting data! So now we need to mount the array, and copy everything from the old system to the new system

=== Mount the array ===

=== Mount the array ===

−

[root@arch ~]# mkdir /mnt/new-raid

+

−

[root@arch ~]# mount /dev/md0 /mnt/new-raid

+

# mkdir /mnt/new-raid

+

# mount /dev/md0 /mnt/new-raid

=== Copy the data ===

=== Copy the data ===

−

[root@arch ~]# rsync -avxHAXS --delete --progress / /mnt/new-raid

+

+

# rsync -avxHAXS --delete --progress / /mnt/new-raid

Note that by using the '''-x''' option you are limiting rsync to a single file system. If you have a more traditional file system layout, with different partitions for /boot, /home, and perhaps others, you will need to rsync those file systems separately. For example:

Note that by using the '''-x''' option you are limiting rsync to a single file system. If you have a more traditional file system layout, with different partitions for /boot, /home, and perhaps others, you will need to rsync those file systems separately. For example:

Alternatively, you can use tar instead of the above rsync command if you prefer. rsync will, however, be quicker if you are only copying over changes. The tar command is: <code>tar -C / -clspf - . | tar -xlspvf -</code>

+

Alternatively, you can use tar instead of the above rsync command if you prefer. rsync will, however, be quicker if you are only copying over changes. The tar command is: {{ic|<nowiki>tar -C / -clspf - . | tar -xlspvf -</nowiki>}}

=== Update GRUB legacy ===

=== Update GRUB legacy ===

+

Use your preferred text editor to open {{ic|/mnt/new-raid/boot/grub/menu.lst}}.

Use your preferred text editor to open {{ic|/mnt/new-raid/boot/grub/menu.lst}}.

You need to tell fstab on the '''new''' disk where to find the new devices. It's better to use UUID codes here, which should not change, even if our partition detection order changes or a drive gets removed.

You need to tell fstab on the '''new''' disk where to find the new devices. It's better to use UUID codes here, which should not change, even if our partition detection order changes or a drive gets removed.

Look for the partition labeled "NEW-SWAP", on /dev/sdb1, that we created above. Copy your swap partition's UUID into the new fstab, as shown below. Of course we also add /dev/md0, as our root mount point.

+

Look for the partition labeled "NEW-SWAP", on {{ic|/dev/sdb1}}, that we created above. Copy your swap partition's UUID into the new fstab, as shown below. Of course we also add {{ic|/dev/md0}}, as our root mount point.

−

[root@arch ~]# cat /mnt/new-raid/etc/fstab

+

{{hc|# cat /mnt/new-raid/etc/fstab|<nowiki>

/dev/md0 / ext3 defaults 0 1

/dev/md0 / ext3 defaults 0 1

UUID=9ff5682b-d5a1-4ed5-8d63-d1df911e0142 none swap sw 0 0

UUID=9ff5682b-d5a1-4ed5-8d63-d1df911e0142 none swap sw 0 0

−

<!-- /dev/sdb1 swap swap defaults 0 0 -->

+

</nowiki>}}

=== Rebuild initcpio or initramfs ===

=== Rebuild initcpio or initramfs ===

Line 196:

Line 213:

==== First chroot into the RAID system ====

==== First chroot into the RAID system ====

−

[root@arch ~]# mount --bind /sys /mnt/new-raid/sys

+

# mount --bind /sys /mnt/new-raid/sys

−

[root@arch ~]# mount --bind /proc /mnt/new-raid/proc

+

# mount --bind /proc /mnt/new-raid/proc

−

[root@arch ~]# mount --bind /dev /mnt/new-raid/dev

+

# mount --bind /dev /mnt/new-raid/dev

−

[root@arch ~]# chroot /mnt/new-raid/

+

# chroot /mnt/new-raid/

−

[root /]#

−

If the chroot command gives you an error like <code>chroot: failed to run command `/bin/zsh': No such file or directory</code>, then use <code>chroot /mnt/new-raid/ /bin/bash</code> instead.

+

If the chroot command gives you an error like {{ic|chroot: failed to run command `/bin/zsh': No such file or directory}}, then use {{ic|chroot /mnt/new-raid/ /bin/bash}} instead.

You are now chrooted in what will become the root of your RAID-1 system. Complete the appropriate section below for your distribution (almost every other step is identical, regardless of the Linux variant).

You are now chrooted in what will become the root of your RAID-1 system. Complete the appropriate section below for your distribution (almost every other step is identical, regardless of the Linux variant).

Line 208:

Line 224:

==== Record mdadm's config ====

==== Record mdadm's config ====

−

For Arch Linux, use "/etc/mdadm.conf", for Ubuntu or Debian, use "/etc/mdadm/mdadm.conf"

+

For Arch Linux, use {{ic|/etc/mdadm.conf}}, for Ubuntu or Debian, use {{ic|/etc/mdadm/mdadm.conf}}}

−

nano /etc/mdadm.conf

+

# nano /etc/mdadm.conf

... and change the "MAILADDR" line to be your email address, if you want emailed alerts of problems with the RAID-1.

... and change the "MAILADDR" line to be your email address, if you want emailed alerts of problems with the RAID-1.

Then save the array configuration with UUIDs to make it easier for the system to find /dev/md0 on boot-up. If you do not do this, you can get an "ALERT! /dev/md0 does not exist" error when booting :

Then save the array configuration with UUIDs to make it easier for the system to find /dev/md0 on boot-up. If you do not do this, you can get an "ALERT! /dev/md0 does not exist" error when booting :

−

mdadm --detail --scan >> /etc/mdadm.conf

+

# mdadm --detail --scan >> /etc/mdadm.conf

−

==== For Arch Linux: Rebuild initcpio ====

+

==== For Arch Linux: rebuild initcpio ====

Edit {{ic|/etc/mkinitcpio.conf}} to include {{Ic|mdadm}} in the HOOKS array. Place it after {{Ic|autodetect}}, {{Ic|sata}}, {{Ic|scsi}} and {{Ic|pata}} (whichever is appropriate for your hardware).

Edit {{ic|/etc/mkinitcpio.conf}} to include {{Ic|mdadm}} in the HOOKS array. Place it after {{Ic|autodetect}}, {{Ic|sata}}, {{Ic|scsi}} and {{Ic|pata}} (whichever is appropriate for your hardware).

−

[root /]# mkinitcpio -g /boot/kernel26.img

+

# mkinitcpio -p linux

−

[root /]# exit

+

# exit

−

==== For Ubuntu or Debian: Rebuild initramfs ====

+

==== For Ubuntu or Debian: rebuild initramfs ====

Then rebuild initramfs, incorporating the two above changes:

Then rebuild initramfs, incorporating the two above changes:

−

update-initramfs -k `uname -r` -c -t

+

# update-initramfs -k $(uname -r) -c -t

This will rebuild your running version - to rebuild others, this will show a listing (in Ubuntu):

This will rebuild your running version - to rebuild others, this will show a listing (in Ubuntu):

Then substitute/script in these others so that all are available for use with the new RAID setup.

Then substitute/script in these others so that all are available for use with the new RAID setup.

=== Install GRUB on the RAID array ===

=== Install GRUB on the RAID array ===

−

Start grub:

−

[root@arch ~]# grub --no-floppy

+

Start GRUB:

+

+

# grub --no-floppy

Then we find our two partitions - the current one (hd0,0) (I.e. first disk, first partition), and (hd1,1) (i.e. the partition we just added above, on the second partition of the second drive). Check you get two results here:

Then we find our two partitions - the current one (hd0,0) (I.e. first disk, first partition), and (hd1,1) (i.e. the partition we just added above, on the second partition of the second drive). Check you get two results here:

Line 241:

Line 258:

(hd1,1)

(hd1,1)

−

Then we tell grub to assume the new second drive is (hd0), i.e. the first disk in the system (when it is not currently the case). If your first disk fails, however, and you remove it, or you change the order disks are detected in the BIOS so that you can boot from your second disk, then your second disk will become the first disk in the system. The MBR will then be correct, your new second drive will have become your first drive, and you will be able to boot from this disk.

+

Then we tell GRUB to assume the new second drive is (hd0), i.e. the first disk in the system (when it is not currently the case). If your first disk fails, however, and you remove it, or you change the order disks are detected in the BIOS so that you can boot from your second disk, then your second disk will become the first disk in the system. The MBR will then be correct, your new second drive will have become your first drive, and you will be able to boot from this disk.

grub> device (hd0) /dev/sdb

grub> device (hd0) /dev/sdb

Line 259:

Line 276:

== Verify success ==

== Verify success ==

+

Reboot your computer, making sure it boots from the new RAID disk ({{ic|/dev/sdb}}) and not the original disk ({{ic|/dev/sda}}). You may need to change the boot device priorities in your BIOS to do this.

Reboot your computer, making sure it boots from the new RAID disk ({{ic|/dev/sdb}}) and not the original disk ({{ic|/dev/sda}}). You may need to change the boot device priorities in your BIOS to do this.

Line 264:

Line 282:

Verify you have booted from the RAID array by looking at the output of mount. Also check mdstat again only to confirm which disk is in the array.

Verify you have booted from the RAID array by looking at the output of mount. Also check mdstat again only to confirm which disk is in the array.

Syncing can take a while. If the machine is not needed for other tasks the speed limit can be increased.

Syncing can take a while. If the machine is not needed for other tasks the speed limit can be increased.

−

[root@arch ~]# cat /proc/mdstat

+

{{hc|# cat /proc/mdstat|<nowiki>

Personalities : [raid1]

Personalities : [raid1]

md0 : active raid1 sda3[2] sdb3[1]

md0 : active raid1 sda3[2] sdb3[1]

Line 328:

Line 351:

unused devices: <none>

unused devices: <none>

+

</nowiki>}}

Check the current speed limit.

Check the current speed limit.

−

[root@arch ~]# cat /proc/sys/dev/raid/speed_limit_min

+

{{hc|# cat /proc/sys/dev/raid/speed_limit_min|

−

1000

+

1000

−

[root@arch ~]# cat /proc/sys/dev/raid/speed_limit_max

+

}}

−

200000

+

{{hc|# cat /proc/sys/dev/raid/speed_limit_max|

+

200000

+

}}

Increase the limits.

Increase the limits.

−

[root@arch ~]# echo 400000 >/proc/sys/dev/raid/speed_limit_min

+

# echo 400000 >/proc/sys/dev/raid/speed_limit_min

−

[root@arch ~]# echo 400000 >/proc/sys/dev/raid/speed_limit_max

+

# echo 400000 >/proc/sys/dev/raid/speed_limit_max

Then check out the syncing speed and estimated finish time.

Then check out the syncing speed and estimated finish time.

−

[root@arch ~]# cat /proc/mdstat

+

{{hc|# cat /proc/mdstat|<nowiki>

Personalities : [raid1]

Personalities : [raid1]

md0 : active raid1 sda3[2] sdb3[1]

md0 : active raid1 sda3[2] sdb3[1]

Line 350:

Line 376:

unused devices: <none>

unused devices: <none>

+

</nowiki>}}

=== Add second swap partition ===

=== Add second swap partition ===

+

The partition was created with sfdisk, but it still has to be formatted for swap.

The partition was created with sfdisk, but it still has to be formatted for swap.

−

[root@arch ~]# mkswap -L SWAP /dev/sda1

+

{{hc|# mkswap -L SWAP /dev/sda1|<nowiki>

Setting up swapspace version 1, size = 271314 kB

Setting up swapspace version 1, size = 271314 kB

LABEL=SWAP, UUID=1acd55dc-f73f-4639-94bc-3f30c33710c9

LABEL=SWAP, UUID=1acd55dc-f73f-4639-94bc-3f30c33710c9

+

</nowiki>}}

Then add this UUID to the fstab exactly like the other one earlier. When done, it should look similar to this:

Then add this UUID to the fstab exactly like the other one earlier. When done, it should look similar to this:

−

[root@arch ~]# cat /mnt/new-raid/etc/fstab

+

{{hc|# cat /mnt/new-raid/etc/fstab|<nowiki>

/dev/md0 / ext3 defaults 0 1

/dev/md0 / ext3 defaults 0 1

UUID=1acd55dc-f73f-4639-94bc-3f30c33710c9 none swap sw 0 0

UUID=1acd55dc-f73f-4639-94bc-3f30c33710c9 none swap sw 0 0

UUID=9ff5682b-d5a1-4ed5-8d63-d1df911e0142 none swap sw 0 0

UUID=9ff5682b-d5a1-4ed5-8d63-d1df911e0142 none swap sw 0 0

+

</nowiki>}}

It can be activated immediately:

It can be activated immediately:

−

[root@arch ~]# swapon /dev/sda1

+

# swapon /dev/sda1

=== Verify that email alerts are working ===

=== Verify that email alerts are working ===

If you run this command, then you should get a notification email showing the contents of /proc/mdstat :

If you run this command, then you should get a notification email showing the contents of /proc/mdstat :

−

[root@arch ~]# mdadm --monitor --test --oneshot /dev/md0

+

# mdadm --monitor --test --oneshot /dev/md0

Check that you get the test email notification. This way you can be aware if one of the disks in the array fails (otherwise it may fail silently, putting you at risk of data loss if another drive should also fail). If you do not get an email notification, check the "MAILADDR" line in mdadm.conf, and also check that sending an email to this address from the command line (e.g. using the "mail" command) works.

Check that you get the test email notification. This way you can be aware if one of the disks in the array fails (otherwise it may fail silently, putting you at risk of data loss if another drive should also fail). If you do not get an email notification, check the "MAILADDR" line in mdadm.conf, and also check that sending an email to this address from the command line (e.g. using the "mail" command) works.

== Automatic conversion tool alternative - Raider ==

== Automatic conversion tool alternative - Raider ==

−

You may consider to use '''Raider''', which is a tool that is able to convert a single disk in to a Raid system ('''1, 4, 5, 6''' or '''10''') with a two-pass command.

+

You may consider to use {{AUR|Raider}}, which is a tool that is able to convert a single disk in to a Raid system ('''1, 4, 5, 6''' or '''10''') with a two-pass command.

−

*Website: http://raider.sourceforge.net/

+

* Website: http://raider.sourceforge.net/

−

*It is also available in AUR: {{AUR|Raider}}

Revision as of 13:07, 10 June 2013

You already have a fully functional system setup on a single drive, but you would like to add some redundancy to the setup by using RAID-1 to mirror your data across 2 drives. This guide follows the following steps to make the required changes, without losing data.

Create a single-disk RAID-1 array with our new disk

Move all your data from the old-disk to the new RAID-1 array

Verify the data move was successful

Wipe the old disk and add it to the new RAID-1 array

Warning: Make a backup first. Even though our aim is to convert to a RAID setup without losing data, there's no guarantee the process will be perfect, and there is a high risk of accidents happening.

The swap partitions will not be in a RAID array as having swap on RAID serves no purpose. Refer to this article for reasons why.

To minimize the risk of Data on Disk (DoD) changing in the middle of our changes, I suggest you drop to single user mode before you start by using the telinit 1 command.

You will need to be the root user for the entire process.

Create new RAID array

First we need to create a single-disk RAID array using the new disk.

Partition the Disk

Use fdisk or your partitioning program of choice to setup 2 primary partitions on your new disk. Make the swap partition half the size of the total swap you want (the other half will go on the other disk).

Drop to single user mode:

# telinit 1

To see the current partitions:

# fdisk -l

To partition the new disk

# fdisk /dev/sdb

Then the fdisk commands to partition the new disk. Note that everything after the "#" is an explanation of what the command is doing:

c # Turn off DOS compatibility (optional).
n # new
p # primary
1 # first partition
1 # start at first cylinder
101 # end cylinder, 0.1% of the disk. Note: update this number as appropriate for your disk.
n # new
p # primary
2 # second partition
press enter # Uses the default start from the end of the first partition
press enter # Uses the default of using all the remain space on the disk.
t # set the partition type
1 # for partition number 1
82 # ... and set it to be swap
t # set the partition type
2 # for partition number 2 ...
fd # ... and set it to be "linux raid auto"
a # Toggle the bootable flag to be "on"
2 # for partition number 2.
p # print what the partition table will look like
w # now write all of the above changes to disk

At the end of partitioning, your partitions should look something like this:

Copy data

The new RAID-1 array is ready to start accepting data! So now we need to mount the array, and copy everything from the old system to the new system

Mount the array

# mkdir /mnt/new-raid
# mount /dev/md0 /mnt/new-raid

Copy the data

# rsync -avxHAXS --delete --progress / /mnt/new-raid

Note that by using the -x option you are limiting rsync to a single file system. If you have a more traditional file system layout, with different partitions for /boot, /home, and perhaps others, you will need to rsync those file systems separately. For example:

Alternatively, you can use tar instead of the above rsync command if you prefer. rsync will, however, be quicker if you are only copying over changes. The tar command is: tar -C / -clspf - . | tar -xlspvf -

Update GRUB legacy

Use your preferred text editor to open /mnt/new-raid/boot/grub/menu.lst.

Update GRUB

Alter fstab

You need to tell fstab on the new disk where to find the new devices. It's better to use UUID codes here, which should not change, even if our partition detection order changes or a drive gets removed.

Look for the partition labeled "NEW-SWAP", on /dev/sdb1, that we created above. Copy your swap partition's UUID into the new fstab, as shown below. Of course we also add /dev/md0, as our root mount point.

Rebuild initcpio or initramfs

First chroot into the RAID system

If the chroot command gives you an error like chroot: failed to run command `/bin/zsh': No such file or directory, then use chroot /mnt/new-raid/ /bin/bash instead.

You are now chrooted in what will become the root of your RAID-1 system. Complete the appropriate section below for your distribution (almost every other step is identical, regardless of the Linux variant).

Record mdadm's config

For Arch Linux, use /etc/mdadm.conf, for Ubuntu or Debian, use /etc/mdadm/mdadm.conf}

# nano /etc/mdadm.conf

... and change the "MAILADDR" line to be your email address, if you want emailed alerts of problems with the RAID-1.

Then save the array configuration with UUIDs to make it easier for the system to find /dev/md0 on boot-up. If you do not do this, you can get an "ALERT! /dev/md0 does not exist" error when booting :

# mdadm --detail --scan >> /etc/mdadm.conf

For Arch Linux: rebuild initcpio

Edit /etc/mkinitcpio.conf to include mdadm in the HOOKS array. Place it after autodetect, sata, scsi and pata (whichever is appropriate for your hardware).

# mkinitcpio -p linux
# exit

For Ubuntu or Debian: rebuild initramfs

Then rebuild initramfs, incorporating the two above changes:

# update-initramfs -k $(uname -r) -c -t

This will rebuild your running version - to rebuild others, this will show a listing (in Ubuntu):

Then substitute/script in these others so that all are available for use with the new RAID setup.

Install GRUB on the RAID array

Start GRUB:

# grub --no-floppy

Then we find our two partitions - the current one (hd0,0) (I.e. first disk, first partition), and (hd1,1) (i.e. the partition we just added above, on the second partition of the second drive). Check you get two results here:

grub> find /boot/grub/stage1
(hd0,0)
(hd1,1)

Then we tell GRUB to assume the new second drive is (hd0), i.e. the first disk in the system (when it is not currently the case). If your first disk fails, however, and you remove it, or you change the order disks are detected in the BIOS so that you can boot from your second disk, then your second disk will become the first disk in the system. The MBR will then be correct, your new second drive will have become your first drive, and you will be able to boot from this disk.

grub> device (hd0) /dev/sdb

Then we install GRUB onto the MBR of our new second drive. Check that the "partition type" is detected as "0xfd", as shown below, to make sure you have the right partition:

Note it is the swap partition on sdb that is in use, nothing from sda.

If system boots fine, and the output of the above commands is correct, then congratulations! You're now running off the degraded RAID array. We can add the original disk to the array now to bring it up to full performance.

Add original disk to array

Partition original disk

Copy the partition table from /dev/sdb (newly implemented RAID disk) to /dev/sda (second disk we are adding to the array) so that both disks have exactly the same layout.

# sfdisk -d /dev/sdb | sfdisk /dev/sda

Alternate method - this will output the /dev/sdb partition layout to a file, then it's used as input for partitioning /dev/sda.

Verify that email alerts are working

If you run this command, then you should get a notification email showing the contents of /proc/mdstat :

# mdadm --monitor --test --oneshot /dev/md0

Check that you get the test email notification. This way you can be aware if one of the disks in the array fails (otherwise it may fail silently, putting you at risk of data loss if another drive should also fail). If you do not get an email notification, check the "MAILADDR" line in mdadm.conf, and also check that sending an email to this address from the command line (e.g. using the "mail" command) works.

Automatic conversion tool alternative - Raider

You may consider to use RaiderAUR, which is a tool that is able to convert a single disk in to a Raid system (1, 4, 5, 6 or 10) with a two-pass command.