Better Home Storage: MergerFS + SnapRAID on OpenMediaVault

There is a way to enjoy some benefits from RAID without "real RAID" - and a supplement to scheduled cold backup for home data storage.

· 4 min read
Better Home Storage: MergerFS + SnapRAID on OpenMediaVault

I decided to build a NAS with old hard drives chiefly for serving media files like movie, TV shows and music.

When building a multi-bay system, it's inevitable to consider which kind of disk redundancy to use. Traditional RAID (both hard and soft ones) would work, but it also have several cons:

  • Data are distributed onto disks so that it could be not read from a single drive
  • Requires several disks to work simultaneously when reading or writing
  • If all parity disks plus 1 more disk corrupt, all data will be lost

For home data storage, I'd prefer scheduled cold backup. But is there any way to enjoy some benefits from RAID without "real RAID"? Yes, there is!

Backgrounds

MergerFS

MergerFS could pool multiple filesystems or folders into one place. It allows us to create RAID0-like filesystems but keeps data intact when some drives are broken.

mergerfs is a union filesystem geared towards simplifying storage and management of files across numerous commodity storage devices. It is similar to mhddfs, unionfs, and aufs.

SnapRAID

SnapRAID is short for "Snapshot RAID". It is a backup program for disk arrays and capable of recovering from up to six disk failures. It is mainly targeted for a home media center, with a lot of big files that rarely change.

Some highlights of SnapRAID are:

  • The disks can have different sizes
  • Add disks at any time
  • Quit SnapRAID without moving data and formatting disks
  • Anti-delete function

Differences between mergerfsfolders & unionfilesystems

There are 2 similar plugins based on mergerfs in OMV-Extras:

  • mergerfsfolders pools folders
  • unionfilesystems pools drives

With  blank new drives here, we would use unionfilesystems for simplified configuration.

Prepare System and Drives

I have 4×2TiB hard drives and plan to use 3 of them for data and 1 for parity check.

In OpenMediaVault:

  1. Install and configure OpenMediaVault as needed.
  2. Install OMV-Extras repository.
  3. Install openmediavault-snapraid and openmediavault-unionfilesystem plugin.
  4. Wipe all the disks and create filesystems on them.

I also labeled all file systems with DXX for data disks and PXX for the lonely parity disk.

Configure MergerFS

In OpenMediaVault control panel:

  1. Navigate to Storage > Union Filesystems.
  2. Click on Add.
  3. Type a pool name in Name.
  4. Check all data disks (but exclude parity disks) in Branches.
  5. Select Existing path, most free space (or epmfs) in Create Policy.
  6. Save and apply.

Here I set create policy to epmfs because I want the files gather together as possible as they could be. Under epmfs policy, new files will be created together in the same drive when the upper level folder path exists. You can check other available create policies upon your needs.

Note that some applications (`libttorrent` applications, sqlite3, etc.) requires mmap. Since I would use qBittorrent, I need to add some options:

allow_other,use_ino,cache.files=partial,dropcacheonclose=true,...

Also as I have labeled data disks as D01, D02 and D03, I expect MergerFS write files into them in that given order. Unfortunately this order could not be specified or modified via the control panel. I have to edit /etc/fstab:

....
/srv/dev-disk-by-uuid-${D01}:/srv/dev-disk-by-uuid-${D02}:/srv/dev-disk-by-uuid-${D03}            /srv/${MERGERFS_UUID}       fuse.mergerfs   defaults,allow_other,cache.files=off,use_ino,category.create=ff,minfreespace=4G,fsname=${MERGERFS_NAME}:${MERGERFS_UUID},x-systemd.requires=/srv/dev-disk-by-uuid-${D01},x-systemd.requires=/srv/dev-disk-by-uuid-${D02},x-systemd.requires=/srv/dev-disk-by-uuid-${D03} 0 0
....

Save the file, umount and mount MergerFS mountpoint. Now the writing order is corrected.

Configure SnapRAID

First configure data disks. In OpenMediaVault control panel:

  1. Navigate to Services > SnapRAID.
  2. Navigate to Disks tab.
  3. Click on Add.
  4. Select a data disk for Disk.
  5. Check Content and Data, then click on Save.

Repeat this until all data disks are added. Then configure parity disks:

  1. Repeat step 1-4 when configuring data disks.
  2. Check Parity , then click on Save.

Also repeat this until all parity disks are configured.

Initialize and Use SnapRAID

First Start

After SnapRAID is set up for the first time, you should synchronize the data, create and check parity. There are several useful commands:

  • sync hashes files and save the parity.
  • scrub verifies the data with the generated hashes and detect errors
  • status shows the current array status

So in combined:

snapraid {sync, scrub, status}

Rules

Not every piece of data worth backing up. Most of the time for home usage, only "static files" need to be backed up. For example, there is no strong necessity to back up container images, caches and other temporary files. These directories or files could be excluded in Rules tab.

Scheduled Diff and Cron Jobs

Click on Scheduled diff button and SnapRAID plugin will generate a cron job in Scheduled Jobs tab. This job only contains diff operation which calculates file differences from the last snap. To maintain the parity synchronized,  sync and scrub jobs are also needed. If you plan to synchronize automatically, you would like to check a helper script when creating jobs.

Sometimes empty files could be created by programs and cause errors:

The file '${FILENAME}' has unexpected zero size!
It's possible that after a kernel crash this file was lost,
and you can use 'snapraid fix -f ${FILENAME}' to recover it.
If this an expected condition you can 'sync' anyway using 'snapraid --force-zero sync'

If that file seems to be fine with a zero content with you, simply add the --force-zero parameter to synchronize.

Related Articles

Install FreeDOS - but Manually
· 2 min read

Labels in OpenShift 3 Web Console

Labels are used to match application names and resources in OpenShift 3 Web Console.

· 1 min read