Fix/Restore/Recompile/Remount RAID10 Array via Webmin>Hardware>Linux Raid

Hi All,

Most important part first: I’ve lost access to my main **DATA** array after trying to create a new, unrelated, array using (it appears) malfunctioning SSDs. The Webmin Linux Raid page might be able to save my **DATA** array, I hope. In a nutshell: Can I use Webmin to get sda, sdb, sdc and sdd back and hooked into a new /dev/md/*x* mount-point somehow?

Background:

My **DATA** array is composed of four 4TB spinners currently showing up like this:

sda           3.6T linux_raid_member disk   
└─md0         7.3T                   raid10 
  └─md0p1     7.3T ext4              part   
sdb           3.6T linux_raid_member disk   
└─md0         7.3T                   raid10 
  └─md0p1     7.3T ext4              part   
sdc           3.6T linux_raid_member disk   
└─md0         7.3T                   raid10 
  └─md0p1     7.3T ext4              part   
sdd           3.6T linux_raid_member disk   
└─md0         7.3T                   raid10 
  └─md0p1     7.3T ext4              part   

The array was originally created with the following mdadm command:

$ sudo mdadm --create --verbose /dev/md0 --level=10 --layout=f2 --raid-devices=4 /dev/sda /dev/sdb /dev/sdc /dev/sdd

Yesterday, after a few hours trying spent trying to create a new drive comprised of 3 partitions with a RAID1 array for each partition using two SSDs, meant to function as a “portable” TimeShift device to backup three of my LAN computers, I came to the realization that one or both of the SSDs were seriously corrupted/broken/kaput.

ROOT OF MY PROBLEM: At this point, I tried to “diagnose” these two SSDs by throwing a bunch of commands at them, using Disks and GParted - changing filesystem types, adding /removing flags, creating new UUIDs, and finally, I tried to wipe and format them. But nothing worked; they’re both in the garbage now.

Somehow, during all of these changes, it appears my **DATA** array entry (i.e., the typical /dev/md0 entry), no longer appeared as it was supposed to: there was now some empty space, a large, empty partition, and some more dead space. And this is where I stand right now.

Running cat /proc/mdstat gives me the following:

Personalities : [raid10] [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] 
md0 : active raid10 sdc[2] sdb[1] sdd[3] sda[0]
      7813771264 blocks super 1.2 512K chunks 2 far-copies [4/4] [UUUU]
      bitmap: 0/59 pages [0KB], 65536KB chunk

unused devices: <none>

There is some information not present there which I seem to recall seeing in the past? My /etc/mdadm/mdadm.conf entry:

ARRAY /dev/md/0 metadata=1.2 name=nas:0 UUID=dfb605b0-6029-4227-9f3a-622ca38f3606

…is no longer valid as the md0 entry shows up now as noted above, i.e., some blank space, a large partition, and some more blank space (it WAS a single contiguous partition before yesterday). And the large partition no longer has the UUID as above or in fstab:

UUID=dfb605b0-6029-4227-9f3a-622ca38f3606 /media/nas ext4 rw,suid,dev,exec,auto,user,group,async,nofail,discard 0 0

I THINK it is possible to fix my array by “re-attaching” the four 4Tb spinners (sda, sdb, sdc and sdd) to a new/fixed “md0” but I have no idea, and Webmin’s Linux Raid page appears to offer me some hope. This is absolutely critical as **DATA** is my ONLY copy of all my files and the last 3 years of my life, work, documents, images, music, i. e., everything.

So I’m very open to options right now, please and thank you.

Merci:)
Shawn

SYSTEM INFORMATION
OS type and version Ub22.04
Virtualmin version 7.7

@shawngibson,

Sadly, the simple answer here is…

Webmin cannot do anything more then you can do on the command line. There is no magic here, Webmin is merely a GUI to tools you already have installed on your system.

As for data recovery, that would really depend a lot on how broken things really are.

Things were, unfortunately, broken utterly. After two days of scouring the Internet for solutions, it became obvious to me that, somewhere along the line, I commanded the computer to do something which removed any possibility of retrieving my data. Sad day, and some very real hatred aimed intently at Pandora, the little minx.

Fortunately, my main work at present, my website, is on another box and another array. I can accept the loss of 3 years’ work otherwise, but images of my brother (who committed suicide @ Covid19), all of my tax and employment records for 20-odd years, websites & WP blogs I made going back to around 2005, ~60,000 negative scans from a now-unusable 35mm Scanner…and items I have not even considered which will show their painful non-existence many times in the future, I am sure…

I’ve always known that even a solid RAID10 setup is only “safe” if it is combined with at least one more (preferably off-site) storage solution, but such safety requires more money than I have right now.

Lesson learned: never underestimate the unexpected, and do what can be done to mitigate potential damage. While pissing around with those two malfunctioning SSDs in trying to build another array (for backup storage, ironically), I should have unmounted and powered-down my Data array…I did not. I knew, at the time, that I should, but I did not.

I don’t know how much data you have but I back up my mirrored raid with an external 3 gig drive. 4 gig is now $100 at Newegg. And that’s just a quick search.

You’re lucky:)

My “Data” array was 16Tb, about 75% full. The array on my VirtualMin/website box is 6Tb, about 35% full, but was meant to have a bunch of information passed over from the “Data” array. My Music array (I have a huge DAW setup) is 4Tb, mostly samples, and almost full lol. And as noted above, I was hoping to build a portable 2Tb array for all of my TimeShift backups, but that went tits-up (if you’ll pardon the expression).

I don’t have much of a tan, of course :rofl:

Yeah. My point was only that you might consider this for your absolute, can’t lose’ stuff. I’m considering setting up backup at my son’s and vice versa. Dynamic DNS and I pay for plenty of bandwidth I can’t possibly use anyhow.

This topic was automatically closed 60 days after the last reply. New replies are no longer allowed.