Volume 1 Crashed but drives are all normal.


Recommended Posts

As the title suggests when I woke up today my Plex was no longer working so I logged into my DSM and it warns that Volume 1 has crashed.

 

Now this is the only volume I Have which is made up of 7 8TB HDDs, and in Storage pool everything is normal.

 

I can boot the DSM and still access the files locally through the network but seems to be in limp mode.

 

No apps can run within DSM.

 

Other than setting this up initially about a year ago I have no Idea how to use Linux or even using terminal from outside of the NAS.

 

Have never had an issue with this and it bypassed just having normal HDD's fail and straight to oh sorry but your screwed.

 

Any help would be greatly appreciated.

 

Moochin.

Volume1.png

HDD Status.png

DSM.png

Basic Info.png

Overview.png

Edited by Moochin
Link to post
Share on other sites

check the logs in /var/log

when the self healing of btrfs fails its a sign of other trouble like ram, cable or controller problems

use you backup and offload whats not in the backup , repairing a btrfs is usually not a long term solution

locate the (hardware?) problem, fix it, recreate the volume with backup and offloaded data

 

Link to post
Share on other sites
Posted (edited)

Thanks for the help, so no way to continue without moving 30TB of data off then back into the NAS?

 

Running extended HDD tests atm and will move to memtest once this is done.

 

So if something else in the NAS fails your data redundancy doesn't help and you can just lose the lot with 2 weeks worth of work getting it all out and back if it even allows that depending what goes wrong?

 

I'd be fine with redoing the DSM and that, is there no way to recover the Volume without the coping off then back on though?

Edited by Moochin
Link to post
Share on other sites
On 4/12/2021 at 1:34 AM, Moochin said:

So if something else in the NAS fails your data redundancy doesn't help and you can just lose the lot with 2 weeks worth of work getting it all out and back if it even allows that depending what goes wrong?

thats what backup is for and the scale depends on then work done and the kind of data (juggling 8k videos needs more space then a photoshop project or some office documents

in most cases the really important data is just e few TB and the daily amount is not that much

have the data on your computer, backup that to nas (daily) and then backup the nas (weekly)

its also possible to use a usb hdd for backup of important data, it does not have to be a 2nd nas

 

On 4/12/2021 at 1:34 AM, Moochin said:

is there no way to recover the Volume without the coping off then back on though?

you can read on the internet about recovering btrfs and we have some cases here in the forum (2-3?) you can search for

i had that myself last year and even after trying to repair, the data last written where damaged or gone and the file system seemed beyond repair to me

if you have all the time to learn everything about btrfs and enough disk space to keep the old malfunctioning btrfs - but thats usually not the case, business continuity plan is also about letting go at some point and continue

you can try to involve data recovery specialists like kroll ontrack but i guess the money that would take is beyond what you can afford, data recovery is often compromising too, time and money are often the limit

 

its more important to find the source of the problem

 

 

Edited by IG-88
Link to post
Share on other sites
20 hours ago, Moochin said:

Could I just build a new system and install the raid into the new PC and merge it in or that won't work cause its crashed?

for sure, if the btrfs file system is damaged just using a different or new installed OS in is not going to change that

you can keep the system and delete the raid, but finding the source of the corruption is still to do and might be the more important task for now, without fixing that even trying to repair the file system in any way might be a fruitless effort

Link to post
Share on other sites
Posted (edited)
6 hours ago, IG-88 said:

for sure, if the btrfs file system is damaged just using a different or new installed OS in is not going to change that

you can keep the system and delete the raid, but finding the source of the corruption is still to do and might be the more important task for now, without fixing that even trying to repair the file system in any way might be a fruitless effort

Would not replacing the entire PC + sata cables solve that issue? all the hard drives are fine as they have been fully tested.

 

When I say system I mean the PC, I will buy all new parts and set up with a new boot loader and everything so the only thing that remains is the HDD's

Edited by Moochin
Link to post
Share on other sites
13 hours ago, Moochin said:

When I say system I mean the PC, I will buy all new parts and set up with a new boot loader and everything so the only thing that remains is the HDD's

its usually suggested to test it to some extend, i would at least run memtest over night. install dsm to a single hdd, copy some date and look into the logs in /var/log/ if there is anything unusual

Link to post
Share on other sites
5 hours ago, Moochin said:

I assume through terminal?

yes, you activate ssh in dsm web gui and use putty

maybe install package synocli-file and use midnight commander (mc) to access the files (F3 is view a file iirc)

/var/log/

dmesg and messages are the files to look into

Link to post
Share on other sites

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.

Guest
Reply to this topic...

×   Pasted as rich text.   Restore formatting

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.