Jump to content
XPEnology Community

Volume crashed after reboot


costyy

Recommended Posts

 mdadm --run /dev/md3
mdadm: failed to start array /dev/md3: Invalid argument
root@NAS:~# cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] [raidF1]
md1 : active raid1 sda2[0] sdb2[1] sdc2[2] sdd2[3] sdf2[5] sdg2[6] sdj2[7] sdk2[8] sdl2[9]
      2097088 blocks [12/9] [UUUU_UUUUU__]

md0 : active raid1 sda1[0] sdb1[5] sdc1[1] sdd1[2] sdf1[4] sdk1[8] sdl1[9]
      2490176 blocks [12/7] [UUU_UU__UU__]

unused devices: <none>
 

Link to comment
Share on other sites

mdadm --assemble /dev/md3
mdadm: /dev/md3 not identified in config file.
root@NAS:~# cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] [raidF1]
md1 : active raid1 sda2[0] sdb2[1] sdc2[2] sdd2[3] sdf2[5] sdg2[6] sdj2[7] sdk2[8] sdl2[9]
      2097088 blocks [12/9] [UUUU_UUUUU__]

md0 : active raid1 sda1[0] sdb1[5] sdc1[1] sdd1[2] sdf1[4] sdk1[8] sdl1[9]
      2490176 blocks [12/7] [UUU_UU__UU__]

unused devices: <none>
 

Link to comment
Share on other sites

 mdadm --assemble --scan
mdadm: /dev/md/3 has been started with 9 drives (out of 10).
mdadm: Found some drive for an array that is already active: /dev/md/0_0
mdadm: giving up.
root@NAS:~# cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] [raidF1]
md3 : active raid5 sda3[0] sdg3[11] sdk3[9] sdb3[10] sdl3[6] sdj3[8] sdf3[3] sdd3[2] sdc3[1]
      35119473984 blocks super 1.2 level 5, 64k chunk, algorithm 2 [10/9] [UUUU_UUUUU]

md1 : active raid1 sda2[0] sdb2[1] sdc2[2] sdd2[3] sdf2[5] sdg2[6] sdj2[7] sdk2[8] sdl2[9]
      2097088 blocks [12/9] [UUUU_UUUUU__]

md0 : active raid1 sda1[0] sdb1[5] sdc1[1] sdd1[2] sdf1[4] sdk1[8] sdl1[9]
      2490176 blocks [12/7] [UUU_UU__UU__]

unused devices: <none>
 

e1.jpeg

Edited by costyy
Link to comment
Share on other sites

mount
/dev/md0 on / type ext4 (rw,relatime,journal_checksum,barrier,data=ordered)
none on /dev type devtmpfs (rw,nosuid,noexec,relatime,size=16459916k,nr_inodes=4114979,mode=755)
none on /dev/pts type devpts (rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000)
none on /proc type proc (rw,nosuid,nodev,noexec,relatime)
none on /sys type sysfs (rw,nosuid,nodev,noexec,relatime)
/tmp on /tmp type tmpfs (rw,relatime)
/run on /run type tmpfs (rw,nosuid,nodev,relatime,mode=755)
/dev/shm on /dev/shm type tmpfs (rw,nosuid,nodev,relatime)
none on /sys/fs/cgroup type tmpfs (rw,relatime,size=4k,mode=755)
cgmfs on /run/cgmanager/fs type tmpfs (rw,relatime,size=100k,mode=755)
cgroup on /sys/fs/cgroup/cpuset type cgroup (rw,relatime,cpuset,release_agent=/run/cgmanager/agents/cgm-release-agent.cpuset,clone_children)
cgroup on /sys/fs/cgroup/cpu type cgroup (rw,relatime,cpu,release_agent=/run/cgmanager/agents/cgm-release-agent.cpu)
cgroup on /sys/fs/cgroup/cpuacct type cgroup (rw,relatime,cpuacct,release_agent=/run/cgmanager/agents/cgm-release-agent.cpuacct)
cgroup on /sys/fs/cgroup/memory type cgroup (rw,relatime,memory,release_agent=/run/cgmanager/agents/cgm-release-agent.memory)
cgroup on /sys/fs/cgroup/devices type cgroup (rw,relatime,devices,release_agent=/run/cgmanager/agents/cgm-release-agent.devices)
cgroup on /sys/fs/cgroup/freezer type cgroup (rw,relatime,freezer,release_agent=/run/cgmanager/agents/cgm-release-agent.freezer)
cgroup on /sys/fs/cgroup/blkio type cgroup (rw,relatime,blkio,release_agent=/run/cgmanager/agents/cgm-release-agent.blkio)
none on /proc/bus/usb type devtmpfs (rw,nosuid,noexec,relatime,size=16459916k,nr_inodes=4114979,mode=755)
none on /sys/kernel/debug type debugfs (rw,relatime)
securityfs on /sys/kernel/security type securityfs (rw,relatime)
none on /config type configfs (rw,relatime)
root@NAS:~# df -v
Filesystem     1K-blocks    Used Available Use% Mounted on
/dev/md0         2385528 1624852    641892  72% /
none            16459916       0  16459916   0% /dev
/tmp            16475452     928  16474524   1% /tmp
/run            16475452    4896  16470556   1% /run
/dev/shm        16475452       4  16475448   1% /dev/shm
none                   4       0         4   0% /sys/fs/cgroup
cgmfs                100       0       100   0% /run/cgmanager/fs
 

Link to comment
Share on other sites

 ls -la /
total 100
drwxr-xr-x  30 root root  4096 Apr  5 18:30 .
drwxr-xr-x  30 root root  4096 Apr  5 18:30 ..
lrwxrwxrwx   1 root root     7 Aug 27  2021 bin -> usr/bin
drwxr-xr-x   7 root root     0 Apr  5 18:30 config
drwxr-xr-x  13 root root 19340 Apr  5 19:29 dev
drwxr-xr-x  49 root root  4096 Apr  5 18:34 etc
drwxr-xr-x  43 root root  4096 Aug 27  2021 etc.defaults
drwxr-xr-x   2 root root  4096 May 12  2020 initrd
lrwxrwxrwx   1 root root     7 Aug 27  2021 lib -> usr/lib
lrwxrwxrwx   1 root root     9 Aug 27  2021 lib32 -> usr/lib32
lrwxrwxrwx   1 root root     7 Aug 27  2021 lib64 -> usr/lib
drwxr-xr-x   2 root root  4096 Aug 27  2021 .log.junior
drwx------   2 root root  4096 May 12  2020 lost+found
-rw-r--r--   1 root root     0 Apr  2 00:28 Media
drwxr-xr-x   2 root root  4096 May 12  2020 mnt
drwxr-xr-x   2 root root  4096 Aug 27  2021 .old_patch_info
drwx--x--x   4 root root  4096 Sep  2  2021 opt
dr-xr-xr-x 248 root root     0 Apr  5 18:27 proc
-rw-------   1 root root  1024 Aug 27  2021 .rnd
drwx------   5 root root  4096 Dec 18 13:08 root
drwxr-xr-x  26 root root  1360 Apr  5 20:00 run
lrwxrwxrwx   1 root root     8 Aug 27  2021 sbin -> usr/sbin
-rw-r--r--   1 root root     0 Apr  2 00:28 Server
drwxr-xr-x   4 root root  4096 Aug 27  2021 .syno
dr-xr-xr-x  12 root root     0 Apr  5 18:30 sys
drwxr-xr-x   2 root root  4096 Aug 27  2021 .system_info
drwxrwxrwt  12 root root  1440 Apr  5 20:08 tmp
drwxr-xr-x   2 root root  4096 Aug 27  2021 tmpRoot
drwxr-xr-x  11 root root  4096 May 12  2020 usr
drwxr-xr-x  17 root root  4096 Apr  5 18:30 var
drwxr-xr-x  14 root root  4096 Aug 27  2021 var.defaults
drwxrwxrwx   2 root root  4096 Apr  1 14:17 volume1
drwxr-xr-x   4 root root  4096 Jan 14 20:38 volume2
drwxr-xr-x   5 root root  4096 Nov  1 20:05 volumeSATA1
drwxr-xr-x   3 root root  4096 Nov  1 19:39 volumeSATA2
drwxr-xr-x   3 root root  4096 Nov  1 19:39 volumeSATA3
drwxr-xr-x   3 root root  4096 Apr  2 00:31 volumeUSB1
drwxr-xr-x   3 root root  4096 Jan 19 09:45 volumeUSB2
 

Link to comment
Share on other sites

Ok. So this is good news.  You should be able to access your shares on /volume2 normally via the network, or via File Station.

 

BUT, your nas is not in good shape.  The /dev/md3 aka /volume2 array is critical (no redundancy).  Any disturbance could cause loss of some or all of your data.  So please consider the following data recovery process going forward:

 

1. Do not reboot

2. Do not attempt to fix anything in Storage Manager

3. Copy ALL your data off the nas

4. Delete your Volume2/Storage Pool 2

5. Replace any disks with bad sectors

6. Fix the System Partition using the button in Storage Manager

7. Build a new Storage Pool (array), testing the integrity of the drives

8. Build a new btrfs volume

9. Copy your files back onto the nas

 

Please note that none of this does anything for your SSD, which I think is /dev/sde and normally mapped to /volume1.  I don't know what is wrong with it, or if you have any important data on it.  DSM doesn't even believe it exists at the moment.  However, it isn't really affected by what is happening with your main data array.

Edited by flyride
Link to comment
Share on other sites

If there is a drive retrying a bad sector for a long time, the transfer rate will be slow while those sectors are being encountered.  That is what you want (the drive trying to recover the data).  Usually DSM will retry and then fail a particular file.  But it might decide to crash the array if the drive error is significant enough.  Are all the member drives still visible or has one or more disks gone offline?

 

If drives are offline, that means you have some sort of continued hardware problem.  Either a cable, controller or drive is failing.  Nobody can diagnose the issue without you posting more information about the crash and circumstances.  This needs to be fixed before trying anything again.  If a drive is unusable at this point, your data is incomplete.  You could send the array to a forensic recovery specialist at great cost, and they will retrieve some data from the array, but not all of it.

 

"All the data is rubbish."  Based on the signatures of the drives, this should not be the case, unless you modified one or more of the drives while DSM was not in control of them.  I can't advise you on that.  There will be some minor corruption based on the stale drive I talked about earlier but it should only be a few files if any.

Link to comment
Share on other sites

A simple bad sector won't cause the problems you seem to be having.  I think you have a power problem or a cable problem.  Please try and resolve this first before trying to restart the array again.

 

The information you need to investigate and recover the array just like we have done is in this thread now.  Post if you need help.

Link to comment
Share on other sites

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.

Guest
Reply to this topic...

×   Pasted as rich text.   Restore formatting

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.

×
×
  • Create New...